Skip to main content

Constraints

Constraints enforce data integrity in a database. Spice supports constraints on locally accelerated tables to ensure data quality and configure behavior for data updates that violate constraints.

Constraints are specified using column references in the Spicepod via the primary_key field in the acceleration configuration. Additional unique constraints are specified via the indexes field with the value unique. Data that violates these constraints will result in a conflict.

If multiple rows in the incoming data violate any constraint, the entire incoming batch of data will be dropped.

Example Spicepod:

datasets:
- from: spice.ai/eth.recent_blocks
name: eth.recent_blocks
acceleration:
enabled: true
engine: sqlite
primary_key: hash # Define a primary key on the `hash` column
indexes:
'(number, timestamp)': unique # Add a unique index with a multicolumn key comprised of the `number` and `timestamp` columns

Column References​

Column references can be used to specify which columns are part of the constraint. The column reference can be a single column name or a multicolumn key. The column reference must be enclosed in parentheses if it is a multicolumn key.

Examples

  • number: Reference a constraint on the number column
  • (hash, timestamp): Reference a constraint on the hash and timestamp columns

Handling conflicts​

The behavior of inserting data that violates the constraint can be configured via the on_conflict field to either drop the data that violates the constraint or upsert that data into the accelerated table (i.e. update all values other than the columns that are part of the constraint to match the incoming data).

warning

If there are multiple rows in the incoming data that violate any constraint, the entire incoming batch of data will be dropped.

Example Spicepod:

datasets:
- from: spice.ai/eth.recent_blocks
name: eth.recent_blocks
acceleration:
enabled: true
engine: sqlite
primary_key: hash # Define a primary key on the `hash` column
indexes:
'(number, timestamp)': unique # Add a unique index with a multicolumn key comprised of the `number` and `timestamp` columns
on_conflict:
# Upsert the incoming data when the primary key constraint on "hash" is violated,
# alternatively "drop" can be used instead of "upsert" to drop the data update.
hash: upsert

Advanced Upsert Options​

By default, even when upsert is configured, if there are constraint violations, such as duplicates within the same batch of ingested data, it will result in a constraint violation - as attempting to upsert data into the target acceleration engine results in an error if done in a single statement. (i.e. PostgreSQL does not allow the same row to be proposed for insertion more than once)

Spice provides two upsert options to resolve duplicates within a single update:

  • upsert_dedup: Removes exact duplicates in the incoming batch if there is a constraint violation. (i.e. the equivalent of running SELECT DISTINCT * FROM [batch])
  • upsert_dedup_by_row_id: Resolves conflicts by taking the row with the greatest row id. This is the behavior that would occur if the upsert were applied row-by-row. This guarantees that no constraint violations would result in an error, but it has the tradeoff of being effectively "random" if the incoming data is not ordered.

The new behavior is only triggered when an incoming batch has a constraint violation, minimizing the effect of applying these computations to only when its necessary. However, they can have a performance impact and are not enabled by default.

Full configuration example:

acceleration:
enabled: true
engine: duckdb
mode: file
primary_key: id
on_conflict:
id: upsert_dedup # upsert_dedup_by_row_id
Examples for advanced upsert behavior

Take these two CSV files:

one.csv:

foo,bar
a,1
b,2
a,1

Behavior on one.csv with a primary key on foo and on_conflict set to:

  • upsert: Will error with: Constraint Violation: Incoming data violates uniqueness constraint on column(s): foo
  • upsert_dedup: Will succeed in loading 2 rows, the a,1 row is reduced to a single instance.
  • upsert_dedup_by_row_id: Same as upsert_dedup

two.csv:

foo,bar
a,1
b,2
a,10

Behavior on one.csv with a primary key on foo and on_conflict set to:

  • upsert: Will error with: Constraint Violation: Incoming data violates uniqueness constraint on column(s): foo
  • upsert_dedup: Will error with: Constraint Violation: Incoming data violates uniqueness constraint on column(s): foo
  • upsert_dedup_by_row_id: Will succeed in loading 2 rows, a,10 and b,2. The primary key violation is resolved to the row that occurred later.

Limitations​

  • Single on_conflict target supported: Only a single on_conflict target can be specified, unless all on_conflict targets are specified with drop.

    • Examples for valid/invalid on_conflict targets

      The following Spicepod is invalid because it specifies multiple on_conflict targets with upsert:

      Invalid
      datasets:
      - from: spice.ai/eth.recent_blocks
      name: eth.recent_blocks
      acceleration:
      enabled: true
      engine: sqlite
      primary_key: hash
      indexes:
      '(number, timestamp)': unique
      on_conflict:
      hash: upsert
      '(number, timestamp)': upsert

      The following Spicepod is valid because it specifies multiple on_conflict targets with drop, which is allowed:

      Valid
      datasets:
      - from: spice.ai/eth.recent_blocks
      name: eth.recent_blocks
      acceleration:
      enabled: true
      engine: sqlite
      primary_key: hash
      indexes:
      '(number, timestamp)': unique
      on_conflict:
      hash: drop
      '(number, timestamp)': drop

      The following Spicepod is invalid because it specifies multiple on_conflict targets with upsert and drop:

      Invalid
      datasets:
      - from: spice.ai/eth.recent_blocks
      name: eth.recent_blocks
      acceleration:
      enabled: true
      engine: sqlite
      primary_key: hash
      indexes:
      '(number, timestamp)': unique
      on_conflict:
      hash: upsert
      '(number, timestamp)': drop
  • DuckDB Limitations:

    • DuckDB does not support upsert for datasets with List or Map types.
    • Standard indexes unexpectedly act like unique indexes and block updates when upsert is configured.
      • Standard indexes blocking updates

        The following Spicepod specifies a standard index on the number column, which blocks updates when upsert is configured for the hash column:

        datasets:
        - from: spice.ai/eth.recent_blocks
        name: eth.recent_blocks
        acceleration:
        enabled: true
        engine: duckdb
        primary_key: hash
        indexes:
        number: enabled
        on_conflict:
        hash: upsert

        The following error is returned when attempting to upsert data into the eth.recent_blocks table:

        ERROR runtime::accelerated_table::refresh: Error adding data for eth.recent_blocks: External error:
        Unable to insert into duckdb table: Binder Error: Can not assign to column 'number' because
        it has a UNIQUE/PRIMARY KEY constraint

        This is a limitation of DuckDB.