Skip to main content
Version: Next

Spice.ai Features

Spice provides a set of features for building data-driven applications and AI agents. This page gives an overview of each feature area.

Data Query and Federationโ€‹

Query Federation connects multiple data sourcesโ€”databases, data warehouses, and data lakesโ€”through a single SQL interface. Write one query that joins data across PostgreSQL, Snowflake, S3, and other sources. Spice pushes query operations to source databases when possible to reduce data transfer.

Data Acceleration and Cachingโ€‹

Data Acceleration materializes remote datasets locally in memory or on disk using engines like Arrow, DuckDB, SQLite, or PostgreSQL. Accelerated datasets stay current through scheduled refreshes, append mode, or Change Data Capture (CDC). Caching stores query and search results in memory with configurable TTLs and eviction policies to avoid redundant computation.

AI and Language Modelsโ€‹

Large Language Models provides an OpenAI-compatible API gateway for hosted models (OpenAI, Anthropic, xAI) and locally served models (Llama, Phi) with CUDA and Metal acceleration. Models can call tools to query datasets, run SQL, and retrieve schemas. Embeddings generates vector representations of text for semantic search and RAG workflows.

Search supports three methods: vector search (semantic similarity using embeddings), full-text search (keyword matching with BM25 scoring), and hybrid search (combining both with Reciprocal Rank Fusion). All search methods are accessible through SQL UDTFs like vector_search() and text_search().

Monitoring and Observabilityโ€‹

Observability exposes Prometheus-compatible metrics, OpenTelemetry metric export, and distributed tracing with Zipkin. Integrations are available for Datadog, Grafana, and other monitoring platforms.