Spice.ai Use Cases
Spice supports a range of use cases across data infrastructure, search, and AI. Each use case below describes a specific scenario with architecture guidance, configuration examples, and links to relevant cookbook recipes.
For hands-on examples, see the Spice.ai Cookbook.
Data Federation, Acceleration, and SQL Query​
- Reverse-ETL: Serve data from warehouses and data lakes to operational systems, applications, and dashboards, eliminating complex pipelines.
- ETL-free Workflows and Data Migrations: Enable data migrations and workflows without ETL federating legacy and modern systems for faster time-to-market and lower operational overhead.
- Database CDN: Locally replicate working sets of data for operational applications, caching dynamic data for high performance, low-latency, and resilience.
- Data Mesh: Unified data access across disparate sources with acceleration.
- Object-Store Native Database: Federates, accelerates, and queries object-store data for real-time data access without centralized warehouses.
Caching​
- Write-Through Cache: Write data through Spice to both a local accelerator and the upstream source, keeping both layers consistent.
- Read-Through Cache: Fetch data from the upstream source on cache miss, with stale-while-revalidate and stale-if-error semantics.
- SQL/Database Cache: Cache SQL database tables locally with acceleration and cache SQL query results in memory.
- S3 Cache: Cache S3 and object store data locally with smart refresh skip for unchanged files.
- HTTP Cache: Cache HTTP API responses locally with request filtering, TTL, and stale-while-revalidate support.
Search and Retrieval​
- Enterprise Search: Semantic and full-text-search search with hybrid vector and keyword capabilities.
- Object-Store Native Search: Enables SQL queries, hybrid search, and LLM inference on object-store data for security applications, delivering real-time insights.
- Simplifying Real-Time Data Collection and Search: Processes streaming and static data with integrated search for real-time insights in health-tech, focusing on application logic.
Retrieval-Augmented-Generation (RAG)​
- RAG for Contextual Applications: Combines structured and unstructured data for context-rich AI outputs in SaaS chatbots, improving user interactions.
- RAG for AI-Powered Reporting: Generates dynamic, context-aware AI-driven reports for operational insights in health-tech, ensuring compliance and precision.
AI Applications and Agents​
- Real-Time Decision-Making for Intelligent Applications: Powers instant, context-aware decisions for security applications by grounding AI in federated, low-latency datasets.
- Edge-Enabled AI Applications and Agents: Deploys AI applications across cloud and edge for low-latency decisions in security IoT use cases.
- Tool-Augmented AI with Model Context Protocol Server: Extends AI with custom tools via MCP server in finserv, integrating domain-specific APIs for enhanced functionality.
- Agentic AI Applications and Agents: Builds intelligent, autonomous agents for SaaS applications, enabling context-aware automation and decision-making.
- Multi-Tenant AI Agents: Deploy AI agents across many SaaS tenants with strict isolation and no per-tenant ETL pipelines.
