Batch ML Predictions
Perform a batch of ML predictions, using multiple models, in one request. This is useful for ensembling or A/B testing different models.
Documentation related to Models including machine learning, LLMs, etc.
View all tagsPerform a batch of ML predictions, using multiple models, in one request. This is useful for ensembling or A/B testing different models.
Creates a model response for the given chat conversation.
Learn how Spice evaluates, tracks, compares, and improves language model performance for specific tasks
Learn how to provide LLMs with memory
Learn how to override default LLM hyperparameters in Spice.
Learn how LLMs interact with the Spice runtime.
Learn how to configure large language models (LLMs)
Return all evals available to run in the runtime.
List all models, both machine learning and language models, available in the runtime.
Get a list of spicepods and their details. In CSV format, it will return a summarised form.
Learn how to load and serve large learning models.
Spice supports loading and serving ONNX models for inference, from sources including local filesystems, Hugging Face, and the Spice.ai Cloud platform.
Make a ML prediction using a specific model.
Learn how Spice can search across datasets using database-native and vector-search methods.
Learn how Spice can perform searches using vector-based methods.