Quick Links
- API Reference — Explore the API endpoints in these docs
- Full API Docs — Complete OpenAPI documentation
- Studio — Access all resources via the UI
Decompose with Extractors
Break complex objects into semantic layers. A single video becomes searchable transcripts, visual embeddings, scene descriptions, and detected entities—each layer independently queryable.Breaking down a video into semantic layers
- Search within any modality (find spoken words, visual moments, or document sections)
- Extract structured data from unstructured content
- Build multiple indexes from a single source file
Recompose with Retrievers
Reassemble layers based on semantic relevance. Chain search stages, apply filters across modalities, and enrich results—turning decomposed content back into meaningful answers.Chaining search stages to recompose results
- Multi-stage retrieval pipelines (search → filter → rank → enrich)
- Cross-modal queries (text query finds video moments)
- Dynamic result enrichment without re-indexing
How It Works
The platform has three main pieces that work together:1
2
Processing
Create Collections that run ML models (CLIP, Whisper, LayoutLM, etc.) on your objects. Models extract embeddings and structured data, then store everything as searchable documents.
3
Retrieval
Build Retrievers to search across your documents. Chain multiple search stages together, apply filters, and enrich results—all through configuration, no code.
Example Flow
Key Features
Multi-Tenant by Default
Every request includes a namespace header. Keep customers, environments, and projects completely isolated.
Configuration Over Code
Define pipelines, extractors, and retrievers with JSON configs. No infrastructure to manage, no model deployments to worry about.
Enrichment & Discovery
Add taxonomies for classification or clusters for content discovery—both use the same semantic JOIN primitives.
Built for Production
Analytics, caching, webhooks, and monitoring built in. Track every request, A/B test retrieval strategies, and roll back configs when needed.
What You Get
- No infrastructure work: No Ray clusters to manage, no model serving to configure, no vector DB ops
- Mix any models: CLIP for images, Whisper for audio, LayoutLM for documents—use them together in the same collection
- Semantic JOINs: Connect documents across collections using vector similarity instead of foreign keys
- Complete lineage: Trace any result back through document → object → source file

