Single Object Upload
skip_duplicatescompares blob hashes to avoid storing identical uploads.key_prefixprovides logical grouping for later queries.- Object creation is idempotent—safe to retry failed requests.
Batch Upload
Bulk ingestion is faster when you create objects in batch and submit a single processing job.task_id; poll /v1/tasks/{task_id} or watch the batch resource directly.
Upload Sources
| Source | Strategy |
|---|---|
| Browser / client app | Upload to your storage, then register object URLs with Mixpeek |
| Data lake (S3/GCS) | Point blobs to existing URIs and set skip_duplicates=true |
| Streaming ingestion | Combine incremental object creation with periodic batch submissions |
| Large archives | Pre-process externally, split into multiple objects, and enqueue batches |
Tips for High-Volume Imports
- Use multiple smaller batches (1k–10k objects) to keep pollers responsive.
- Parallelize object registration—the API layer scales horizontally.
- Monitor tasks with the Tasks API or webhooks to coordinate downstream systems.
- Leverage retries; object creation and batch submission are safe to repeat on failure.
- Tag metadata that will help with retrieval filters or taxonomy enrichment later.

