Taxonomies
Create and manage hierarchical classifications for multimodal objects
Taxonomies in Mixpeek provide a structured way to organize and classify multimodal content using customizable hierarchical categories. Each node can use multiple embedding models to capture both visual and semantic aspects of content.
Quick Start Video
Watch this 5-minute live walkthrough to see taxonomies in action. The video covers creating, managing, and implementing taxonomies for multimodal content classification.
Sample Multimodal Taxonomy
Implementation
Advanced Implementation
Searching with Taxonomies
When filtering by entities[].node
, you can use either the node’s name (e.g., “serve_technique”) or its ID (e.g., “node_abc123”). The system will automatically match either format.
Internal Taxonomy Structure
Features store taxonomy classifications with hierarchical information in a simplified array structure:
Each classification contains:
node_id
: Unique node identifierdepth
: Level in the taxonomy tree (0 = root)score
: Classification confidence scoreorder
: Array representing path in tree
This structure enables several filtering patterns:
Automatic Classification During Ingestion
Best Practices for Video Taxonomies
Video Description
- Describe key visual elements
- Include relevant actions and movements
- Specify important technical details
Multimodal Configuration
- Use video embeddings for visual content
- Add text embeddings for semantic context
- Combine multiple models for better accuracy
Hierarchy Design
- Group similar techniques together
- Create logical progression paths
- Maintain consistent categorization
Video embedding processing can be resource-intensive. Consider using key frames or segments for initial classification.
The vertex-multimodal model can process both video frames and text descriptions, making it ideal for video content classification.
Performance Considerations
Was this page helpful?