Process data through a collection - works for both bucket-sourced and collection-sourced collections.
For bucket-sourced collections:
Discovers objects from source bucket(s), creates a batch, and submits for processing.
Use include_buckets to limit which source buckets to process from.
For collection-sourced collections:
Processes existing documents from upstream collection(s).
Use include_collections to limit which source collections to process from.
Filtering:
source_filters: Field-level filters using LogicalOperator format{"AND": [{"field": "status", "operator": "eq", "value": "pending"}]}{"AND": [{"field": "object_id", "operator": "in", "value": ["obj_1", "obj_2"]}]}Returns:
Documentation Index
Fetch the complete documentation index at: https://docs.mixpeek.com/docs/llms.txt
Use this file to discover all available pages before exploring further.
REQUIRED: Bearer token authentication using your API key. Format: 'Bearer sk_xxxxxxxxxxxxx'. You can create API keys in the Mixpeek dashboard under Organization Settings.
"Bearer YOUR_API_KEY"
"Bearer YOUR_STRIPE_API_KEY"
Namespace identifier for scoping this request. All resources (collections, buckets, taxonomies, etc.) are scoped to a namespace. You can provide either the namespace name or namespace ID. Format: ns_xxxxxxxxxxxxx (ID) or a custom name like 'my-namespace'. Falls back to ?namespace= query parameter if the header is omitted.
"ns_abc123def456"
"production"
"my-namespace"
The ID or name of the collection to trigger
Request to trigger (re)processing through a collection.
For bucket-sourced collections (tier 0):
Discovers objects from source bucket(s) and creates a batch for processing.
Use include_buckets to limit which source buckets to process from.
For collection-sourced collections (tier N):
Processes existing documents from upstream collection(s).
Use include_collections to limit which source collections to process from.
Use source_filters for field-level filtering on objects or documents.
Document Overwrite Behavior:
unique_key configured: Documents are UPSERTED (overwrites existing)unique_key: New documents are CREATED (may cause duplicates)To enable idempotent re-processing, configure unique_key on the source bucket.
Limit processing to objects from these specific buckets (IDs or names). Only applies to bucket-sourced collections. If not provided, all configured source buckets are used.
Limit processing to documents from these specific collections (IDs or names). Only applies to collection-sourced collections. If not provided, all configured source collections are used.
Limit processing to these specific object IDs. Only applies to bucket-sourced collections. This is a convenience shorthand — equivalent to using source_filters with {"AND": [{"field": "object_id", "operator": "in", "value": [...]}]}.
Field-level filters for objects (bucket-sourced) or documents (collection-sourced). Uses LogicalOperator format (AND/OR/NOT). Use this to filter by metadata fields, status, or any other object/document properties.
{
"AND": [
{
"field": "status",
"operator": "eq",
"value": "pending"
}
]
}Successful Response
Response after triggering collection processing.
Use batch_id or task_id to monitor progress via GET /v1/batches/{batch_id}
or GET /v1/tasks/{task_id}.
ID of the created batch for tracking progress.
Task ID for monitoring via GET /v1/tasks/{task_id}.
ID of the collection being processed.
Number of processing tiers in the DAG.
Human-readable status message.
Bucket IDs that objects were discovered from (bucket-sourced collections).
Collection IDs that documents were read from (collection-sourced collections).
Total number of objects included in the batch (bucket-sourced collections).
Total number of documents to process (collection-sourced collections).