Subcategories:
Showing 10 of 10 tools
Mixpeek
Multimodal Infrastructure
Multimodal data infrastructure platform that indexes, processes, and retrieves across video, image, audio, and text with unified pipelines and search.
Key features:
OpenAI
Foundation Models
AI research and deployment company behind GPT-4, DALL-E, and Whisper, providing multimodal AI models through APIs and ChatGPT.
Key features:
Google Vertex AI
ML Platform
Google Cloud ML platform providing access to Gemini models, AutoML, and custom training for building multimodal AI applications at scale.
Key features:
Anthropic
Foundation Models
AI safety company providing Claude, a multimodal AI assistant capable of analyzing text, images, and code with a focus on helpfulness and safety.
Key features:
Coactive AI
Visual Data
Visual data platform that enables teams to search, analyze, and organize image and video content using multimodal AI understanding.
Key features:
Amazon Bedrock
ML Platform
Fully managed service from AWS providing access to foundation models from leading AI companies for building generative AI applications.
Key features:
Meta AI
Foundation Models
Meta open-source AI research lab behind LLaMA, Segment Anything, and ImageBind, advancing multimodal understanding across text, image, and video.
Key features:
Microsoft Azure AI
ML Platform
Comprehensive cloud AI platform from Microsoft providing vision, speech, language, and generative AI services including Azure OpenAI Service.
Key features:
Hugging Face
ML Platform
Open-source AI platform and community hub hosting models, datasets, and spaces, providing tools for building and deploying ML applications.
Key features:
Replicate
Model Hosting
Cloud platform for running open-source machine learning models via API, making it easy to deploy and scale models without managing infrastructure.
Key features:
Explore Other Categories
Need a Multimodal Solution?
Mixpeek processes video, image, audio, and text through unified pipelines. See how it compares to the tools listed in this directory.
