The ability of a model to generalize to unseen tasks or modalities without explicit training data for them. Common in large pretrained multimodal models.
Zero-shot learning enables models to generalize to new tasks or modalities without explicit training data. This capability is achieved by leveraging knowledge from related tasks or modalities, allowing models to make predictions in unfamiliar scenarios.
Zero-shot learning often uses transfer learning and multimodal embeddings to extend model capabilities. Techniques include using shared representations and semantic mappings to bridge the gap between known and unknown tasks.