A dense vector representation of data (text, image, audio, video) in a shared semantic space that enables similarity operations and cross-modal queries.
Embeddings convert high-dimensional data into dense vectors that capture semantic meaning. These vectors enable similarity comparisons and can be used for search, clustering, and other machine learning tasks.
Generated using neural networks trained on large datasets. Different architectures are used for different modalities (e.g., BERT for text, ResNet for images). Vectors typically range from 256 to 1536 dimensions.