A search technique that converts data into high-dimensional vector embeddings and retrieves results by finding the nearest vectors in embedding space, enabling semantic understanding beyond keyword matching.
Vector search converts content (text, images, audio, video) into dense numerical representations called embeddings using neural networks. These embeddings capture semantic meaning in a high-dimensional space where similar concepts cluster together. When a query arrives, it is embedded using the same model, and approximate nearest neighbor (ANN) algorithms like HNSW find the closest vectors in the index, returning semantically relevant results regardless of exact keyword overlap.
The core components of a vector search system are: an embedding model (CLIP, SigLIP, sentence-transformers, or custom models) that maps content to vectors, a vector index (Qdrant, FAISS, or similar) that organizes vectors for fast retrieval using algorithms like HNSW or IVF, and a distance metric (cosine similarity, Euclidean distance, or dot product) that determines how similarity is measured. Modern systems combine vector search with metadata filtering for hybrid retrieval, and use quantization techniques like scalar or product quantization to reduce memory footprint while maintaining accuracy.