A foundational neural network model that learns vector representations of words from large text corpora, capturing semantic relationships. Word2Vec laid the groundwork for modern embedding techniques used across multimodal AI systems.
Word2Vec trains a shallow two-layer neural network on large text corpora using one of two architectures: Continuous Bag of Words (CBOW), which predicts a target word from surrounding context words, or Skip-gram, which predicts context words given a target word. The learned weight matrix becomes the word embedding table, where each word maps to a dense vector that encodes semantic meaning.
Word2Vec typically produces 100-300 dimensional vectors trained on sliding windows of 5-10 words. It uses negative sampling or hierarchical softmax to make training efficient on large vocabularies. The resulting vectors exhibit linear algebraic properties, such as king - man + woman = queen, demonstrating that the model captures relational semantics in vector space.