The use of AI to translate text from one natural language to another while preserving meaning. Machine translation enables multilingual content processing and cross-lingual search in global multimodal systems.
Neural machine translation uses encoder-decoder transformer models to translate text. The encoder processes the source language sentence into contextualized representations, and the decoder generates the target language sentence token by token. Attention mechanisms align source and target tokens. Modern systems handle over 100 languages and produce near-human quality for well-resourced language pairs.
State-of-the-art systems include NLLB (No Language Left Behind, 200 languages), mBART, and M2M-100. Commercial APIs (Google Translate, DeepL) use proprietary large-scale models. Multilingual models share parameters across languages, enabling zero-shot translation between unseen language pairs. Quality is measured using BLEU, chrF, and COMET scores, with human evaluation for production systems.