The practice of anchoring AI model outputs to specific, verifiable source data or real-world references. Grounding is essential for building trustworthy multimodal AI systems that produce accurate, attributable, and verifiable results.
Grounding connects model predictions or generated text to specific evidence in the source data. For language models, this means linking claims in generated text to retrieved documents or database records. For multimodal models, grounding connects textual descriptions to specific regions, timestamps, or elements in visual or audio content. This enables users to verify outputs and builds trust in AI-generated information.
Implementation approaches include citation generation (linking statements to source documents), visual grounding (bounding boxes for referred objects), temporal grounding (timestamps for described events), and fact verification against knowledge bases. Grounded generation models produce outputs with inline references to source material. Evaluation measures attribution accuracy, source relevance, and faithfulness to cited evidence.