A machine learning approach that combines predictions from multiple models to achieve better accuracy and robustness than any individual model. Ensemble methods improve reliability in multimodal AI systems where no single model excels at all aspects of the task.
Ensemble methods aggregate predictions from multiple models using strategies like majority voting (classification), averaging (regression), stacking (meta-learner), or boosting (sequential correction). Each model may specialize in different aspects of the data, and their combined output reduces individual model errors. The diversity of the ensemble members is key to improvement: models that make different errors provide the most complementary signal.
Common approaches include bagging (random subsets, e.g., Random Forest), boosting (sequential error correction, e.g., XGBoost), and stacking (training a meta-model on base model outputs). For deep learning, ensembles combine models with different architectures, training data, or hyperparameters. Weighted ensembles optimize weights on a validation set. Ensemble size is typically 3-10 models, with diminishing returns beyond that.