Source: wikibot/mixture-of-experts
= Mixture of experts
{wiki=Mixture_of_experts}
Mixture of Experts (MoE) is a machine learning architecture designed to improve model performance by leveraging multiple sub-models, or "experts," each specialized in different aspects of the data. The idea is to use a gating mechanism to dynamically select which expert(s) to utilize for a given input, allowing the model to adaptively allocate resources based on the complexity of the task at hand.