Tag
MoE
MoE, or Mixture of Experts, is an architecture that activates only a subset of experts per token or task, balancing total parameter count, inference cost, and quality. It shows up in open coding models, long-context agents, and other systems built for efficient scaling.
1 articles
