TH
← Back
news 2026-04-23 Β· huggingface-papers

🧩 Train Giant AI Models 3x Cheaper by Cloning Their Own Experts

🧩 Train Giant AI Models 3x Cheaper by Cloning Their Own Experts

What if you could triple your AI model's capacity without tripling the training cost?

A new paper from Amazon researchers introduces "Expert Upcycling" β€” a method that grows Mixture-of-Experts (MoE) models by duplicating their existing experts and letting them specialize further, instead of training larger models from scratch.


The problem is clear: training frontier AI models costs millions in compute. Every time you want a bigger model, you start over β€” throwing away everything the smaller model already learned.

Expert Upcycling flips this entirely. Take a trained MoE model, clone its experts (8 become 16 or 32), and continue training. The clones inherit their parent's knowledge, then gradually develop unique specializations.


🎯 The results are striking:


Think of it like a restaurant: instead of hiring and training 24 new chefs, your 8 best chefs each train a protΓ©gΓ© who starts at their level and develops their own style. The kitchen triples in capacity, but each dish still takes the same time.

This could be a turning point in making large-scale AI accessible beyond just the biggest labs.

πŸ“„ Source

huggingface-papers
Share: Facebook 𝕏
← Previous
A Cyberpunk Short Film Made Entirely with Open-Sou
Next β†’
GPT-5.5 System Card: OpenAI's Smartest Model Gets