[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-mixture-of-experts":3},{"tag":4,"articles":9},{"id":5,"name":6,"slug":6,"article_count":7,"description_zh":8,"description_en":8},"3558c57d-a009-4f0e-bf6a-20cea0b81732","mixture-of-experts",2,null,[10],{"id":11,"slug":12,"title":13,"summary":14,"category":15,"image_url":16,"cover_image":16,"language":17,"created_at":18},"072a2114-1f7f-4d61-99f7-be82c686c286","unipool-shared-expert-pool-moe-zh","UniPool：共享 MoE 專家池","UniPool 把 MoE 的分層專家改成全域共享池，減少重複參數，並在五個 LLaMA 規模模型上改善驗證損失。","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778221269156-lam7.png","zh","2026-05-08T06:20:40.070989+00:00"]