[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-compute-cost":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"facd8e74-8c13-425c-8283-c5ef7c9d8974","compute cost","compute-cost",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"37045a8c-9166-4ba7-8f62-fcd8e0593665","ae-llm-adaptive-efficiency-optimization-zh","AE-LLM 要讓大模型更省算力","AE-LLM 主打大型語言模型的自適應效率最佳化，想在不固定耗算力的前提下，讓模型依工作負載調整效率；但摘要沒有公開完整 benchmark 細節。","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778051455312-7tw1.png","zh","2026-05-06T07:10:32.541013+00:00"]