[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-glm-5":3},{"tag":4,"articles":11},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":10},"d79e91f0-2882-4696-ae41-cf7f4c3163b8","GLM-5","glm-5",4,"GLM-5 是 Z.AI 的旗艦大型語言模型，主打程式碼生成、代理式工作流與長上下文推理。它在 SWE-bench、Terminal Bench 等基準的表現，讓開源模型在實務開發與自架部署上更值得比較。","GLM-5 is Z.AI’s flagship large language model, built for coding, agent workflows, and long-context reasoning. Its results on SWE-bench and Terminal Bench make it a relevant benchmark when comparing open models for real development and self-hosted deployment.",[12,21],{"id":13,"slug":14,"title":15,"summary":16,"category":17,"image_url":18,"cover_image":18,"language":19,"created_at":20},"57576af6-0bf2-4616-ac89-8435e39a8aa7","glm-5-zai-flagship-coding-agents-zh","GLM-5 登場：Z.AI 的寫程式旗艦","GLM-5 是 Z.AI 的新旗艦模型。744B 總參數、200K context、SWE-bench Verified 77.8、Terminal Bench 2.0 56.2，直接挑戰頂級 coding 模型。","model-release","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775135063109-o1yh.png","zh","2026-04-02T13:03:42.135022+00:00",{"id":22,"slug":23,"title":24,"summary":25,"category":17,"image_url":26,"cover_image":26,"language":19,"created_at":27},"710ff4cc-d333-4bd8-b50a-e5522d430161","open-source-llm-comparison-2026-zh","2026 開源 LLM 誰領先","Qwen 3.5、GLM-5、DeepSeek R1、Llama 4 讓開源 LLM 進入實戰。這篇整理 2026 年主流模型的 benchmark、上下文長度、授權條款與自架表現。","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775131800331-8pqc.png","2026-04-02T12:09:39.445524+00:00"]