[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-coding-agents":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"152858af-94a9-47e1-9e9a-c78b3c4256c0","coding agents","coding-agents",0,null,[11,20,28],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"72550eec-094c-4e09-84c4-bf451c5d3b7e","mistral-cloud-coding-agents-vibe-medium-35-en","Mistral Moves Coding Agents to the Cloud","Mistral’s Vibe now runs coding agents in the cloud with Medium 3.5, async PRs, and sandboxed sessions.","ai-agent","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778454649398-xlmv.png","en","2026-05-10T23:10:33.940319+00:00",{"id":21,"slug":22,"title":23,"summary":24,"category":25,"image_url":26,"cover_image":26,"language":18,"created_at":27},"13519d21-7023-407c-8974-7c633ebede9f","why-open-source-llms-should-be-judged-by-workload-not-hype-en","Why Open-Source LLMs Must Be Judged by Workload, Not Hype","Open-source LLMs in 2026 should be chosen by workload fit, not benchmark hype.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778095240831-b3wm.png","2026-05-06T19:20:22.800006+00:00",{"id":29,"slug":30,"title":31,"summary":32,"category":33,"image_url":34,"cover_image":34,"language":18,"created_at":35},"3dd30f79-85d5-4658-9d03-e7075c9cf646","why-claude-code-should-use-deepseek-v4-for-1m-context-en","Why Claude Code Should Use DeepSeek v4 for 1M Context","Claude Code should route through DeepSeek v4 when teams need 1M-context coding sessions.","tools","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1777867835777-eky0.png","2026-05-04T04:10:20.245382+00:00"]