[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-deepseek-v4":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"0ac20b1f-f360-43fb-b37f-3df933328d51","DeepSeek v4","deepseek-v4",1,null,[11,20],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"3dd30f79-85d5-4658-9d03-e7075c9cf646","why-claude-code-should-use-deepseek-v4-for-1m-context-en","Why Claude Code Should Use DeepSeek v4 for 1M Context","Claude Code should route through DeepSeek v4 when teams need 1M-context coding sessions.","tools","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1777867835777-eky0.png","en","2026-05-04T04:10:20.245382+00:00",{"id":21,"slug":22,"title":23,"summary":24,"category":25,"image_url":26,"cover_image":26,"language":18,"created_at":27},"61ed1d6b-505f-4cf5-b132-2d57964ca4c2","gpt-5-4-vs-claude-opus-4-6-ai-benchmark-en","GPT-5.4 vs Claude Opus 4.6: 75% Win Rate","We tested GPT-5.4, Claude Opus 4.6, DeepSeek V4, and Gemini 3.1 across 12 benchmarks. One model won 9 of them.","model-release","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775127830823-xco3.png","2026-04-02T09:12:38.725884+00:00"]