[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-llama-3":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"5de316be-5f51-4126-9bfb-47727314e149","Llama 3","llama-3",2,null,[11,20],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"a0660205-5b41-49a6-8119-ee9105a7e1f5","chatgpt-ads-format-standardization-data-zh","ChatGPT 廣告越來越一致","40,000 筆廣告版位分析顯示，ChatGPT 廣告正變得更短、更直白、更標準化。這反映 OpenAI 在優化轉換，也透露 LLM 使用習慣正在往任務導向收斂。","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775218190861-p9x8.png","zh","2026-04-03T12:09:37.164139+00:00",{"id":21,"slug":22,"title":23,"summary":24,"category":25,"image_url":26,"cover_image":26,"language":18,"created_at":27},"e7d8242f-edab-4282-8317-9a27fec3cb91","sebastian-raschka-llm-architecture-gallery-zh","Sebastian Raschka 的 LLM 架構圖鑑","Raschka 的 LLM Architecture Gallery 把 GPT-2、Llama 3、OLMo 2、DeepSeek、Qwen 等模型的層數、注意力與 KV cache 數字攤開來比，工程師一眼就能看出部署差異。","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775121663540-srg4.png","2026-04-02T07:27:33.561537+00:00"]