[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-llms":3},{"tag":4,"articles":11},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":10},"909f2c64-0896-45ad-b655-081c53ab7f21","LLMs","llms",7,"LLM（大型語言模型）是當前生成式 AI 的核心，從聊天助理、企業代理流程到廣告投放與內容生成都離不開它。這個主題也涵蓋偏誤、對齊、越獄與模型內部機制，因為它直接影響系統可靠性與實際部署風險。","LLMs are the core engine behind modern generative AI, powering chat assistants, enterprise agents, ad systems, and content generation. This tag also covers bias, alignment, jailbreak resistance, and internal model behavior, all of which shape reliability in real deployments.",[12,21,29,36,43,50,57,64,71,78,85,92],{"id":13,"slug":14,"title":15,"summary":16,"category":17,"image_url":18,"cover_image":18,"language":19,"created_at":20},"afddc8c2-ae3d-416b-bacd-63d8d4e4899b","autotts-llms-discover-test-time-scaling-en","AutoTTS lets LLMs discover test-time scaling","AutoTTS turns test-time scaling into an environment search problem, letting LLMs discover cheaper reasoning strategies automatically.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778479852627-3ju7.png","en","2026-05-11T06:10:31.579371+00:00",{"id":22,"slug":23,"title":24,"summary":25,"category":26,"image_url":27,"cover_image":27,"language":19,"created_at":28},"2d033835-7c64-4e54-82cf-c19145e4a2d0","why-small-language-models-should-replace-llm-first-enterpris-en","Why small language models should replace LLM-first enterprise AI","Enterprise AI should default to small language models, not giant LLMs, because they are cheaper, faster, and safer for most workflows.","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778461855512-rkkc.png","2026-05-11T01:10:24.598783+00:00",{"id":30,"slug":31,"title":32,"summary":33,"category":17,"image_url":34,"cover_image":34,"language":19,"created_at":35},"fcba2ffc-9687-40b6-b58c-a36dc8b4926b","retrieval-augmented-generation-explained-en","Retrieval-Augmented Generation, Explained Simply","RAG lets large language models pull fresh facts from documents before answering, which cuts hallucinations and adds citations.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778083860476-4o28.png","2026-05-06T16:10:34.177377+00:00",{"id":37,"slug":38,"title":39,"summary":40,"category":17,"image_url":41,"cover_image":41,"language":19,"created_at":42},"86e88a6b-78cc-45d7-9ee0-ec903e69928e","selective-llm-regularization-recommenders-en","Selective LLM Regularization for Recommenders","A paper on using selective LLM-guided regularization to improve recommendation models without overhauling the recommender stack.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778053271931-89py.png","2026-05-06T07:40:37.319427+00:00",{"id":44,"slug":45,"title":46,"summary":47,"category":17,"image_url":48,"cover_image":48,"language":19,"created_at":49},"f414aa1a-27e8-45d9-b407-d542121915d2","llms-procedural-execution-diagnostic-study-en","When LLMs Stop Following Procedural Steps","A diagnostic benchmark shows LLMs lose procedural fidelity as step counts grow, even when the arithmetic stays simple.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1777875670060-pmbt.png","2026-05-04T06:20:27.84519+00:00",{"id":51,"slug":52,"title":53,"summary":54,"category":17,"image_url":55,"cover_image":55,"language":19,"created_at":56},"bafbed2b-3504-43e8-aecf-aefd70bbf2d9","llm-narratives-global-majority-nationalities-en","How LLMs Stereotype Global Majority Nationalities","A study finds widely used LLMs produce harmful, one-sided narratives about national origins, especially when US cues appear in prompts.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1777270205101-zesr.png","2026-04-27T06:09:37.615061+00:00",{"id":58,"slug":59,"title":60,"summary":61,"category":17,"image_url":62,"cover_image":62,"language":19,"created_at":63},"4c790996-8883-4535-ac50-350e7882f0a3","llms-harmful-content-unified-mechanism-en","How LLMs encode harmful behavior internally","A pruning study suggests harmful output lives in a compact, shared weight set—helping explain jailbreak brittleness and emergent misalignment.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1776060239898-zqk2.png","2026-04-13T06:03:38.491469+00:00",{"id":65,"slug":66,"title":67,"summary":68,"category":26,"image_url":69,"cover_image":69,"language":19,"created_at":70},"d8586da7-8b63-459f-b221-8b2d3f0e054f","chatgpt-ads-format-standardization-data-en","ChatGPT Ads Are Getting More Uniform","New data from 40,000 ad placements shows ChatGPT ads are becoming shorter, clearer, and more standardized as OpenAI optimizes for conversion.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775218197706-et7g.png","2026-04-03T12:09:37.828183+00:00",{"id":72,"slug":73,"title":74,"summary":75,"category":26,"image_url":76,"cover_image":76,"language":19,"created_at":77},"3f8ce806-cc89-41ac-aae7-83ed9afd34b7","what-agentic-workflows-actually-do-enterprise-ai-en","What Agentic Workflows Actually Do in Enterprise AI","Agentic workflows let AI agents plan, act, and adapt with little human input, changing how teams handle support, ops, and data work.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775149805983-zsy9.png","2026-04-02T17:09:38.973748+00:00",{"id":79,"slug":80,"title":81,"summary":82,"category":17,"image_url":83,"cover_image":83,"language":19,"created_at":84},"ea6494a5-5f7a-4896-8fe8-c26737159834","duplicate-prompts-can-lift-accuracy-fast-en","Duplicate Prompts Can Lift Accuracy Fast","A Google study found repeating prompts once improved 47 of 70 model-benchmark pairs, with one task jumping from 21% to 97%.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122510439-jn67.png","2026-04-02T08:39:34.706953+00:00",{"id":86,"slug":87,"title":88,"summary":89,"category":17,"image_url":90,"cover_image":90,"language":19,"created_at":91},"27f0d044-b9f9-4a58-99e8-1a181ea32f19","universal-yoco-efficient-depth-scaling-en","Universal YOCO aims to scale depth without cache bloat","YOCO-U mixes recursive computation with efficient attention to scale LLM depth while keeping inference overhead and KV cache growth in check.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775115621645-wqql.png","2026-04-02T06:06:26.960639+00:00",{"id":93,"slug":94,"title":95,"summary":96,"category":97,"image_url":98,"cover_image":98,"language":19,"created_at":99},"9afcaae9-1e3d-47c1-8cca-08a2a4cb3185","what-ai-agents-are-how-they-work-en","What AI Agents Are and How They Work","AI agents combine LLMs, memory, tools, and planning. IBM says they can call APIs, search data, and coordinate tasks autonomously.","ai-agent","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775200012337-2kcd.png","2026-04-02T05:42:30.056009+00:00"]