[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"glossary-articles-inference-zh-TW":3},[4,20,35,45,57],{"id":5,"slug":6,"title":7,"summary":8,"source":9,"category":10,"image_url":11,"cover_image":11,"language":12,"views":13,"created_at":14,"published_at":15,"topic_cluster_id":16,"topic_clusters":17},"5b5fa24f-5259-4e9e-8270-b08b6805f281","minimax-m1-open-hybrid-attention-reasoning-model-zh","MiniMax-M1：開源 1M Token 推理模型","MiniMax 推出 M1 開源推理模型，主打 100 萬 Token 上下文、8 萬 Token 輸出與低價 API。","www.minimax.io","model-release","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778797859209-ea1g.png","zh",2,"2026-05-14T22:30:38.636592+00:00","2026-05-14T22:30:38.611+00:00","0ccb5d2e-69f1-4354-a3e0-cb370221cd95",{"slug":18,"title":19},"-xiaomi-mimo-v25-pro-coding-","為什麼 Xiaomi 的 MiMo-V2.5-Pro 改變的是 Coding …",{"id":21,"slug":22,"title":23,"summary":24,"source":25,"category":26,"image_url":27,"cover_image":27,"language":12,"views":28,"created_at":29,"published_at":30,"topic_cluster_id":31,"topic_clusters":32},"d75b5708-d4ec-4c46-9592-fa0a68d4bc26","judge-reliability-harness-stress-tests-llm-judges-zh","LLM 評審也會不穩","這篇論文做了一個壓力測試工具，檢查 LLM 當評審時，會不會因為格式、改寫、篇幅或標籤翻轉而判斷不一致。","arxiv.org","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778740856189-g1zr.png",0,"2026-05-14T06:40:32.198872+00:00","2026-05-14T06:40:32.036+00:00","0c35a120-52fc-41fc-afa3-d404eb934158",{"slug":33,"title":34},"longmemeval-v2-agent-","LongMemEval-V2：測 agent 長期記憶",{"id":36,"slug":37,"title":38,"summary":39,"source":25,"category":26,"image_url":40,"cover_image":40,"language":12,"views":41,"created_at":42,"published_at":43,"topic_cluster_id":31,"topic_clusters":44},"941f698a-1dcf-4807-bd56-5295c07d2dee","taming-black-box-llm-inference-scheduling-zh","黑箱 LLM 排程更聰明了","這篇論文用「預測輸出長度」來改善黑箱 LLM 推論排程，想在看不到模型內部的情況下，減少排隊摩擦、提升大規模服務效率。","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778740253221-wgy6.png",1,"2026-05-14T06:30:31.546746+00:00","2026-05-14T06:30:31.423+00:00",{"slug":33,"title":34},{"id":46,"slug":47,"title":48,"summary":49,"source":50,"category":26,"image_url":51,"cover_image":51,"language":12,"views":28,"created_at":52,"published_at":53,"topic_cluster_id":54,"topic_clusters":55},"1de70fef-f10d-4ff6-b162-d58b527a168c","ollama-memory-leak-cve-2026-7482-zh","Ollama 漏洞可遠端洩漏記憶體","Ollama 出現 CVE-2026-7482，遠端攻擊者可透過惡意 GGUF 檔讀出程序記憶體，可能外洩金鑰、提示詞與使用者資料。","thehackernews.com","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778713858981-cpex.png","2026-05-13T23:10:35.420032+00:00","2026-05-13T23:10:35.166+00:00","8575dc44-8be8-4a20-a937-db265af3e9e9",{"slug":56,"title":48},"ollama-",{"id":58,"slug":59,"title":60,"summary":61,"source":62,"category":63,"image_url":64,"cover_image":64,"language":12,"views":28,"created_at":65,"published_at":66,"topic_cluster_id":67,"topic_clusters":68},"4adef3ab-9f07-4970-91cf-77b8b581b348","why-databricks-model-serving-is-right-default-zh","為什麼 Databricks Model Serving 是生產推論的正確預設","Databricks Model Serving 應該成為生產推論的預設選項，因為它把部署、治理與擴展整合在同一個平台，降低多模型團隊的營運成本。","docs.databricks.com","tools","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778692245329-a2wt.png","2026-05-13T17:10:30.659153+00:00","2026-05-13T17:10:30.443+00:00","6d7e007a-da26-455a-af52-fcbe1ddc66a0",{"slug":69,"title":60},"-databricks-model-serving-"]