[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-inference-cost":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"28e70747-db4a-4095-b5c1-a082bd9f32f5","inference cost","inference-cost",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"5b27896f-ad48-4a9a-8b6e-823568d8c669","wei-shen-me-lu-you-cai-shi-mo-xing-fu-wu-de-zhen-zheng-ping-zh","為什麼路由才是模型服務的真正瓶頸","模型服務的主要限制不是推理本身，而是路由決策；誰、何時、送到哪個模型與副本，才決定延遲、成本與穩定性。","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778278838578-nms7.png","zh","2026-05-08T22:20:22.020009+00:00"]