[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-triton-inference-server":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"89369a95-ecc7-40d5-985c-c39121a7034f","Triton Inference Server","triton-inference-server",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"a4380666-3f3c-4465-be35-903068c7045e","how-to-reduce-ai-model-serving-friction-zh","怎麼降低 AI 模型部署摩擦","這篇教你把訓練好的 AI 模型穩定送進 production，透過 ONNX、TensorRT、動態輸入、版本鎖定與 Triton 檢查，建立可重複的部署流程。","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778922836413-ff99.png","zh","2026-05-16T09:13:31.665292+00:00"]