[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-dynamo":3},{"tag":4,"articles":11},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":10},"3c3befd2-247b-4dac-9387-1cba94b18d03","Dynamo","dynamo",3,"Dynamo 指的是 NVIDIA 用來提升推論效率的軟體層與執行優化，常與 TensorRT-LLM、Blackwell Ultra、GB300 NVL72 一起出現。它的重要性在於，AI 伺服器的速度與成本，已不只取決於 GPU 規格，也取決於排程、記憶體管理與模型執行策略。","Dynamo refers to NVIDIA’s software layer and execution optimizations for inference, often discussed alongside TensorRT-LLM, Blackwell Ultra, and GB300 NVL72. It matters because AI server speed and cost now depend not only on GPU hardware, but also on scheduling, memory handling, and model execution strategy.",[12,21],{"id":13,"slug":14,"title":15,"summary":16,"category":17,"image_url":18,"cover_image":18,"language":19,"created_at":20},"a15782d7-4678-4415-9a0b-4c642e46b022","nvidia-mlperf-software-inference-benchmarks-en","Nvidia’s MLPerf Gains Show Software Still Matters","Nvidia posted up to 2.77x MLPerf gains on GB300 NVL72, with software tricks like Dynamo and TensorRT-LLM doing heavy lifting.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775185791842-obyu.png","en","2026-04-03T03:09:35.154603+00:00",{"id":22,"slug":23,"title":24,"summary":25,"category":26,"image_url":27,"cover_image":27,"language":19,"created_at":28},"3e10b782-08fe-4a58-aabc-0f4ca77eaa50","nvidia-sets-new-mlperf-inference-records-en","NVIDIA Sets New MLPerf Inference Records","Blackwell Ultra hit new MLPerf Inference v6.0 highs, with GB300 NVL72 gaining 2.7x on DeepSeek-R1 server tests and 1.5x on Llama 3.1 405B.","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122498583-yuhr.png","2026-04-02T08:48:38.893048+00:00"]