[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-mlperf-inference":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"96989e04-ffa0-4898-bb21-4b3aaabeb8fd","MLPerf Inference","mlperf-inference",1,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"3e10b782-08fe-4a58-aabc-0f4ca77eaa50","nvidia-sets-new-mlperf-inference-records-en","NVIDIA Sets New MLPerf Inference Records","Blackwell Ultra hit new MLPerf Inference v6.0 highs, with GB300 NVL72 gaining 2.7x on DeepSeek-R1 server tests and 1.5x on Llama 3.1 405B.","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122498583-yuhr.png","en","2026-04-02T08:48:38.893048+00:00"]