[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"article-kubernetes-becoming-ais-control-plane-en":3,"tags-kubernetes-becoming-ais-control-plane-en":30,"related-lang-kubernetes-becoming-ais-control-plane-en":40,"related-posts-kubernetes-becoming-ais-control-plane-en":44,"series-industry-ebda74d3-8122-455a-addd-1ade341b2542":81},{"id":4,"title":5,"content":6,"summary":7,"source":8,"source_url":9,"author":10,"image_url":11,"keywords":12,"language":18,"translated_content":10,"views":19,"is_premium":20,"created_at":21,"updated_at":21,"cover_image":11,"published_at":22,"rewrite_status":23,"rewrite_error":10,"rewritten_from_id":24,"slug":25,"category":26,"related_article_id":27,"status":28,"google_indexed_at":29,"x_posted_at":10,"tweet_text":10,"title_rewritten_at":10,"title_original":10,"key_takeaways":10,"topic_cluster_id":10,"embedding":10,"is_canonical_seed":20},"ebda74d3-8122-455a-addd-1ade341b2542","Kubernetes Is Becoming AI’s Control Plane","\u003Cp>KubeCon Europe 2026 drew more than 13,500 attendees, up about 10% year over year, and the message from \u003Ca href=\"https:\u002F\u002Fwww.cncf.io\u002F\" target=\"_blank\" rel=\"noopener\">CNCF\u003C\u002Fa> was plain: Kubernetes is no longer just where cloud-native apps run. It is becoming the control plane for AI infrastructure, especially inference.\u003C\u002Fp>\u003Cp>That shift matters because the AI conversation has moved past demos and into operations. The hard problem now is serving models reliably, routing requests efficiently, and keeping GPU-heavy systems from turning into an expensive mess.\u003C\u002Fp>\u003Ch2>Why KubeCon’s AI message landed so hard\u003C\u002Fh2>\u003Cp>The event in Amsterdam was the largest KubeCon to date, with more than 13,500 attendees from over 100 countries, 3,000-plus organizations, and nearly 900 sessions. CNCF also said the cloud-native developer base is approaching 20 million people. Those numbers are a reminder that this ecosystem has real gravity.\u003C\u002Fp>\n\u003Cfigure class=\"my-6\">\u003Cimg src=\"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775178585987-q0fi.png\" alt=\"Kubernetes Is Becoming AI’s Control Plane\" class=\"rounded-xl w-full\" loading=\"lazy\" \u002F>\u003C\u002Ffigure>\n\u003Cp>What changed this year is where that gravity is pointing. CNCF leadership framed AI infrastructure as the next major workload for cloud native, with inference taking center stage. Training still matters, but inference is what enterprises will run every day, at scale, and under pressure from latency, cost, and reliability targets.\u003C\u002Fp>\u003Cp>That is why Kubernetes keeps coming up in AI conversations. It already solves scheduling, service discovery, scaling, and policy for \u003Ca href=\"\u002Fnews\u002Fwhy-distributed-systems-feel-so-weird-en\">distributed systems\u003C\u002Fa>. AI workloads now need those same capabilities, plus better handling for GPUs, stateful routing, and model serving patterns that do not behave like ordinary web traffic.\u003C\u002Fp>\u003Cul>\u003Cli>13,500+ attendees at KubeCon Europe 2026\u003C\u002Fli>\u003Cli>100+ countries represented\u003C\u002Fli>\u003Cli>3,000+ organizations in attendance\u003C\u002Fli>\u003Cli>Nearly 900 sessions across the conference\u003C\u002Fli>\u003Cli>Cloud-native developer base approaching 20 million\u003C\u002Fli>\u003C\u002Ful>\u003Ch2>The AI announcements were more than conference theater\u003C\u002Fh2>\u003Cp>The most important signal came from the ecosystem itself. \u003Ca href=\"https:\u002F\u002Fwww.nvidia.com\u002F\" target=\"_blank\" rel=\"noopener\">NVIDIA\u003C\u002Fa> joined CNCF as a platinum member, donated its GPU driver to Kubernetes SIG Node as a reference implementation for the vendor-neutral DRA API, and pledged $4 million over three years to support GPU access for CNCF projects.\u003C\u002Fp>\u003Cp>That is the kind of move that tells you the hardware layer wants in on the standards conversation. If Kubernetes is going to remain the operating layer for AI infrastructure, GPU access cannot stay trapped inside vendor-specific tooling. The same goes for scheduling, routing, and serving.\u003C\u002Fp>\u003Cp>CNCF also announced \u003Ca href=\"https:\u002F\u002Fgithub.com\u002Fllmd\u002Fllmd\" target=\"_blank\" rel=\"noopener\">LLMD\u003C\u002Fa> as a new sandbox project. LLMD is pitched as a distributed inference system built around Kubernetes, which is a strong sign that the community is now treating inference as a first-class infrastructure problem, not an application afterthought.\u003C\u002Fp>\u003Cblockquote>“Inference is where the money is,” said Jensen Huang, CEO of NVIDIA, onstage at \u003Ca href=\"https:\u002F\u002Fwww.nvidia.com\u002Fen-us\u002Fon-demand\u002Fsession\u002Fgtcspring25-s81939\u002F\" target=\"_blank\" rel=\"noopener\">GTC 2025\u003C\u002Fa>. The line fits this moment almost too well.\u003C\u002Fblockquote>\u003Cp>CNCF also expanded its Kubernetes AI conformance work, adding requirements around Gateway API support, inference-aware routing, and disaggregated inference. Those are dry terms, but they point to a real operational headache: traditional load balancing assumes stateless traffic, while AI inference cares about cache reuse, prompt latency, and GPU memory efficiency.\u003C\u002Fp>\u003Cp>In other words, the stack is changing because the workload changed first.\u003C\u002Fp>\u003Ch2>The numbers behind the inference shift\u003C\u002Fh2>\u003Cp>The keynote leaned hard on one forecast: in 2023, about two-thirds of AI compute went to training and one-third to inference. By the end of 2026, that ratio is expected to flip. By decade’s end, inference demand was projected to reach 93.3 gigawatts of compute capacity.\u003C\u002Fp>\n\u003Cfigure class=\"my-6\">\u003Cimg src=\"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775178586492-pyjq.png\" alt=\"Kubernetes Is Becoming AI’s Control Plane\" class=\"rounded-xl w-full\" loading=\"lazy\" \u002F>\u003C\u002Ffigure>\n\u003Cp>That forecast should be treated as directional, not destiny. Still, the trend is easy to understand. Chatbots introduced people to AI. Agents will drive sustained usage, and that means far more repeated requests, more token generation, and much higher infrastructure load.\u003C\u002Fp>\u003Cp>For operators, this changes the economics. Training is a burst. Inference is the bill that keeps arriving.\u003C\u002Fp>\u003Cul>\u003Cli>2023: roughly 2\u002F3 of AI compute went to training\u003C\u002Fli>\u003Cli>2023: roughly 1\u002F3 of AI compute went to inference\u003C\u002Fli>\u003Cli>By end of 2026: inference is expected to overtake training\u003C\u002Fli>\u003Cli>By decade’s end: inference demand projected at 93.3 gigawatts\u003C\u002Fli>\u003C\u002Ful>\u003Cp>This is also why Kubernetes is being recast as a programmable control plane for AI infrastructure rather than a simple container scheduler. Once inference becomes the dominant steady-state workload, the winning platform is the one that can coordinate GPUs, route requests intelligently, and keep utilization high without turning every deployment into a custom project.\u003C\u002Fp>\u003Ch2>What operators can learn from Uber and the cloud-native stack\u003C\u002Fh2>\u003Cp>The Uber segment at KubeCon gave the keynote some much-needed reality. Uber said its Michelangelo platform supports 100% of mission-critical ML at the company, with 20,000 models trained per month, 5,300 in production, and more than 30 million peak predictions per second across roughly 1,000 serving nodes.\u003C\u002Fp>\u003Cp>Those are not small numbers, and they show why AI infrastructure has to be treated like production infrastructure from day one. The scale is not hypothetical. It already exists inside large companies that have spent years building internal platforms around model training, deployment, and serving.\u003C\u002Fp>\u003Cp>For teams trying to make sense of all this, the comparisons are useful because they show where the pressure sits:\u003C\u002Fp>\u003Cul>\u003Cli>\u003Ca href=\"https:\u002F\u002Fkubernetes.io\u002F\" target=\"_blank\" rel=\"noopener\">Kubernetes\u003C\u002Fa> already handles scheduling and orchestration across mixed workloads\u003C\u002Fli>\u003Cli>\u003Ca href=\"https:\u002F\u002Fwww.cncf.io\u002Fprojects\u002Fkyverno\u002F\" target=\"_blank\" rel=\"noopener\">Kyverno\u003C\u002Fa> and \u003Ca href=\"https:\u002F\u002Fwww.cncf.io\u002Fprojects\u002Ftekton\u002F\" target=\"_blank\" rel=\"noopener\">Tekton\u003C\u002Fa> help standardize policy and pipelines, which matters when models change often\u003C\u002Fli>\u003Cli>\u003Ca href=\"https:\u002F\u002Fwww.cncf.io\u002Fprojects\u002Fdragonfly\u002F\" target=\"_blank\" rel=\"noopener\">Dragonfly\u003C\u002Fa> helps with distribution patterns that become more important as model artifacts and inference traffic grow\u003C\u002Fli>\u003Cli>\u003Ca href=\"https:\u002F\u002Fwww.cncf.io\u002Fprojects\u002Ffluid\u002F\" target=\"_blank\" rel=\"noopener\">Fluid\u003C\u002Fa> addresses data access patterns that matter when AI jobs need fast, repeated reads\u003C\u002Fli>\u003C\u002Ful>\u003Cp>That mix is the real story. AI infrastructure does not need a brand-new operating model from scratch. It needs the cloud-native stack to stretch into GPU scheduling, inference routing, and policy enforcement without losing the portability that made Kubernetes useful in the first place.\u003C\u002Fp>\u003Cp>Europe’s role also matters here. CNCF said Europe is currently the largest regional contributor across CNCF projects, which fits the broader sovereignty conversation. If AI infrastructure is going to be regulated, audited, and deployed across national boundaries, open standards will matter more, not less.\u003C\u002Fp>\u003Ch2>Kubernetes is moving from app hosting to AI operations\u003C\u002Fh2>\u003Cp>The takeaway from KubeCon Europe 2026 is simple: the AI discussion has left the hype phase and entered the operations phase. The important questions now are about inference, GPU access, routing, and control, not just model quality or benchmark scores.\u003C\u002Fp>\u003Cp>That means platform teams should stop asking whether Kubernetes can support AI and start asking where the bottlenecks will appear first. Is it GPU allocation? Is it cache-aware routing? Is it policy drift across clusters? Those are the decisions that will define the next generation of AI infrastructure.\u003C\u002Fp>\u003Cp>If CNCF gets this right, Kubernetes becomes the standard control layer for enterprise AI. If it does not, vendors will fill the gap with closed systems that make portability harder and costs less predictable. My bet is that the open stack still has the edge, but only if it keeps standardizing around the messy reality of inference.\u003C\u002Fp>\u003Cp>For now, the actionable move is clear: treat inference as a production workload, not a pilot. The teams that design for routing, GPU economics, and policy today will have far fewer surprises when agents and specialized models start driving most of the traffic.\u003C\u002Fp>","KubeCon Europe 2026 showed Kubernetes moving from app orchestration to AI ops, with inference, GPUs, and open standards leading the shift.","www.efficientlyconnected.com","https:\u002F\u002Fwww.efficientlyconnected.com\u002Fkubecon-europe-2026-keynote-cloud-natives-ai-pivot-moves-from-hype-to-infrastructure\u002F",null,"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775178585987-q0fi.png",[13,14,15,16,17],"Kubernetes","AI infrastructure","inference","CNCF","GPU computing","en",0,false,"2026-04-03T01:09:30.778998+00:00","2026-04-03T01:09:30.754+00:00","done","a898f165-91d6-4e22-8f6e-042a9d516fdf","kubernetes-becoming-ais-control-plane-en","industry","b2f9469b-f74a-44b1-9e08-8b1539632542","published","2026-04-07T07:41:13.597+00:00",[31,32,34,36,38],{"name":15,"slug":15},{"name":13,"slug":33},"kubernetes",{"name":17,"slug":35},"gpu-computing",{"name":16,"slug":37},"cncf",{"name":14,"slug":39},"ai-infrastructure",{"id":27,"slug":41,"title":42,"language":43},"kubernetes-becoming-ais-control-plane-zh","Kubernetes 正在變成 AI 控制平面","zh",[45,51,57,63,69,75],{"id":46,"slug":47,"title":48,"cover_image":49,"image_url":49,"created_at":50,"category":26},"cf1863f5-624d-4b5f-bc32-d469c2149866","why-ai-infrastructure-is-now-the-real-moat-en","Why AI infrastructure is now the real moat","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778875858866-4ikl.png","2026-05-15T20:10:38.090619+00:00",{"id":52,"slug":53,"title":54,"cover_image":55,"image_url":55,"created_at":56,"category":26},"6ff3920d-c8ea-4cf3-8543-9cf9efc3fe36","circles-agent-stack-targets-machine-speed-payments-en","Circle’s Agent Stack targets machine-speed payments","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778871659638-hur1.png","2026-05-15T19:00:44.756112+00:00",{"id":58,"slug":59,"title":60,"cover_image":61,"image_url":61,"created_at":62,"category":26},"1270e2f4-6f3b-4772-9075-87c54b07a8d1","iren-signs-nvidia-ai-infrastructure-pact-en","IREN signs Nvidia AI infrastructure pact","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778871059665-3vhi.png","2026-05-15T18:50:38.162691+00:00",{"id":64,"slug":65,"title":66,"cover_image":67,"image_url":67,"created_at":68,"category":26},"b308c85e-ee9c-4de6-b702-dfad6d8da36f","circle-agent-stack-ai-payments-en","Circle launches Agent Stack for AI payments","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778870450891-zv1j.png","2026-05-15T18:40:31.462625+00:00",{"id":70,"slug":71,"title":72,"cover_image":73,"image_url":73,"created_at":74,"category":26},"f7028083-46ba-493b-a3db-dd6616a8c21f","why-nebius-ai-pivot-is-more-real-than-hype-en","Why Nebius’s AI Pivot Is More Real Than Hype","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778823055711-tbfv.png","2026-05-15T05:30:26.829489+00:00",{"id":76,"slug":77,"title":78,"cover_image":79,"image_url":79,"created_at":80,"category":26},"b63692ed-db6a-4dbd-b771-e1babdc94af7","nvidia-backs-corning-factories-with-billions-en","Nvidia backs Corning factories with billions","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778822444685-tvx6.png","2026-05-15T05:20:28.914908+00:00",[82,87,92,97,102,107,112,117,122,127],{"id":83,"slug":84,"title":85,"created_at":86},"d35a1bd9-e709-412e-a2df-392df1dc572a","ai-impact-2026-developments-market-en","AI's Impact in 2026: Key Developments and Market Shifts","2026-03-25T16:20:33.205823+00:00",{"id":88,"slug":89,"title":90,"created_at":91},"5ed27921-5fd6-492e-8c59-78393bf37710","trumps-ai-legislative-framework-en","Trump's AI Legislative Framework: What's Inside?","2026-03-25T16:22:20.005325+00:00",{"id":93,"slug":94,"title":95,"created_at":96},"e454a642-f03c-4794-b185-5f651aebbaca","nvidia-gtc-2026-key-highlights-innovations-en","NVIDIA GTC 2026: Key Highlights and Innovations","2026-03-25T16:22:47.882615+00:00",{"id":98,"slug":99,"title":100,"created_at":101},"0ebb5b16-774a-4922-945d-5f2ce1df5a6d","claude-usage-diversifies-learning-curves-en","Claude Usage Diversifies, Learning Curves Emerge","2026-03-25T16:25:50.770376+00:00",{"id":103,"slug":104,"title":105,"created_at":106},"69934e86-2fc5-4280-8223-7b917a48ace8","openclaw-ai-commoditization-concerns-en","OpenClaw's Rise Raises Concerns of AI Model Commoditization","2026-03-25T16:26:30.582047+00:00",{"id":108,"slug":109,"title":110,"created_at":111},"b4b2575b-2ac8-46b2-b90e-ab1d7c060797","google-gemini-ai-rollout-2026-en","Google's Gemini AI Rollout Extended to 2026","2026-03-25T16:28:14.808842+00:00",{"id":113,"slug":114,"title":115,"created_at":116},"6e18bc65-42ae-4ad0-b564-67d7f66b979e","meta-llama4-fabricated-results-scandal-en","Meta's Llama 4 Scandal: Fabricated AI Test Results Unveiled","2026-03-25T16:29:15.482836+00:00",{"id":118,"slug":119,"title":120,"created_at":121},"bf888e9d-08be-4f47-996c-7b24b5ab3500","accenture-mistral-ai-deployment-en","Accenture and Mistral AI Team Up for AI Deployment","2026-03-25T16:31:01.894655+00:00",{"id":123,"slug":124,"title":125,"created_at":126},"5382b536-fad2-49c6-ac85-9eb2bae49f35","mistral-ai-high-stakes-2026-en","Mistral AI: Facing High Stakes in 2026","2026-03-25T16:31:39.941974+00:00",{"id":128,"slug":129,"title":130,"created_at":131},"9da3d2d6-b669-4971-ba1d-17fdb3548ed5","cursors-meteoric-rise-pressures-en","Cursor's Meteoric Rise Faces Industry Pressures","2026-03-25T16:32:21.899217+00:00"]