[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"article-xiaomi-mimo-ai-agentic-software-push-en":3,"tags-xiaomi-mimo-ai-agentic-software-push-en":31,"related-lang-xiaomi-mimo-ai-agentic-software-push-en":42,"related-posts-xiaomi-mimo-ai-agentic-software-push-en":46,"series-industry-b8ea2224-db14-4d1d-a9df-60d6f9f499c7":83},{"id":4,"title":5,"content":6,"summary":7,"source":8,"source_url":9,"author":10,"image_url":11,"keywords":12,"language":18,"translated_content":10,"views":19,"is_premium":20,"created_at":21,"updated_at":21,"cover_image":22,"published_at":23,"rewrite_status":24,"rewrite_error":10,"rewritten_from_id":25,"slug":26,"category":27,"related_article_id":28,"status":29,"google_indexed_at":30,"x_posted_at":10,"tweet_text":10,"title_rewritten_at":10,"title_original":10,"key_takeaways":10,"topic_cluster_id":10,"embedding":10,"is_canonical_seed":20},"b8ea2224-db14-4d1d-a9df-60d6f9f499c7","Xiaomi’s MiMo AI Push Targets Agentic Software","\u003Cp>Xiaomi just put a very large number on the table: more than 1 trillion parameters for \u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo\" target=\"_blank\" rel=\"noopener\">MiMo-V2-Pro\u003C\u002Fa>. The bigger surprise is not the scale, though. It is the price, with input tokens listed at $1 per million and output tokens at $3 per million, which undercuts several premium frontier models by a wide margin.\u003C\u002Fp>\u003Cp>That matters because the AI race is moving from chatbots to agents that can click, read, reason, and act. Xiaomi’s new stack, which also includes \u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-Omni\" target=\"_blank\" rel=\"noopener\">MiMo-V2-Omni\u003C\u002Fa> and \u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-TTS\" target=\"_blank\" rel=\"noopener\">MiMo-V2-TTS\u003C\u002Fa>, is aimed at exactly that kind of software worker.\u003C\u002Fp>\u003Ch2>What Xiaomi actually launched\u003C\u002Fh2>\u003Cp>The launch is broader than a single model release. Xiaomi is packaging language reasoning, multimodal perception, and speech synthesis into one family of systems that can support agentic workflows across apps, browsers, and eventually physical devices.\u003C\u002Fp>\u003Cp>\u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo\" target=\"_blank\" rel=\"noopener\">MiMo-V2-Pro\u003C\u002Fa> is the headline model. It uses a Mixture-of-Experts design, claims more than 1 trillion total parameters, and activates 42 billion parameters per request. Xiaomi says it supports a context window of up to 1 million tokens, which is the sort of range you need for long-running agent tasks, codebases, or document-heavy enterprise work.\u003C\u002Fp>\u003Cp>\u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-Omni\" target=\"_blank\" rel=\"noopener\">MiMo-V2-Omni\u003C\u002Fa> extends that idea into text, image, audio, and video. \u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-TTS\" target=\"_blank\" rel=\"noopener\">MiMo-V2-TTS\u003C\u002Fa> handles speech generation with emotional control and nonverbal cues like laughter and hesitation.\u003C\u002Fp>\u003Cul>\u003Cli>MiMo-V2-Pro: 1T+ total parameters, 42B active per request\u003C\u002Fli>\u003Cli>Context window: up to 1 million tokens\u003C\u002Fli>\u003Cli>MiMo-V2-Pro pricing: $1 input, $3 output per million tokens\u003C\u002Fli>\u003Cli>MiMo-V2-Omni: text, vision, audio, and video in one model\u003C\u002Fli>\u003Cli>MiMo-V2-TTS: trained on more than 100 million hours of speech data\u003C\u002Fli>\u003C\u002Ful>\u003Ch2>Why agentic AI is the real story\u003C\u002Fh2>\u003Cp>The interesting shift here is not raw model size. It is the move toward agents that can do work instead of just answering prompts. That means planning a sequence, calling tools, checking results, correcting mistakes, and continuing without a human babysitting every step.\u003C\u002Fp>\u003Cp>That is the direction the whole field is moving in. OpenAI, \u003Ca href=\"https:\u002F\u002Fwww.anthropic.com\" target=\"_blank\" rel=\"noopener\">Anthropic\u003C\u002Fa>, and \u003Ca href=\"https:\u002F\u002Fdeepmind.google\" target=\"_blank\" rel=\"noopener\">Google DeepMind\u003C\u002Fa> have all been pushing models that can reason over longer contexts and interact with software. Xiaomi is now trying to win on a mix of scale, multimodality, and price.\u003C\u002Fp>\u003Cp>The company’s pitch is practical: let the model read a browser page, compare products, write a document, generate audio, and hand off the result. In other words, build AI that behaves more like a junior operator than a text box.\u003C\u002Fp>\u003Cblockquote>“AI is the new electricity.” — Andrew Ng\u003C\u002Fblockquote>\u003Cp>That quote gets used a lot because it still fits. Xiaomi is betting that AI will matter less as a standalone app and more as an invisible layer inside products, services, and workflows.\u003C\u002Fp>\u003Ch2>How MiMo-V2-Pro compares on price and scope\u003C\u002Fh2>\u003Cp>Pricing is where Xiaomi gets aggressive. The company is not trying to outspend the biggest labs. It is trying to make a model that developers can actually afford to run at scale.\u003C\u002Fp>\u003Cp>According to the launch details, MiMo-V2-Pro costs $1 per million input tokens and $3 per million output tokens. That is far below the pricing tier of several premium models that developers use for coding and agent tasks.\u003C\u002Fp>\u003Cp>Here is the comparison Xiaomi is implicitly making:\u003C\u002Fp>\u003Cul>\u003Cli>MiMo-V2-Pro: $1 input, $3 output per million tokens\u003C\u002Fli>\u003Cli>\u003Ca href=\"https:\u002F\u002Fwww.anthropic.com\u002Fclaude\" target=\"_blank\" rel=\"noopener\">Claude Sonnet\u003C\u002Fa>: $3 input, $15 output per million tokens\u003C\u002Fli>\u003Cli>\u003Ca href=\"https:\u002F\u002Fwww.anthropic.com\u002Fclaude\" target=\"_blank\" rel=\"noopener\">Claude Opus\u003C\u002Fa>: $5 input, $25 output per million tokens\u003C\u002Fli>\u003C\u002Ful>\u003Cp>That gap changes the economics of experimentation. A startup building an agent that runs dozens of model calls per task will care a lot more about token pricing than a demo user chatting once a day.\u003C\u002Fp>\u003Cp>Xiaomi also says MiMo-V2-Pro performs near the top tier on coding and agent benchmarks. The article source places it close to Claude Opus 4.6 in those tasks, which is a serious claim if the real-world performance holds up outside curated tests.\u003C\u002Fp>\u003Ch2>Omni and TTS make the stack more useful\u003C\u002Fh2>\u003Cp>\u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-Omni\" target=\"_blank\" rel=\"noopener\">MiMo-V2-Omni\u003C\u002Fa> is the part that makes Xiaomi’s plan feel less like a lab demo and more like a product strategy. A model that can process text, images, audio, and video can work across interfaces that humans already use every day.\u003C\u002Fp>\u003Cp>That opens up concrete use cases: checking dashcam footage for hazards, reading a browser page and filling out forms, separating speakers in a meeting recording, or scanning long video streams for important moments. The source article says Omni can handle continuous long-audio analysis beyond 10 hours, which is useful for support centers, meetings, and media review.\u003C\u002Fp>\u003Cp>\u003Ca href=\"https:\u002F\u002Fopenxlab.org.cn\u002Fmodels\u002Fdetail\u002FMiMo-V2-TTS\" target=\"_blank\" rel=\"noopener\">MiMo-V2-TTS\u003C\u002Fa> fills another gap. If agents are going to talk to people all day, the voice layer has to sound natural. Xiaomi says users can describe voice style in plain language, and the system can handle dialects, singing, laughter, and hesitation. That is a meaningful step beyond the fixed emotion presets most TTS tools still use.\u003C\u002Fp>\u003Cul>\u003Cli>Omni can analyze images, video, and long audio streams\u003C\u002Fli>\u003Cli>Omni supports native audio-video joint reasoning\u003C\u002Fli>\u003Cli>TTS adds emotional control through text instructions\u003C\u002Fli>\u003Cli>TTS includes dialects, tones, and paralinguistic sounds\u003C\u002Fli>\u003Cli>Combined, the stack targets software agents and voice agents\u003C\u002Fli>\u003C\u002Ful>\u003Cp>The interesting part is how these pieces fit together. A browser agent needs language reasoning. A customer support agent needs voice. A robotics system eventually needs both, plus perception from cameras and microphones. Xiaomi is building toward that chain.\u003C\u002Fp>\u003Ch2>What the market should watch next\u003C\u002Fh2>\u003Cp>Xiaomi’s move matters because it shows how fast the agent market is splitting into layers. Some companies will sell premium reasoning models. Others will sell cheaper models that developers can run often. Xiaomi is clearly aiming for the second group while still chasing top-tier capability.\u003C\u002Fp>\u003Cp>There is also a branding twist. The model first appeared anonymously on OpenRouter under the codename Hunter Alpha, and users speculated it was DeepSeek V4 before Xiaomi revealed the real source. That kind of mystery launch only happens when a model is good enough to create its own rumor mill.\u003C\u002Fp>\u003Cp>For developers, the practical question is simple: will MiMo integrate cleanly into agent frameworks and hold up under real workloads? Xiaomi says it is working with tools such as \u003Ca href=\"https:\u002F\u002Fgithub.com\u002Fcline\u002Fcline\" target=\"_blank\" rel=\"noopener\">Cline\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fwww.blackbox.ai\" target=\"_blank\" rel=\"noopener\">Blackbox AI\u003C\u002Fa>, and \u003Ca href=\"https:\u002F\u002Fgithub.com\u002FKilo-Org\u002Fkilocode\" target=\"_blank\" rel=\"noopener\">Kilo Code\u003C\u002Fa>, which suggests the company understands that distribution matters as much as raw model quality.\u003C\u002Fp>\u003Cp>My read: if MiMo-V2-Pro really holds near-frontier coding performance at these prices, the strongest adoption will come from teams building internal agents, not consumer chat apps. The next test is whether Xiaomi can turn that technical win into a developer habit. If it can, the company may become a much bigger AI player than its phone business alone would suggest.\u003C\u002Fp>","Xiaomi’s MiMo-V2-Pro, Omni, and TTS models pair 1T+ parameters with low pricing, aiming squarely at agentic AI workloads.","www.1950.ai","https:\u002F\u002Fwww.1950.ai\u002Fpost\u002Fxiaomi-enters-the-ai-power-league-with-mimo-v2-pro-and-omni-a-trillion-parameter-strategy-to-disrup",null,"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1774618930188-17in.png",[13,14,15,16,17],"Xiaomi","MiMo-V2-Pro","agentic AI","multimodal AI","speech synthesis","en",0,false,"2026-03-28T13:10:35.375168+00:00","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1774498042556-7ne9.png","2026-03-28T13:40:03.693+00:00","done","00e96d88-a556-4d57-88e6-4334e340a102","xiaomi-mimo-ai-agentic-software-push-en","industry","d57cf951-6f7c-4cac-b52b-79a62dd1d361","published","2026-04-09T09:00:58.555+00:00",[32,34,36,38,40],{"name":13,"slug":33},"xiaomi",{"name":16,"slug":35},"multimodal-ai",{"name":17,"slug":37},"speech-synthesis",{"name":14,"slug":39},"mimo-v2-pro",{"name":15,"slug":41},"agentic-ai",{"id":28,"slug":43,"title":44,"language":45},"xiaomi-mimo-ai-agentic-software-push-zh","小米 MiMo 盯上代理式軟體","zh",[47,53,59,65,71,77],{"id":48,"slug":49,"title":50,"cover_image":51,"image_url":51,"created_at":52,"category":27},"6ff3920d-c8ea-4cf3-8543-9cf9efc3fe36","circles-agent-stack-targets-machine-speed-payments-en","Circle’s Agent Stack targets machine-speed payments","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778871659638-hur1.png","2026-05-15T19:00:44.756112+00:00",{"id":54,"slug":55,"title":56,"cover_image":57,"image_url":57,"created_at":58,"category":27},"1270e2f4-6f3b-4772-9075-87c54b07a8d1","iren-signs-nvidia-ai-infrastructure-pact-en","IREN signs Nvidia AI infrastructure pact","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778871059665-3vhi.png","2026-05-15T18:50:38.162691+00:00",{"id":60,"slug":61,"title":62,"cover_image":63,"image_url":63,"created_at":64,"category":27},"b308c85e-ee9c-4de6-b702-dfad6d8da36f","circle-agent-stack-ai-payments-en","Circle launches Agent Stack for AI payments","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778870450891-zv1j.png","2026-05-15T18:40:31.462625+00:00",{"id":66,"slug":67,"title":68,"cover_image":69,"image_url":69,"created_at":70,"category":27},"f7028083-46ba-493b-a3db-dd6616a8c21f","why-nebius-ai-pivot-is-more-real-than-hype-en","Why Nebius’s AI Pivot Is More Real Than Hype","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778823055711-tbfv.png","2026-05-15T05:30:26.829489+00:00",{"id":72,"slug":73,"title":74,"cover_image":75,"image_url":75,"created_at":76,"category":27},"b63692ed-db6a-4dbd-b771-e1babdc94af7","nvidia-backs-corning-factories-with-billions-en","Nvidia backs Corning factories with billions","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778822444685-tvx6.png","2026-05-15T05:20:28.914908+00:00",{"id":78,"slug":79,"title":80,"cover_image":81,"image_url":81,"created_at":82,"category":27},"26ab4480-2476-4ec7-b43a-5d46def6487e","why-anthropic-gates-foundation-ai-public-goods-en","Why Anthropic and the Gates Foundation should fund AI public goods","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778796645685-wbw0.png","2026-05-14T22:10:22.60302+00:00",[84,89,94,99,104,109,114,119,124,129],{"id":85,"slug":86,"title":87,"created_at":88},"d35a1bd9-e709-412e-a2df-392df1dc572a","ai-impact-2026-developments-market-en","AI's Impact in 2026: Key Developments and Market Shifts","2026-03-25T16:20:33.205823+00:00",{"id":90,"slug":91,"title":92,"created_at":93},"5ed27921-5fd6-492e-8c59-78393bf37710","trumps-ai-legislative-framework-en","Trump's AI Legislative Framework: What's Inside?","2026-03-25T16:22:20.005325+00:00",{"id":95,"slug":96,"title":97,"created_at":98},"e454a642-f03c-4794-b185-5f651aebbaca","nvidia-gtc-2026-key-highlights-innovations-en","NVIDIA GTC 2026: Key Highlights and Innovations","2026-03-25T16:22:47.882615+00:00",{"id":100,"slug":101,"title":102,"created_at":103},"0ebb5b16-774a-4922-945d-5f2ce1df5a6d","claude-usage-diversifies-learning-curves-en","Claude Usage Diversifies, Learning Curves Emerge","2026-03-25T16:25:50.770376+00:00",{"id":105,"slug":106,"title":107,"created_at":108},"69934e86-2fc5-4280-8223-7b917a48ace8","openclaw-ai-commoditization-concerns-en","OpenClaw's Rise Raises Concerns of AI Model Commoditization","2026-03-25T16:26:30.582047+00:00",{"id":110,"slug":111,"title":112,"created_at":113},"b4b2575b-2ac8-46b2-b90e-ab1d7c060797","google-gemini-ai-rollout-2026-en","Google's Gemini AI Rollout Extended to 2026","2026-03-25T16:28:14.808842+00:00",{"id":115,"slug":116,"title":117,"created_at":118},"6e18bc65-42ae-4ad0-b564-67d7f66b979e","meta-llama4-fabricated-results-scandal-en","Meta's Llama 4 Scandal: Fabricated AI Test Results Unveiled","2026-03-25T16:29:15.482836+00:00",{"id":120,"slug":121,"title":122,"created_at":123},"bf888e9d-08be-4f47-996c-7b24b5ab3500","accenture-mistral-ai-deployment-en","Accenture and Mistral AI Team Up for AI Deployment","2026-03-25T16:31:01.894655+00:00",{"id":125,"slug":126,"title":127,"created_at":128},"5382b536-fad2-49c6-ac85-9eb2bae49f35","mistral-ai-high-stakes-2026-en","Mistral AI: Facing High Stakes in 2026","2026-03-25T16:31:39.941974+00:00",{"id":130,"slug":131,"title":132,"created_at":133},"9da3d2d6-b669-4971-ba1d-17fdb3548ed5","cursors-meteoric-rise-pressures-en","Cursor's Meteoric Rise Faces Industry Pressures","2026-03-25T16:32:21.899217+00:00"]