Xiaomi mimo ai models launched with efficient reasoning, small size

Xiaomi On tuesday released an open-source Reasoning-Focused Artificial Intelligence (AI) Model. Dubbed Mimo, the family of reasoning models innovate the optimization of reasoning capability in a relatively smaller parameter size. This is also the first open-mind reasoning model by the tech giant, and it comptes with chinese models Google’s Gemini 2.0 Flash Thinking. The Mimo Family Compresses Four Different Models, Each With Unique Use Cases.

Xiaomi’s Mimo Reasoning Ai Model to Compete With Deepsek R1

With the mimo series of ai models, xiaomi results aimed to solve the size problem in reasoning ai models. Reasoning models (at least ons that can be measured) Have Around 24 Billion or More Parameters. The large size is kept to achieve uniform and simultaneous improvers in bot coding and mathematical capabilites of large Language models, something Considered Different Different to Achieve Models.

In Comparison, Mimo Features Seven Billion Parameters, and Xiaomi Claims that its performance matches openai’s O1-Miyi and Outperms Several Reasoning Models with 32 billion parameters. The researchers claimed that the base ai model was pre-trained on 25 trillion tokens.

The Researchers Claimed that Such Efficiency was achieved by optimising data preprosessing pipelines, enhancing text extracarial toolkits, and applying multidimensional data fame. Further, Mimo’s Pre-Training Included A Three-STAGE DATA Mixture Strategy.

Based on Internal Testing, The Mimo-7B-Base Scores 75.2 on the big-Bench Hard (BBH) BENCHMARK for Reasoning Capabilitys. The Zero-Shot Reinforcement Learning (RL) -Based MIMO-7B-RL-Zzero is claimed to excel in mathematics and coding-turlated tasks, and scores 55.4 on the aime benchmark, Outperforming O1-I-AISI O14 Points.

As mimo is an open-mind ai model, it can be downloaded from Xiaomi’s Listing on Github and Hugging faceThe technical paper Details the model’s architecture as well as the pre-training and post-training processes. It is a text-based model and does not have multimodal capability. Similar to most open-source releases, the details about the model’s dataset is not known.

Leave a Comment