This post covers the MiniMax M2.7 release (March 18, 2026). Sources: MiniMax official announcement, VentureBeat, MiniMax platform docs.
What’s New This Week
MiniMax released M2.7 on March 18, 2026 – a proprietary reasoning model that, by the company’s account, helped build itself. The self-evolving RL loop claim is getting most of the coverage. The proprietary shift is getting less. Both matter.
Changelog
| Date | Summary |
|---|---|
| 18 Mar 2026 | Initial post covering M2.7 release, self-evolving RL loop, and Chinese AI open-source strategy shift. |
MiniMax released M2.7 this week. Two things about it are worth your attention, and they’re not the same thing.
The first is the headline: earlier versions of M2.7 were used to build the reinforcement learning research harness that trained M2.7. The model handled 30-50% of its own development workflow – log-reading, debugging, metric analysis, code modification – autonomously, across iterative loops of 100 rounds or more. That’s a specific and unusual claim, and the technical detail behind it is more interesting than the marketing framing.
The second is quieter: a Chinese AI startup that built its reputation on open-source models just went proprietary. That’s a strategic signal worth tracking separately from the benchmark numbers.
What “self-evolving” actually means
The phrase “self-evolving AI” will trigger very different reactions depending on what you’ve been reading. Let’s be precise about what MiniMax actually did.
MiniMax used an internal version of M2.7 to build a research agent harness – the scaffolding that manages data pipelines, training environments, evaluation infrastructure, and cross-team coordination for RL experiments. Once built, that harness ran the daily RL research workflow: a researcher proposes an experiment, the agent handles literature review, data pipelines, experiment launch, monitoring, log-reading, debugging, code fixes, merge requests, and smoke tests. Human researchers stayed in the loop for critical decisions. The model handled the rest.
The figure they cite is 30-50% of the workflow automated. That’s the “self-evolving” claim in its operational form.
There’s a second, more specific example in the official announcement that’s worth quoting directly. MiniMax had M2.7 optimize a model’s programming performance on an internal scaffold. The model ran entirely autonomously, executing a loop of “analyze failure trajectories, plan changes, modify scaffold code, run evaluations, compare results, decide to keep or revert changes” for over 100 rounds. It discovered effective optimizations: searching for optimal sampling parameter combinations, designing workflow guidelines, adding loop detection. The result was a 30% performance improvement on internal evaluation sets.
This is not recursive self-improvement in the science-fiction sense. The model did not rewrite its own weights. It did something more mundane but arguably more immediately relevant: it automated the software engineering work required to improve the system it runs on. AI-assisted research applied to AI development itself.
MiniMax Head of Engineering Skyler Miao described the intent directly: “We intentionally trained the model to be better at planning and at clarifying requirements with the user. Next step is a more complex user simulator to push this even further.”
The direction of travel is clear. The current state is a model that handles a significant fraction of its own development pipeline. The goal is more.
Why automating 30-50% of RL research matters
Reinforcement learning research is expensive, slow, and highly iterative. The cycle is: run experiment, wait, read logs, identify failure modes, adjust hyperparameters or scaffold code, re-run. The bottleneck is rarely compute – it’s researcher time and attention. Each experimental cycle requires someone to read the logs, understand what went wrong, and decide what to change.
If a model can handle the log-reading, failure analysis, and code modification steps autonomously – and do it across 100+ rounds without human oversight – the economics of RL research shift. Researchers aren’t replaced; they move to higher-level direction. But the experimental cycle compresses significantly.
This matters beyond MiniMax. If the pattern holds, labs that deploy this kind of AI-assisted research infrastructure will iterate faster than those that don’t. The compounding effect over 12-18 months could be substantial. This is the “AI-assisted research” pattern that people have been predicting for years, and it’s starting to appear in production workflows rather than papers.
The open-source to proprietary shift
This is the bigger strategic signal, and it’s getting less coverage than the benchmark numbers.
Chinese AI startups spent 2024-2025 establishing themselves as the open-source leaders. DeepSeek, Qwen, and MiniMax’s own M2 series were the flagships: frontier-capable models released with open weights, free to download, free to fine-tune. The strategy worked. Enterprises adopted them for cost and customization reasons. Chinese labs were, as VentureBeat put it, “standard-bearers in the world of the open source AI frontier.”
M2.7 is proprietary. MiniMax becomes the second Chinese startup to close its frontier model in recent months, following z.ai with GLM-5 Turbo. Alibaba’s Qwen team is reportedly heading the same direction following leadership departures.
Why shift? A few plausible reasons. Open weights are expensive to maintain and distribute at scale. The business model for API revenue requires keeping the frontier model proprietary – you can’t charge for API access to a model anyone can run locally. And the competitive advantage from releasing open weights diminishes as everyone does it: if DeepSeek releases open weights, and Qwen releases open weights, and MiniMax releases open weights, the differentiation collapses.
The more cynical reading: the open-source positioning was partly a land-grab strategy. Release open models to drive adoption, build developer mindshare, establish the ecosystem – then close the frontier when the economics require it. That’s not unique to Chinese labs; it’s the same pattern Mistral has navigated, and to some degree Meta. But the pace of the shift in China is notable.
The question is whether this reverses China’s positioning as the open-source alternative to US proprietary models, or whether it’s a transitional phase. For the model landscape signal, see AI Model Landscape.
The cost and performance economics
M2.7’s pricing is $0.30 per million input tokens and $1.20 per million output tokens – unchanged from M2.5. At current Claude Sonnet pricing of $3/$15 per million, that’s a 10x cost difference on input and 12.5x on output.
For agentic workloads, this matters in ways that simple chat usage doesn’t. An agent running a coding task might make hundreds of model calls. At Claude Sonnet pricing, that adds up quickly. At M2.7 pricing, the same pipeline costs a fraction. For teams building cost-sensitive agent infrastructure – the kind of work discussed in The Agentic Turn – the economics of a capable model at $0.30/$1.20 are genuinely attractive.
The caveat is that pricing holds until it doesn’t. M2.7 is currently positioned to drive adoption. Whether those rates survive when MiniMax needs to convert adoption to revenue is a different question.
On speed, MiniMax offers a high-speed variant at higher subscription tiers. The base API pricing is unchanged from M2.5, so the performance gains in M2.7 come without a cost increase – if the benchmarks hold.
Benchmarks and what they tell you
The headline numbers: M2.7 scored 66.6% medal rate on MLE Bench Lite (OpenAI’s machine learning competition benchmark), tying with Gemini 3.1 and behind only Claude Opus 4.6 at 75.7% and GPT-5.4 at 71.2%. On SWE-Pro (software engineering across multiple languages), it scored 56.22%, matching GPT-5.3-Codex. Artificial Analysis placed it 8th globally on their Intelligence Index, an 8-point improvement over M2.5 in one month.
These are self-reported numbers from MiniMax, using benchmarks they selected. The Reddit community is sceptical – and reasonably so. One comment from the r/singularity thread is representative: “given their willingness to train from reasoning traces from proprietary models, it’d be forgivable for people to assume they use public benchmarks as RLVR environments.” Benchmarking against the same tasks you trained on produces numbers that don’t transfer to real workloads.
Not all the numbers are flattering. On BridgeBench – a third-party benchmark testing natural language to working code – M2.7 placed 19th, worse than M2.5’s 12th. That’s an independent result going the wrong direction, and it’s worth noting alongside the self-reported wins.
The honest position: the MLE Bench and SWE-Pro numbers are interesting, but independent evaluation at scale is pending. The only reliable test for your use case is dropping it into your actual pipeline and measuring task completion on your specific workloads. Given the pricing, the cost of that experiment is low.
What to watch
M2.7 is worth tracking for two separate reasons that happen to arrive in the same release.
The self-evolving RL loop is a genuine architectural shift in how models are built. If 30-50% of RL research workflow automation is real and reproducible, labs that deploy similar infrastructure will iterate faster than those that don’t. The compounding matters. This is early, but the direction is clear, and MiniMax’s stated goal of full autonomy in model training is a concrete roadmap, not a vague aspiration.
The proprietary shift is a strategic signal about the Chinese AI landscape. The open-source race that defined 2024-2025 may be giving way to something more commercially oriented. For engineers and teams making infrastructure decisions – which models to standardise on, which ecosystems to invest in – the stability of the open-source Chinese model ecosystem is now a live question in a way it wasn’t six months ago.
The pricing at $0.30/$1.20 makes M2.7 worth evaluating against current pipeline costs. The self-reported benchmarks warrant appropriate scepticism. The strategy shift is real regardless of where the benchmarks land.
For context on where M2.7 fits in the broader model landscape, see AI Model Landscape. For the Chinese open-source context specifically, see the Leanstral formal verification post and the Mistral Small 4 coverage.