MiniMax, an AI firm based in Shanghai, has released an open-source reasoning model that challenges Chinese rival DeepSeek and US-based Anthropic, OpenAI, and Google in terms of performance and cost.
MiniMax-M1 was released Monday under an Apache software license, and thus is actually open source, unlike Meta's Llama family, offered under a community license that's not open source, and DeepSeek, which is only partially under an open-source license.
"In complex, productivity-oriented scenarios, M1's capabilities are top-tier among open-source models, surpassing domestic closed-source models and approaching the leading overseas models, all while offering the industry's best cost-effectiveness," MiniMax boasts in a blog post.
According to the blog post, M1 is competitive with OpenAI o3, Gemini 2.5 Pro, Claude 4 Opus, DeepSeek R1, DeepSeek R1-0528, and Qwen3-235B on various benchmarks (AIME 2024, LiveCodeBench, SWE-bench Verified, Tau-bench, and MRCR), coming in behind some models and ahead of others to varying degrees. As always, take vendor-supplied benchmark results with a grain of salt, but the source code is available on GitHub should you wish to confirm its performance independently.
But MiniMax makes clear that it's trying to supplant DeepSeek as the leading industry disruptor by noting that its context window (the amount of input it can handle) is one million tokens, which rivals Google Gemini 2.5 Pro and is eight times the capacity of DeepSeek R1.
Day 1/5 of #MiniMaxWeek: We're open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning.- World's longest context window: 1M-token input, 80k-token output- State-of-the-art agentic use among open-source models- RL at unmatched efficiency:… pic.twitter.com/bGfDlZA54n
— MiniMax (official) (@MiniMax__AI) June 16, 2025In terms of output, the model can manage 80,000 tokens, better than DeepSeek's 64,000 token capacity but shy of OpenAI's o3, which can spit out 100,000 tokens in response to a prompt.
- Bots are overwhelming websites with their hunger for AI data
- 'AI is not doing its job and should leave us alone' says Gartner's top analyst
- Defense Department signs OpenAI for $200 million 'frontier AI' pilot project
- UK students flock to AI to help them cheat
Backed by Alibaba Group, Tencent, and IDG Capital, MiniMax claims its Lightning Attention mechanism, a way to calculate attention matrices that improves both training and inference efficiency, gives its M1 model an advantage when computing long context inputs and when trying to reason.
"For example, when performing deep reasoning with 80,000 tokens, it requires only about 30 percent of the computing power of DeepSeek R1," the company claims. "This feature gives us a substantial computational efficiency advantage in both training and inference."
This more efficient computation method, in conjunction with an improved reinforcement learning algorithm called CISPO (detailed in M1's technical report [PDF]), translates to lower computing costs.
"The entire reinforcement learning phase used only 512 [Nvidia] H800s for three weeks, with a rental cost of just $537,400," MiniMax claims. "This is an order of magnitude less than initially anticipated." ®