1. 📘 Topic and Domain: Development of MiniMax-M1, a large-scale hybrid-attention language model with efficient test-time compute scaling, in the domain of natural language processing and machine learning.
2. 💡 Previous Research and New Ideas: Based on MiniMax-Text-01 model and previous attention mechanisms; introduces novel lightning attention mechanism and CISPO reinforcement learning algorithm.
3. ❓ Problem: Addresses the challenge of efficiently scaling language models for extended reasoning processes and long-context understanding while maintaining computational efficiency.
4. 🛠️ Methods: Combines hybrid Mixture-of-Experts architecture with lightning attention mechanism, implements CISPO reinforcement learning algorithm, and uses diverse training data including mathematical reasoning, coding, and software engineering tasks.
5. 📊 Results and Evaluation: Achieves competitive performance against leading models like DeepSeek-R1 and Qwen3-235B, with particular strengths in software engineering, tool use, and long-context tasks; supports 1M token input length and 80K token generation length while using 25% of the FLOPs compared to other models.