Ted Hisokawa
Apr 12, 2026 01:37
MiniMax releases M2.7, a 230B-parameter mixture-of-experts mannequin optimized for NVIDIA GPUs with as much as 2.7x throughput positive aspects on Blackwell {hardware}.

MiniMax has launched M2.7, a 230-billion parameter open-weights AI mannequin designed particularly for autonomous agent workflows, now out there throughout NVIDIA’s inference ecosystem together with the corporate’s newest Blackwell Extremely GPUs.
The mannequin represents a big effectivity play in enterprise AI. Regardless of its large 230B whole parameters, M2.7 prompts solely 10B parameters per token—a 4.3% activation price achieved by mixture-of-experts (MoE) structure with 256 native specialists. This retains inference prices manageable whereas sustaining the reasoning capability of a a lot bigger mannequin.
Efficiency Numbers on Blackwell
NVIDIA collaborated with open supply communities to optimize M2.7 for manufacturing workloads. Two key optimizations—a fused QK RMS Norm kernel and FP8 MoE integration from TensorRT-LLM—delivered substantial throughput enhancements on Blackwell Extremely GPUs.
Testing with a 1K/1K enter/output sequence size dataset confirmed vLLM achieved as much as 2.5x throughput enchancment, whereas SGLang hit 2.7x positive aspects. Each optimizations have been applied inside a single month, suggesting additional efficiency headroom exists.
Technical Structure
M2.7 helps 200K enter context size throughout 62 layers, utilizing multi-head causal self-attention with Rotary Place Embeddings (RoPE). A top-k skilled routing mechanism prompts solely 8 of the 256 specialists for any given enter, which is how the mannequin maintains low inference prices regardless of its scale.
The structure targets coding challenges and sophisticated agentic duties—workflows the place AI programs have to plan, execute, and iterate autonomously moderately than reply to single prompts.
Deployment Choices
Builders can entry M2.7 by a number of channels. NVIDIA’s NemoClaw reference stack offers a one-click deployment for operating autonomous brokers with OpenShell runtime. The mannequin can also be out there by NVIDIA NIM containerized microservices for on-premise, cloud, or hybrid deployments.
For groups eager to customise the mannequin, NVIDIA’s NeMo AutoModel library helps fine-tuning with printed recipes. Reinforcement studying workflows can be found by NeMo RL with pattern configurations for 8K and 16K sequence lengths.
Free GPU-accelerated endpoints on construct.nvidia.com permit testing earlier than committing to infrastructure. The open weights are additionally out there on Hugging Face for self-hosted deployments.
The discharge positions MiniMax as a reputable various to closed fashions from OpenAI and Anthropic for enterprises constructing autonomous AI programs, significantly these already invested in NVIDIA infrastructure.
Picture supply: Shutterstock
