Provides access to large language and vision-language models based on linear attention for various natural language and multimodal tasks.
MiniMax-01 is the official repository for MiniMax-Text-01, a large language model with 456 billion parameters, and MiniMax-VL-01, a vision-language model leveraging a ViT-MLP-LLM framework. MiniMax-Text-01 achieves long-context capabilities through a hybrid architecture integrating Lightning Attention, Softmax Attention, and Mixture-of-Experts (MoE), enabling it to handle up to 4 million tokens during inference. MiniMax-VL-01 extends this with enhanced visual capabilities and dynamic resolution mechanisms, achieving top-level performance on multimodal leaderboards.