# MiniMax M2 Model vLLM Deployment Guide We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to deploy the [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2) model. vLLM is a high-performance inference engine with excellent serving throughput, efficient and intelligent memory management, powerful batch request processing capabilities, and deeply optimized underlying performance. We recommend reviewing vLLM's official documentation to check hardware compatibility before deployment. ## System Requirements - OS: Linux - Python: 3.9 - 3.12 - GPU: - compute capability 7.0 or higher - Memory requirements: 220 GB for weights, 60 GB per 1M context tokens The following are recommended configurations; actual requirements should be adjusted based on your use case: - 4x 96GB GPUs: Supports context input of up to 400K tokens. - 8x 144GB GPUs: Supports context input of up to 3M tokens. ## Deployment with Python It is recommended to use a virtual environment (such as venv, conda, or uv) to avoid dependency conflicts. We recommend installing vLLM in a fresh Python environment: ```bash # Not yet released, please install nightly build uv pip install -U vllm \ --torch-backend=auto \ --extra-index-url https://wheels.vllm.ai/nightly # If released, install using uv uv pip install "vllm" --torch-backend=auto ``` Run the following command to start the vLLM server. vLLM will automatically download and cache the MiniMax-M2 model from Hugging Face. 4-GPU deployment command: ```bash SAFETENSORS_FAST_GPU=1 VLLM_USE_V1=0 vllm serve \ --model MiniMaxAI/MiniMax-M2 \ --trust-remote-code \ --enable-expert-parallel --tensor-parallel-size 4 \ --enable-auto-tool-choice --tool-call-parser minimax_m2 \ --reasoning-parser minimax_m2 ``` ## Testing Deployment After startup, you can test the vLLM OpenAI-compatible API with the following command: ```bash curl http://localhost:8000/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "MiniMaxAI/MiniMax-M2", "messages": [ {"role": "system", "content": [{"type": "text", "text": "You are a helpful assistant."}]}, {"role": "user", "content": [{"type": "text", "text": "Who won the world series in 2020?"}]} ] }' ``` ## Common Issues ### Hugging Face Network Issues If you encounter network issues, you can set up a proxy before pulling the model. ```bash export HF_ENDPOINT=https://hf-mirror.com ``` ### MiniMax-M2 model is not currently supported This vLLM version is outdated. Please upgrade to the latest version. ## Getting Support If you encounter any issues while deploying the MiniMax model: - Contact our technical support team through official channels such as email at api@minimaxi.com - Submit an issue on our [GitHub](https://github.com/MiniMax-AI) repository We continuously optimize the deployment experience for our models. Feedback is welcome!