Upload folder using huggingface_hub
Browse files- README.md +2 -2
- docs/function_call_guide.md +482 -0
- docs/function_call_guide_en.md +482 -0
- docs/vllm_deploy_guide.md +94 -0
- docs/vllm_deploy_guide_cn.md +91 -0
README.md
CHANGED
|
@@ -159,8 +159,8 @@ We look forward to your feedback and to collaborating with developers and resear
|
|
| 159 |
|
| 160 |
## How to Use
|
| 161 |
|
| 162 |
-
- **MiniMax Agent**: Our general agent product, built on MiniMax-M2, is now publicly available and free for a limited time: https://agent.
|
| 163 |
|
| 164 |
-
- **MiniMax Open Platform**: https://www.
|
| 165 |
|
| 166 |
- **MiniMax-M2 Model Weights**: The open-source model weights are available on Hugging Face: https://huggingface.co/MiniMaxAI/
|
|
|
|
| 159 |
|
| 160 |
## How to Use
|
| 161 |
|
| 162 |
+
- **MiniMax Agent**: Our general agent product, built on MiniMax-M2, is now publicly available and free for a limited time: https://agent.minimax.io/
|
| 163 |
|
| 164 |
+
- **MiniMax Open Platform**: https://www.minimax.io/
|
| 165 |
|
| 166 |
- **MiniMax-M2 Model Weights**: The open-source model weights are available on Hugging Face: https://huggingface.co/MiniMaxAI/
|
docs/function_call_guide.md
ADDED
|
@@ -0,0 +1,482 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MiniMax-M2 函数调用(Function Call)功能指南
|
| 2 |
+
|
| 3 |
+
## 简介
|
| 4 |
+
|
| 5 |
+
MiniMax-M2 模型支持函数调用功能,使模型能够识别何时需要调用外部函数,并以结构化格式输出函数调用参数。本文档详细介绍了如何使用 MiniMax-M2 的函数调用功能。
|
| 6 |
+
|
| 7 |
+
## 基础示例
|
| 8 |
+
|
| 9 |
+
以下 Python 脚本基于 OpenAI SDK 实现了一个天气查询函数的调用示例:
|
| 10 |
+
|
| 11 |
+
```python
|
| 12 |
+
from openai import OpenAI
|
| 13 |
+
import json
|
| 14 |
+
|
| 15 |
+
client = OpenAI(base_url="http://localhost:8000/v1", api_key="dummy")
|
| 16 |
+
|
| 17 |
+
def get_weather(location: str, unit: str):
|
| 18 |
+
return f"Getting the weather for {location} in {unit}..."
|
| 19 |
+
|
| 20 |
+
tool_functions = {"get_weather": get_weather}
|
| 21 |
+
|
| 22 |
+
tools = [{
|
| 23 |
+
"type": "function",
|
| 24 |
+
"function": {
|
| 25 |
+
"name": "get_weather",
|
| 26 |
+
"description": "Get the current weather in a given location",
|
| 27 |
+
"parameters": {
|
| 28 |
+
"type": "object",
|
| 29 |
+
"properties": {
|
| 30 |
+
"location": {"type": "string", "description": "City and state, e.g., 'San Francisco, CA'"},
|
| 31 |
+
"unit": {"type": "string", "enum": ["celsius", "fahrenheit"]}
|
| 32 |
+
},
|
| 33 |
+
"required": ["location", "unit"]
|
| 34 |
+
}
|
| 35 |
+
}
|
| 36 |
+
}]
|
| 37 |
+
|
| 38 |
+
response = client.chat.completions.create(
|
| 39 |
+
model=client.models.list().data[0].id,
|
| 40 |
+
messages=[{"role": "user", "content": "What's the weather like in San Francisco? use celsius."}],
|
| 41 |
+
tools=tools,
|
| 42 |
+
tool_choice="auto"
|
| 43 |
+
)
|
| 44 |
+
|
| 45 |
+
print(response)
|
| 46 |
+
|
| 47 |
+
tool_call = response.choices[0].message.tool_calls[0].function
|
| 48 |
+
print(f"Function called: {tool_call.name}")
|
| 49 |
+
print(f"Arguments: {tool_call.arguments}")
|
| 50 |
+
print(f"Result: {get_weather(**json.loads(tool_call.arguments))}")
|
| 51 |
+
```
|
| 52 |
+
|
| 53 |
+
**输出示例:**
|
| 54 |
+
```
|
| 55 |
+
Function called: get_weather
|
| 56 |
+
Arguments: {"location": "San Francisco, CA", "unit": "celsius"}
|
| 57 |
+
Result: Getting the weather for San Francisco, CA in celsius...
|
| 58 |
+
```
|
| 59 |
+
|
| 60 |
+
## 手动解析模型输出
|
| 61 |
+
|
| 62 |
+
如果您无法使用已支持 MiniMax-M2 的推理引擎的内置解析器,或者需要使用其他推理框架(如 transformers、TGI 等),可以使用以下方法手动解析模型的原始输出。这种方法需要您自己解析模型输出的 XML 标签格式。
|
| 63 |
+
|
| 64 |
+
### 使用 Transformers 的示例
|
| 65 |
+
|
| 66 |
+
以下是使用 transformers 库的完整示例:
|
| 67 |
+
|
| 68 |
+
```python
|
| 69 |
+
from transformers import AutoTokenizer
|
| 70 |
+
|
| 71 |
+
def get_default_tools():
|
| 72 |
+
return [
|
| 73 |
+
{
|
| 74 |
+
"name": "get_current_weather",
|
| 75 |
+
"description": "Get the latest weather for a location",
|
| 76 |
+
"parameters": {
|
| 77 |
+
"type": "object",
|
| 78 |
+
"properties": {
|
| 79 |
+
"location": {
|
| 80 |
+
"type": "string",
|
| 81 |
+
"description": "A certain city, such as Beijing, Shanghai"
|
| 82 |
+
}
|
| 83 |
+
},
|
| 84 |
+
}
|
| 85 |
+
"required": ["location"],
|
| 86 |
+
"type": "object"
|
| 87 |
+
}
|
| 88 |
+
]
|
| 89 |
+
|
| 90 |
+
# 加载模型和分词器
|
| 91 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 92 |
+
prompt = "What's the weather like in Shanghai today?"
|
| 93 |
+
messages = [
|
| 94 |
+
{"role": "system", "content": "You are a helpful assistant."},
|
| 95 |
+
{"role": "user", "content": prompt},
|
| 96 |
+
]
|
| 97 |
+
|
| 98 |
+
# 启用函数调用工具
|
| 99 |
+
tools = get_default_tools()
|
| 100 |
+
|
| 101 |
+
# 应用聊天模板,并加入工具定义
|
| 102 |
+
text = tokenizer.apply_chat_template(
|
| 103 |
+
messages,
|
| 104 |
+
tokenize=False,
|
| 105 |
+
add_generation_prompt=True,
|
| 106 |
+
tools=tools
|
| 107 |
+
)
|
| 108 |
+
|
| 109 |
+
# 发送请求(这里使用任何推理服务)
|
| 110 |
+
import requests
|
| 111 |
+
payload = {
|
| 112 |
+
"model": "MiniMaxAI/MiniMax-M2",
|
| 113 |
+
"prompt": text,
|
| 114 |
+
"max_tokens": 4096
|
| 115 |
+
}
|
| 116 |
+
response = requests.post(
|
| 117 |
+
"http://localhost:8000/v1/completions",
|
| 118 |
+
headers={"Content-Type": "application/json"},
|
| 119 |
+
json=payload,
|
| 120 |
+
stream=False,
|
| 121 |
+
)
|
| 122 |
+
|
| 123 |
+
# 模型输出需要手动解析
|
| 124 |
+
raw_output = response.json()["choices"][0]["text"]
|
| 125 |
+
print("原始输出:", raw_output)
|
| 126 |
+
|
| 127 |
+
# 使用下面的解析函数处理输出
|
| 128 |
+
function_calls = parse_tool_calls(raw_output, tools)
|
| 129 |
+
```
|
| 130 |
+
|
| 131 |
+
## 🛠️ 函数调用的定义
|
| 132 |
+
|
| 133 |
+
### 函数结构体
|
| 134 |
+
|
| 135 |
+
函数调用需要在请求体中定义 `tools` 字段,每个函数由以下部分组成:
|
| 136 |
+
|
| 137 |
+
```json
|
| 138 |
+
{
|
| 139 |
+
"tools": [
|
| 140 |
+
{
|
| 141 |
+
"name": "search_web",
|
| 142 |
+
"description": "搜索函数。",
|
| 143 |
+
"parameters": {
|
| 144 |
+
"properties": {
|
| 145 |
+
"query_list": {
|
| 146 |
+
"description": "进行搜索的关键词,列表元素个数为1。",
|
| 147 |
+
"items": { "type": "string" },
|
| 148 |
+
"type": "array"
|
| 149 |
+
},
|
| 150 |
+
"query_tag": {
|
| 151 |
+
"description": "query的分类",
|
| 152 |
+
"items": { "type": "string" },
|
| 153 |
+
"type": "array"
|
| 154 |
+
}
|
| 155 |
+
},
|
| 156 |
+
"required": [ "query_list", "query_tag" ],
|
| 157 |
+
"type": "object"
|
| 158 |
+
}
|
| 159 |
+
}
|
| 160 |
+
]
|
| 161 |
+
}
|
| 162 |
+
```
|
| 163 |
+
|
| 164 |
+
**字段说明:**
|
| 165 |
+
- `name`: 函数名称
|
| 166 |
+
- `description`: 函数功能描述
|
| 167 |
+
- `parameters`: 函数参数定义
|
| 168 |
+
- `properties`: 参数属性定义,key 是参数名,value 包含参数的详细描述
|
| 169 |
+
- `required`: 必填参数列表
|
| 170 |
+
- `type`: 参数类型(通常为 "object")
|
| 171 |
+
|
| 172 |
+
### 模型内部处理格式
|
| 173 |
+
|
| 174 |
+
在 MiniMax-M2 模型内部处理���,函数定义会被转换为特殊格式并拼接到输入文本中。以下是一个完整的示例:
|
| 175 |
+
|
| 176 |
+
```
|
| 177 |
+
]~!b[]~b]system
|
| 178 |
+
You are a helpful assistant.
|
| 179 |
+
|
| 180 |
+
# Tools
|
| 181 |
+
You may call one or more tools to assist with the user query.
|
| 182 |
+
Here are the tools available in JSONSchema format:
|
| 183 |
+
|
| 184 |
+
<tools>
|
| 185 |
+
<tool>{"name": "search_web", "description": "搜索函数。", "parameters": {"type": "object", "properties": {"query_list": {"type": "array", "items": {"type": "string"}, "description": "进行搜索的关键词,列表元素个数为1。"}, "query_tag": {"type": "array", "items": {"type": "string"}, "description": "query的分类"}}, "required": ["query_list", "query_tag"]}}</tool>
|
| 186 |
+
</tools>
|
| 187 |
+
|
| 188 |
+
When making tool calls, use XML format to invoke tools and pass parameters:
|
| 189 |
+
|
| 190 |
+
<minimax:tool_call>
|
| 191 |
+
<invoke name="tool-name-1">
|
| 192 |
+
<parameter name="param-key-1">param-value-1</parameter>
|
| 193 |
+
<parameter name="param-key-2">param-value-2</parameter>
|
| 194 |
+
...
|
| 195 |
+
</invoke>
|
| 196 |
+
[e~[
|
| 197 |
+
]~b]user
|
| 198 |
+
OpenAI 和 Gemini 的最近一次发布会都是什么时候?[e~[
|
| 199 |
+
]~b]ai
|
| 200 |
+
<think>
|
| 201 |
+
```
|
| 202 |
+
|
| 203 |
+
**格式说明:**
|
| 204 |
+
|
| 205 |
+
- `]~!b[]~b]system`: System 消息开始标记
|
| 206 |
+
- `[e~[`: 消息结束标记
|
| 207 |
+
- `]~b]user`: User 消息开始标记
|
| 208 |
+
- `]~b]ai`: Assistant 消息开始标记
|
| 209 |
+
- `]~b]tool`: Tool 结果消息开始标记
|
| 210 |
+
- `<tools>...</tools>`: 工具定义区域,每个工具用 `<tool>` 标签包裹,内容为 JSON Schema
|
| 211 |
+
- `<minimax:tool_call>...</minimax:tool_call>`: 工具调用区域
|
| 212 |
+
- `<think>`: 生成时的思考过程标记(可选)
|
| 213 |
+
|
| 214 |
+
### 模型输出格式
|
| 215 |
+
|
| 216 |
+
MiniMax-M2使用结构化的 XML 标签格式:
|
| 217 |
+
|
| 218 |
+
```xml
|
| 219 |
+
<minimax:tool_call>
|
| 220 |
+
<invoke name="search_web">
|
| 221 |
+
<parameter name="query_tag">["technology", "events"]</parameter>
|
| 222 |
+
<parameter name="query_list">["\"OpenAI\" \"latest\" \"release\""]</parameter>
|
| 223 |
+
</invoke>
|
| 224 |
+
<invoke name="search_web">
|
| 225 |
+
<parameter name="query_tag">["technology", "events"]</parameter>
|
| 226 |
+
<parameter name="query_list">["\"Gemini\" \"latest\" \"release\""]</parameter>
|
| 227 |
+
</invoke>
|
| 228 |
+
</minimax:tool_call>
|
| 229 |
+
```
|
| 230 |
+
|
| 231 |
+
每个函数调用使用 `<invoke name="函数名">` 标签,参数使用 `<parameter name="参数名">` 标签包裹。
|
| 232 |
+
|
| 233 |
+
## 手动解析函数调用结果
|
| 234 |
+
|
| 235 |
+
### 解析函数调用
|
| 236 |
+
|
| 237 |
+
MiniMax-M2使用结构化的 XML 标签,需要不同的解析方式。核心函数如下:
|
| 238 |
+
|
| 239 |
+
```python
|
| 240 |
+
import re
|
| 241 |
+
import json
|
| 242 |
+
from typing import Any, Optional, List, Dict
|
| 243 |
+
|
| 244 |
+
|
| 245 |
+
def extract_name(name_str: str) -> str:
|
| 246 |
+
"""从引号包裹的字符串中提取名称"""
|
| 247 |
+
name_str = name_str.strip()
|
| 248 |
+
if name_str.startswith('"') and name_str.endswith('"'):
|
| 249 |
+
return name_str[1:-1]
|
| 250 |
+
elif name_str.startswith("'") and name_str.endswith("'"):
|
| 251 |
+
return name_str[1:-1]
|
| 252 |
+
return name_str
|
| 253 |
+
|
| 254 |
+
|
| 255 |
+
def convert_param_value(value: str, param_type: str) -> Any:
|
| 256 |
+
"""根据参数类型转换参数值"""
|
| 257 |
+
if value.lower() == "null":
|
| 258 |
+
return None
|
| 259 |
+
|
| 260 |
+
param_type = param_type.lower()
|
| 261 |
+
|
| 262 |
+
if param_type in ["string", "str", "text"]:
|
| 263 |
+
return value
|
| 264 |
+
elif param_type in ["integer", "int"]:
|
| 265 |
+
try:
|
| 266 |
+
return int(value)
|
| 267 |
+
except (ValueError, TypeError):
|
| 268 |
+
return value
|
| 269 |
+
elif param_type in ["number", "float"]:
|
| 270 |
+
try:
|
| 271 |
+
val = float(value)
|
| 272 |
+
return val if val != int(val) else int(val)
|
| 273 |
+
except (ValueError, TypeError):
|
| 274 |
+
return value
|
| 275 |
+
elif param_type in ["boolean", "bool"]:
|
| 276 |
+
return value.lower() in ["true", "1"]
|
| 277 |
+
elif param_type in ["object", "array"]:
|
| 278 |
+
try:
|
| 279 |
+
return json.loads(value)
|
| 280 |
+
except json.JSONDecodeError:
|
| 281 |
+
return value
|
| 282 |
+
else:
|
| 283 |
+
# 尝试 JSON 解析,失败则返回字符串
|
| 284 |
+
try:
|
| 285 |
+
return json.loads(value)
|
| 286 |
+
except json.JSONDecodeError:
|
| 287 |
+
return value
|
| 288 |
+
|
| 289 |
+
|
| 290 |
+
def parse_tool_calls(model_output: str, tools: Optional[List[Dict]] = None) -> List[Dict]:
|
| 291 |
+
"""
|
| 292 |
+
从模型输出中提取所有工具调用
|
| 293 |
+
|
| 294 |
+
Args:
|
| 295 |
+
model_output: 模型的完整输出文本
|
| 296 |
+
tools: 工具定义列表,用于获取参数类型信息,格式可以是:
|
| 297 |
+
- [{"name": "...", "parameters": {...}}]
|
| 298 |
+
- [{"type": "function", "function": {"name": "...", "parameters": {...}}}]
|
| 299 |
+
|
| 300 |
+
Returns:
|
| 301 |
+
解析后的工具调用列表,每个元素包含 name 和 arguments 字段
|
| 302 |
+
|
| 303 |
+
Example:
|
| 304 |
+
>>> tools = [{
|
| 305 |
+
... "name": "get_weather",
|
| 306 |
+
... "parameters": {
|
| 307 |
+
... "type": "object",
|
| 308 |
+
... "properties": {
|
| 309 |
+
... "location": {"type": "string"},
|
| 310 |
+
... "unit": {"type": "string"}
|
| 311 |
+
... }
|
| 312 |
+
... }
|
| 313 |
+
... }]
|
| 314 |
+
>>> output = '''<minimax:tool_call>
|
| 315 |
+
... <invoke name="get_weather">
|
| 316 |
+
... <parameter name="location">San Francisco</parameter>
|
| 317 |
+
... <parameter name="unit">celsius</parameter>
|
| 318 |
+
... </invoke>
|
| 319 |
+
... </minimax:tool_call>'''
|
| 320 |
+
>>> result = parse_tool_calls(output, tools)
|
| 321 |
+
>>> print(result)
|
| 322 |
+
[{'name': 'get_weather', 'arguments': {'location': 'San Francisco', 'unit': 'celsius'}}]
|
| 323 |
+
"""
|
| 324 |
+
# 快速检查是否包含工具调用标记
|
| 325 |
+
if "<minimax:tool_call>" not in model_output:
|
| 326 |
+
return []
|
| 327 |
+
|
| 328 |
+
tool_calls = []
|
| 329 |
+
|
| 330 |
+
try:
|
| 331 |
+
# 匹配所有 <minimax:tool_call> 块
|
| 332 |
+
tool_call_regex = re.compile(r"<minimax:tool_call>(.*?)</minimax:tool_call>", re.DOTALL)
|
| 333 |
+
invoke_regex = re.compile(r"<invoke name=(.*?)</invoke>", re.DOTALL)
|
| 334 |
+
parameter_regex = re.compile(r"<parameter name=(.*?)</parameter>", re.DOTALL)
|
| 335 |
+
|
| 336 |
+
# 遍历所有 tool_call 块
|
| 337 |
+
for tool_call_match in tool_call_regex.findall(model_output):
|
| 338 |
+
# 遍历该块中的所有 invoke
|
| 339 |
+
for invoke_match in invoke_regex.findall(tool_call_match):
|
| 340 |
+
# 提取函数名
|
| 341 |
+
name_match = re.search(r'^([^>]+)', invoke_match)
|
| 342 |
+
if not name_match:
|
| 343 |
+
continue
|
| 344 |
+
|
| 345 |
+
function_name = extract_name(name_match.group(1))
|
| 346 |
+
|
| 347 |
+
# 获取参数配置
|
| 348 |
+
param_config = {}
|
| 349 |
+
if tools:
|
| 350 |
+
for tool in tools:
|
| 351 |
+
tool_name = tool.get("name") or tool.get("function", {}).get("name")
|
| 352 |
+
if tool_name == function_name:
|
| 353 |
+
params = tool.get("parameters") or tool.get("function", {}).get("parameters")
|
| 354 |
+
if isinstance(params, dict) and "properties" in params:
|
| 355 |
+
param_config = params["properties"]
|
| 356 |
+
break
|
| 357 |
+
|
| 358 |
+
# 提取参数
|
| 359 |
+
param_dict = {}
|
| 360 |
+
for match in parameter_regex.findall(invoke_match):
|
| 361 |
+
param_match = re.search(r'^([^>]+)>(.*)', match, re.DOTALL)
|
| 362 |
+
if param_match:
|
| 363 |
+
param_name = extract_name(param_match.group(1))
|
| 364 |
+
param_value = param_match.group(2).strip()
|
| 365 |
+
|
| 366 |
+
# 去除首尾的换行符
|
| 367 |
+
if param_value.startswith('\n'):
|
| 368 |
+
param_value = param_value[1:]
|
| 369 |
+
if param_value.endswith('\n'):
|
| 370 |
+
param_value = param_value[:-1]
|
| 371 |
+
|
| 372 |
+
# 获取参数类型并转换
|
| 373 |
+
param_type = "string"
|
| 374 |
+
if param_name in param_config:
|
| 375 |
+
if isinstance(param_config[param_name], dict) and "type" in param_config[param_name]:
|
| 376 |
+
param_type = param_config[param_name]["type"]
|
| 377 |
+
|
| 378 |
+
param_dict[param_name] = convert_param_value(param_value, param_type)
|
| 379 |
+
|
| 380 |
+
tool_calls.append({
|
| 381 |
+
"name": function_name,
|
| 382 |
+
"arguments": param_dict
|
| 383 |
+
})
|
| 384 |
+
|
| 385 |
+
except Exception as e:
|
| 386 |
+
print(f"解析工具调用失败: {e}")
|
| 387 |
+
return []
|
| 388 |
+
|
| 389 |
+
return tool_calls
|
| 390 |
+
```
|
| 391 |
+
|
| 392 |
+
**使用示例:**
|
| 393 |
+
|
| 394 |
+
```python
|
| 395 |
+
# 定义工具
|
| 396 |
+
tools = [
|
| 397 |
+
{
|
| 398 |
+
"name": "get_weather",
|
| 399 |
+
"parameters": {
|
| 400 |
+
"type": "object",
|
| 401 |
+
"properties": {
|
| 402 |
+
"location": {"type": "string"},
|
| 403 |
+
"unit": {"type": "string"}
|
| 404 |
+
},
|
| 405 |
+
"required": ["location", "unit"]
|
| 406 |
+
}
|
| 407 |
+
}
|
| 408 |
+
]
|
| 409 |
+
|
| 410 |
+
# 模型输出
|
| 411 |
+
model_output = """我来帮你查询天气。
|
| 412 |
+
<minimax:tool_call>
|
| 413 |
+
<invoke name="get_weather">
|
| 414 |
+
<parameter name="location">San Francisco</parameter>
|
| 415 |
+
<parameter name="unit">celsius</parameter>
|
| 416 |
+
</invoke>
|
| 417 |
+
</minimax:tool_call>"""
|
| 418 |
+
|
| 419 |
+
# 解析工具调用
|
| 420 |
+
tool_calls = parse_tool_calls(model_output, tools)
|
| 421 |
+
|
| 422 |
+
# 输出结果
|
| 423 |
+
for call in tool_calls:
|
| 424 |
+
print(f"调用函数: {call['name']}")
|
| 425 |
+
print(f"参数: {call['arguments']}")
|
| 426 |
+
# 输出: 调用函数: get_weather
|
| 427 |
+
# 参数: {'location': 'San Francisco', 'unit': 'celsius'}
|
| 428 |
+
```
|
| 429 |
+
|
| 430 |
+
### 执行函数调用
|
| 431 |
+
|
| 432 |
+
解析完成后,您可以执行对应的函数并构建返回结果:
|
| 433 |
+
|
| 434 |
+
```python
|
| 435 |
+
def execute_function_call(function_name: str, arguments: dict):
|
| 436 |
+
"""执行函数调用并返回结果"""
|
| 437 |
+
if function_name == "get_weather":
|
| 438 |
+
location = arguments.get("location", "未知位置")
|
| 439 |
+
unit = arguments.get("unit", "celsius")
|
| 440 |
+
# 构建函数执行结果
|
| 441 |
+
return {
|
| 442 |
+
"role": "tool",
|
| 443 |
+
"content": [
|
| 444 |
+
{
|
| 445 |
+
"name": function_name,
|
| 446 |
+
"type": "text",
|
| 447 |
+
"text": json.dumps({
|
| 448 |
+
"location": location,
|
| 449 |
+
"temperature": "25",
|
| 450 |
+
"unit": unit,
|
| 451 |
+
"weather": "晴朗"
|
| 452 |
+
}, ensure_ascii=False)
|
| 453 |
+
}
|
| 454 |
+
]
|
| 455 |
+
}
|
| 456 |
+
elif function_name == "search_web":
|
| 457 |
+
query_list = arguments.get("query_list", [])
|
| 458 |
+
query_tag = arguments.get("query_tag", [])
|
| 459 |
+
# 模拟搜索结果
|
| 460 |
+
return {
|
| 461 |
+
"role": "tool",
|
| 462 |
+
"content": [
|
| 463 |
+
{
|
| 464 |
+
"name": function_name,
|
| 465 |
+
"type": "text",
|
| 466 |
+
"text": f"搜索关键词: {query_list}, 分类: {query_tag}\n搜索结果: 相关信息已找到"
|
| 467 |
+
}
|
| 468 |
+
]
|
| 469 |
+
}
|
| 470 |
+
|
| 471 |
+
return None
|
| 472 |
+
```
|
| 473 |
+
|
| 474 |
+
### 将函数执行结果返回给模型
|
| 475 |
+
|
| 476 |
+
成功解析函数调用后,您应将函数执行结果添加到对话历史中,以便模型在后续交互中能够访问和利用这些信息,拼接格式参考chat_template.jinja
|
| 477 |
+
|
| 478 |
+
## 参考资料
|
| 479 |
+
|
| 480 |
+
- [MiniMax-M2 模型仓库](https://github.com/MiniMaxAI/MiniMax-M2)
|
| 481 |
+
- [vLLM 项目主页](https://github.com/vllm-project/vllm)
|
| 482 |
+
- [OpenAI Python SDK](https://github.com/openai/openai-python)
|
docs/function_call_guide_en.md
ADDED
|
@@ -0,0 +1,482 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MiniMax-M2 Function Call Guide
|
| 2 |
+
|
| 3 |
+
## Introduction
|
| 4 |
+
|
| 5 |
+
The MiniMax-M2 model supports function calling capabilities, enabling the model to identify when external functions need to be called and output function call parameters in a structured format. This document provides detailed instructions on how to use the function calling features of MiniMax-M2.
|
| 6 |
+
|
| 7 |
+
## Basic Example
|
| 8 |
+
|
| 9 |
+
The following Python script implements a weather query function call example based on the OpenAI SDK:
|
| 10 |
+
|
| 11 |
+
```python
|
| 12 |
+
from openai import OpenAI
|
| 13 |
+
import json
|
| 14 |
+
|
| 15 |
+
client = OpenAI(base_url="http://localhost:8000/v1", api_key="dummy")
|
| 16 |
+
|
| 17 |
+
def get_weather(location: str, unit: str):
|
| 18 |
+
return f"Getting the weather for {location} in {unit}..."
|
| 19 |
+
|
| 20 |
+
tool_functions = {"get_weather": get_weather}
|
| 21 |
+
|
| 22 |
+
tools = [{
|
| 23 |
+
"type": "function",
|
| 24 |
+
"function": {
|
| 25 |
+
"name": "get_weather",
|
| 26 |
+
"description": "Get the current weather in a given location",
|
| 27 |
+
"parameters": {
|
| 28 |
+
"type": "object",
|
| 29 |
+
"properties": {
|
| 30 |
+
"location": {"type": "string", "description": "City and state, e.g., 'San Francisco, CA'"},
|
| 31 |
+
"unit": {"type": "string", "enum": ["celsius", "fahrenheit"]}
|
| 32 |
+
},
|
| 33 |
+
"required": ["location", "unit"]
|
| 34 |
+
}
|
| 35 |
+
}
|
| 36 |
+
}]
|
| 37 |
+
|
| 38 |
+
response = client.chat.completions.create(
|
| 39 |
+
model=client.models.list().data[0].id,
|
| 40 |
+
messages=[{"role": "user", "content": "What's the weather like in San Francisco? use celsius."}],
|
| 41 |
+
tools=tools,
|
| 42 |
+
tool_choice="auto"
|
| 43 |
+
)
|
| 44 |
+
|
| 45 |
+
print(response)
|
| 46 |
+
|
| 47 |
+
tool_call = response.choices[0].message.tool_calls[0].function
|
| 48 |
+
print(f"Function called: {tool_call.name}")
|
| 49 |
+
print(f"Arguments: {tool_call.arguments}")
|
| 50 |
+
print(f"Result: {get_weather(**json.loads(tool_call.arguments))}")
|
| 51 |
+
```
|
| 52 |
+
|
| 53 |
+
**Output Example:**
|
| 54 |
+
```
|
| 55 |
+
Function called: get_weather
|
| 56 |
+
Arguments: {"location": "San Francisco, CA", "unit": "celsius"}
|
| 57 |
+
Result: Getting the weather for San Francisco, CA in celsius...
|
| 58 |
+
```
|
| 59 |
+
|
| 60 |
+
## Manually Parsing Model Output
|
| 61 |
+
|
| 62 |
+
If you cannot use the built-in parser of inference engines that support MiniMax-M2, or need to use other inference frameworks (such as transformers, TGI, etc.), you can manually parse the model's raw output using the following method. This approach requires you to parse the XML tag format of the model output yourself.
|
| 63 |
+
|
| 64 |
+
### Example Using Transformers
|
| 65 |
+
|
| 66 |
+
Here is a complete example using the transformers library:
|
| 67 |
+
|
| 68 |
+
```python
|
| 69 |
+
from transformers import AutoTokenizer
|
| 70 |
+
|
| 71 |
+
def get_default_tools():
|
| 72 |
+
return [
|
| 73 |
+
{
|
| 74 |
+
"name": "get_current_weather",
|
| 75 |
+
"description": "Get the latest weather for a location",
|
| 76 |
+
"parameters": {
|
| 77 |
+
"type": "object",
|
| 78 |
+
"properties": {
|
| 79 |
+
"location": {
|
| 80 |
+
"type": "string",
|
| 81 |
+
"description": "A certain city, such as Beijing, Shanghai"
|
| 82 |
+
}
|
| 83 |
+
},
|
| 84 |
+
}
|
| 85 |
+
"required": ["location"],
|
| 86 |
+
"type": "object"
|
| 87 |
+
}
|
| 88 |
+
]
|
| 89 |
+
|
| 90 |
+
# Load model and tokenizer
|
| 91 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 92 |
+
prompt = "What's the weather like in Shanghai today?"
|
| 93 |
+
messages = [
|
| 94 |
+
{"role": "system", "content": "You are a helpful assistant."},
|
| 95 |
+
{"role": "user", "content": prompt},
|
| 96 |
+
]
|
| 97 |
+
|
| 98 |
+
# Enable function calling tools
|
| 99 |
+
tools = get_default_tools()
|
| 100 |
+
|
| 101 |
+
# Apply chat template and include tool definitions
|
| 102 |
+
text = tokenizer.apply_chat_template(
|
| 103 |
+
messages,
|
| 104 |
+
tokenize=False,
|
| 105 |
+
add_generation_prompt=True,
|
| 106 |
+
tools=tools
|
| 107 |
+
)
|
| 108 |
+
|
| 109 |
+
# Send request (using any inference service)
|
| 110 |
+
import requests
|
| 111 |
+
payload = {
|
| 112 |
+
"model": "MiniMaxAI/MiniMax-M2",
|
| 113 |
+
"prompt": text,
|
| 114 |
+
"max_tokens": 4096
|
| 115 |
+
}
|
| 116 |
+
response = requests.post(
|
| 117 |
+
"http://localhost:8000/v1/completions",
|
| 118 |
+
headers={"Content-Type": "application/json"},
|
| 119 |
+
json=payload,
|
| 120 |
+
stream=False,
|
| 121 |
+
)
|
| 122 |
+
|
| 123 |
+
# Model output needs manual parsing
|
| 124 |
+
raw_output = response.json()["choices"][0]["text"]
|
| 125 |
+
print("Raw output:", raw_output)
|
| 126 |
+
|
| 127 |
+
# Use the parsing function below to process the output
|
| 128 |
+
function_calls = parse_tool_calls(raw_output, tools)
|
| 129 |
+
```
|
| 130 |
+
|
| 131 |
+
## 🛠️ Function Call Definition
|
| 132 |
+
|
| 133 |
+
### Function Structure
|
| 134 |
+
|
| 135 |
+
Function calls need to define the `tools` field in the request body. Each function consists of the following parts:
|
| 136 |
+
|
| 137 |
+
```json
|
| 138 |
+
{
|
| 139 |
+
"tools": [
|
| 140 |
+
{
|
| 141 |
+
"name": "search_web",
|
| 142 |
+
"description": "Search function.",
|
| 143 |
+
"parameters": {
|
| 144 |
+
"properties": {
|
| 145 |
+
"query_list": {
|
| 146 |
+
"description": "Keywords for search, list should contain 1 element.",
|
| 147 |
+
"items": { "type": "string" },
|
| 148 |
+
"type": "array"
|
| 149 |
+
},
|
| 150 |
+
"query_tag": {
|
| 151 |
+
"description": "Category of query",
|
| 152 |
+
"items": { "type": "string" },
|
| 153 |
+
"type": "array"
|
| 154 |
+
}
|
| 155 |
+
},
|
| 156 |
+
"required": [ "query_list", "query_tag" ],
|
| 157 |
+
"type": "object"
|
| 158 |
+
}
|
| 159 |
+
}
|
| 160 |
+
]
|
| 161 |
+
}
|
| 162 |
+
```
|
| 163 |
+
|
| 164 |
+
**Field Descriptions:**
|
| 165 |
+
- `name`: Function name
|
| 166 |
+
- `description`: Function description
|
| 167 |
+
- `parameters`: Function parameter definition
|
| 168 |
+
- `properties`: Parameter property definition, where key is the parameter name and value contains detailed parameter description
|
| 169 |
+
- `required`: List of required parameters
|
| 170 |
+
- `type`: Parameter type (usually "object")
|
| 171 |
+
|
| 172 |
+
### Internal Processing Format
|
| 173 |
+
|
| 174 |
+
When processing within the MiniMax-M2 model, function definitions are converted to a special format and concatenated to the input text. Here is a complete example:
|
| 175 |
+
|
| 176 |
+
```
|
| 177 |
+
]~!b[]~b]system
|
| 178 |
+
You are a helpful assistant.
|
| 179 |
+
|
| 180 |
+
# Tools
|
| 181 |
+
You may call one or more tools to assist with the user query.
|
| 182 |
+
Here are the tools available in JSONSchema format:
|
| 183 |
+
|
| 184 |
+
<tools>
|
| 185 |
+
<tool>{"name": "search_web", "description": "Search function.", "parameters": {"type": "object", "properties": {"query_list": {"type": "array", "items": {"type": "string"}, "description": "Keywords for search, list should contain 1 element."}, "query_tag": {"type": "array", "items": {"type": "string"}, "description": "Category of query"}}, "required": ["query_list", "query_tag"]}}</tool>
|
| 186 |
+
</tools>
|
| 187 |
+
|
| 188 |
+
When making tool calls, use XML format to invoke tools and pass parameters:
|
| 189 |
+
|
| 190 |
+
<minimax:tool_call>
|
| 191 |
+
<invoke name="tool-name-1">
|
| 192 |
+
<parameter name="param-key-1">param-value-1</parameter>
|
| 193 |
+
<parameter name="param-key-2">param-value-2</parameter>
|
| 194 |
+
...
|
| 195 |
+
</invoke>
|
| 196 |
+
[e~[
|
| 197 |
+
]~b]user
|
| 198 |
+
When were the latest announcements from OpenAI and Gemini?[e~[
|
| 199 |
+
]~b]ai
|
| 200 |
+
<think>
|
| 201 |
+
```
|
| 202 |
+
|
| 203 |
+
**Format Description:**
|
| 204 |
+
|
| 205 |
+
- `]~!b[]~b]system`: System message start marker
|
| 206 |
+
- `[e~[`: Message end marker
|
| 207 |
+
- `]~b]user`: User message start marker
|
| 208 |
+
- `]~b]ai`: Assistant message start marker
|
| 209 |
+
- `]~b]tool`: Tool result message start marker
|
| 210 |
+
- `<tools>...</tools>`: Tool definition area, each tool is wrapped with `<tool>` tag, content is JSON Schema
|
| 211 |
+
- `<minimax:tool_call>...</minimax:tool_call>`: Tool call area
|
| 212 |
+
- `<think>`: Thinking process marker during generation (optional)
|
| 213 |
+
|
| 214 |
+
### Model Output Format
|
| 215 |
+
|
| 216 |
+
MiniMax-M2 uses structured XML tag format:
|
| 217 |
+
|
| 218 |
+
```xml
|
| 219 |
+
<minimax:tool_call>
|
| 220 |
+
<invoke name="search_web">
|
| 221 |
+
<parameter name="query_tag">["technology", "events"]</parameter>
|
| 222 |
+
<parameter name="query_list">["\"OpenAI\" \"latest\" \"release\""]</parameter>
|
| 223 |
+
</invoke>
|
| 224 |
+
<invoke name="search_web">
|
| 225 |
+
<parameter name="query_tag">["technology", "events"]</parameter>
|
| 226 |
+
<parameter name="query_list">["\"Gemini\" \"latest\" \"release\""]</parameter>
|
| 227 |
+
</invoke>
|
| 228 |
+
</minimax:tool_call>
|
| 229 |
+
```
|
| 230 |
+
|
| 231 |
+
Each function call uses the `<invoke name="function_name">` tag, and parameters use the `<parameter name="parameter_name">` tag wrapper.
|
| 232 |
+
|
| 233 |
+
## Manually Parsing Function Call Results
|
| 234 |
+
|
| 235 |
+
### Parsing Function Calls
|
| 236 |
+
|
| 237 |
+
MiniMax-M2 uses structured XML tags, which require a different parsing approach. The core function is as follows:
|
| 238 |
+
|
| 239 |
+
```python
|
| 240 |
+
import re
|
| 241 |
+
import json
|
| 242 |
+
from typing import Any, Optional, List, Dict
|
| 243 |
+
|
| 244 |
+
|
| 245 |
+
def extract_name(name_str: str) -> str:
|
| 246 |
+
"""Extract name from quoted string"""
|
| 247 |
+
name_str = name_str.strip()
|
| 248 |
+
if name_str.startswith('"') and name_str.endswith('"'):
|
| 249 |
+
return name_str[1:-1]
|
| 250 |
+
elif name_str.startswith("'") and name_str.endswith("'"):
|
| 251 |
+
return name_str[1:-1]
|
| 252 |
+
return name_str
|
| 253 |
+
|
| 254 |
+
|
| 255 |
+
def convert_param_value(value: str, param_type: str) -> Any:
|
| 256 |
+
"""Convert parameter value based on parameter type"""
|
| 257 |
+
if value.lower() == "null":
|
| 258 |
+
return None
|
| 259 |
+
|
| 260 |
+
param_type = param_type.lower()
|
| 261 |
+
|
| 262 |
+
if param_type in ["string", "str", "text"]:
|
| 263 |
+
return value
|
| 264 |
+
elif param_type in ["integer", "int"]:
|
| 265 |
+
try:
|
| 266 |
+
return int(value)
|
| 267 |
+
except (ValueError, TypeError):
|
| 268 |
+
return value
|
| 269 |
+
elif param_type in ["number", "float"]:
|
| 270 |
+
try:
|
| 271 |
+
val = float(value)
|
| 272 |
+
return val if val != int(val) else int(val)
|
| 273 |
+
except (ValueError, TypeError):
|
| 274 |
+
return value
|
| 275 |
+
elif param_type in ["boolean", "bool"]:
|
| 276 |
+
return value.lower() in ["true", "1"]
|
| 277 |
+
elif param_type in ["object", "array"]:
|
| 278 |
+
try:
|
| 279 |
+
return json.loads(value)
|
| 280 |
+
except json.JSONDecodeError:
|
| 281 |
+
return value
|
| 282 |
+
else:
|
| 283 |
+
# Try JSON parsing, return string if failed
|
| 284 |
+
try:
|
| 285 |
+
return json.loads(value)
|
| 286 |
+
except json.JSONDecodeError:
|
| 287 |
+
return value
|
| 288 |
+
|
| 289 |
+
|
| 290 |
+
def parse_tool_calls(model_output: str, tools: Optional[List[Dict]] = None) -> List[Dict]:
|
| 291 |
+
"""
|
| 292 |
+
Extract all tool calls from model output
|
| 293 |
+
|
| 294 |
+
Args:
|
| 295 |
+
model_output: Complete output text from the model
|
| 296 |
+
tools: Tool definition list for getting parameter type information, format can be:
|
| 297 |
+
- [{"name": "...", "parameters": {...}}]
|
| 298 |
+
- [{"type": "function", "function": {"name": "...", "parameters": {...}}}]
|
| 299 |
+
|
| 300 |
+
Returns:
|
| 301 |
+
Parsed tool call list, each element contains name and arguments fields
|
| 302 |
+
|
| 303 |
+
Example:
|
| 304 |
+
>>> tools = [{
|
| 305 |
+
... "name": "get_weather",
|
| 306 |
+
... "parameters": {
|
| 307 |
+
... "type": "object",
|
| 308 |
+
... "properties": {
|
| 309 |
+
... "location": {"type": "string"},
|
| 310 |
+
... "unit": {"type": "string"}
|
| 311 |
+
... }
|
| 312 |
+
... }
|
| 313 |
+
... }]
|
| 314 |
+
>>> output = '''<minimax:tool_call>
|
| 315 |
+
... <invoke name="get_weather">
|
| 316 |
+
... <parameter name="location">San Francisco</parameter>
|
| 317 |
+
... <parameter name="unit">celsius</parameter>
|
| 318 |
+
... </invoke>
|
| 319 |
+
... </minimax:tool_call>'''
|
| 320 |
+
>>> result = parse_tool_calls(output, tools)
|
| 321 |
+
>>> print(result)
|
| 322 |
+
[{'name': 'get_weather', 'arguments': {'location': 'San Francisco', 'unit': 'celsius'}}]
|
| 323 |
+
"""
|
| 324 |
+
# Quick check if tool call marker is present
|
| 325 |
+
if "<minimax:tool_call>" not in model_output:
|
| 326 |
+
return []
|
| 327 |
+
|
| 328 |
+
tool_calls = []
|
| 329 |
+
|
| 330 |
+
try:
|
| 331 |
+
# Match all <minimax:tool_call> blocks
|
| 332 |
+
tool_call_regex = re.compile(r"<minimax:tool_call>(.*?)</minimax:tool_call>", re.DOTALL)
|
| 333 |
+
invoke_regex = re.compile(r"<invoke name=(.*?)</invoke>", re.DOTALL)
|
| 334 |
+
parameter_regex = re.compile(r"<parameter name=(.*?)</parameter>", re.DOTALL)
|
| 335 |
+
|
| 336 |
+
# Iterate through all tool_call blocks
|
| 337 |
+
for tool_call_match in tool_call_regex.findall(model_output):
|
| 338 |
+
# Iterate through all invokes in this block
|
| 339 |
+
for invoke_match in invoke_regex.findall(tool_call_match):
|
| 340 |
+
# Extract function name
|
| 341 |
+
name_match = re.search(r'^([^>]+)', invoke_match)
|
| 342 |
+
if not name_match:
|
| 343 |
+
continue
|
| 344 |
+
|
| 345 |
+
function_name = extract_name(name_match.group(1))
|
| 346 |
+
|
| 347 |
+
# Get parameter configuration
|
| 348 |
+
param_config = {}
|
| 349 |
+
if tools:
|
| 350 |
+
for tool in tools:
|
| 351 |
+
tool_name = tool.get("name") or tool.get("function", {}).get("name")
|
| 352 |
+
if tool_name == function_name:
|
| 353 |
+
params = tool.get("parameters") or tool.get("function", {}).get("parameters")
|
| 354 |
+
if isinstance(params, dict) and "properties" in params:
|
| 355 |
+
param_config = params["properties"]
|
| 356 |
+
break
|
| 357 |
+
|
| 358 |
+
# Extract parameters
|
| 359 |
+
param_dict = {}
|
| 360 |
+
for match in parameter_regex.findall(invoke_match):
|
| 361 |
+
param_match = re.search(r'^([^>]+)>(.*)', match, re.DOTALL)
|
| 362 |
+
if param_match:
|
| 363 |
+
param_name = extract_name(param_match.group(1))
|
| 364 |
+
param_value = param_match.group(2).strip()
|
| 365 |
+
|
| 366 |
+
# Remove leading and trailing newlines
|
| 367 |
+
if param_value.startswith('\n'):
|
| 368 |
+
param_value = param_value[1:]
|
| 369 |
+
if param_value.endswith('\n'):
|
| 370 |
+
param_value = param_value[:-1]
|
| 371 |
+
|
| 372 |
+
# Get parameter type and convert
|
| 373 |
+
param_type = "string"
|
| 374 |
+
if param_name in param_config:
|
| 375 |
+
if isinstance(param_config[param_name], dict) and "type" in param_config[param_name]:
|
| 376 |
+
param_type = param_config[param_name]["type"]
|
| 377 |
+
|
| 378 |
+
param_dict[param_name] = convert_param_value(param_value, param_type)
|
| 379 |
+
|
| 380 |
+
tool_calls.append({
|
| 381 |
+
"name": function_name,
|
| 382 |
+
"arguments": param_dict
|
| 383 |
+
})
|
| 384 |
+
|
| 385 |
+
except Exception as e:
|
| 386 |
+
print(f"Failed to parse tool calls: {e}")
|
| 387 |
+
return []
|
| 388 |
+
|
| 389 |
+
return tool_calls
|
| 390 |
+
```
|
| 391 |
+
|
| 392 |
+
**Usage Example:**
|
| 393 |
+
|
| 394 |
+
```python
|
| 395 |
+
# Define tools
|
| 396 |
+
tools = [
|
| 397 |
+
{
|
| 398 |
+
"name": "get_weather",
|
| 399 |
+
"parameters": {
|
| 400 |
+
"type": "object",
|
| 401 |
+
"properties": {
|
| 402 |
+
"location": {"type": "string"},
|
| 403 |
+
"unit": {"type": "string"}
|
| 404 |
+
},
|
| 405 |
+
"required": ["location", "unit"]
|
| 406 |
+
}
|
| 407 |
+
}
|
| 408 |
+
]
|
| 409 |
+
|
| 410 |
+
# Model output
|
| 411 |
+
model_output = """Let me help you query the weather.
|
| 412 |
+
<minimax:tool_call>
|
| 413 |
+
<invoke name="get_weather">
|
| 414 |
+
<parameter name="location">San Francisco</parameter>
|
| 415 |
+
<parameter name="unit">celsius</parameter>
|
| 416 |
+
</invoke>
|
| 417 |
+
</minimax:tool_call>"""
|
| 418 |
+
|
| 419 |
+
# Parse tool calls
|
| 420 |
+
tool_calls = parse_tool_calls(model_output, tools)
|
| 421 |
+
|
| 422 |
+
# Output results
|
| 423 |
+
for call in tool_calls:
|
| 424 |
+
print(f"Function called: {call['name']}")
|
| 425 |
+
print(f"Arguments: {call['arguments']}")
|
| 426 |
+
# Output: Function called: get_weather
|
| 427 |
+
# Arguments: {'location': 'San Francisco', 'unit': 'celsius'}
|
| 428 |
+
```
|
| 429 |
+
|
| 430 |
+
### Executing Function Calls
|
| 431 |
+
|
| 432 |
+
After parsing is complete, you can execute the corresponding function and construct the return result:
|
| 433 |
+
|
| 434 |
+
```python
|
| 435 |
+
def execute_function_call(function_name: str, arguments: dict):
|
| 436 |
+
"""Execute function call and return result"""
|
| 437 |
+
if function_name == "get_weather":
|
| 438 |
+
location = arguments.get("location", "Unknown location")
|
| 439 |
+
unit = arguments.get("unit", "celsius")
|
| 440 |
+
# Build function execution result
|
| 441 |
+
return {
|
| 442 |
+
"role": "tool",
|
| 443 |
+
"content": [
|
| 444 |
+
{
|
| 445 |
+
"name": function_name,
|
| 446 |
+
"type": "text",
|
| 447 |
+
"text": json.dumps({
|
| 448 |
+
"location": location,
|
| 449 |
+
"temperature": "25",
|
| 450 |
+
"unit": unit,
|
| 451 |
+
"weather": "Sunny"
|
| 452 |
+
}, ensure_ascii=False)
|
| 453 |
+
}
|
| 454 |
+
]
|
| 455 |
+
}
|
| 456 |
+
elif function_name == "search_web":
|
| 457 |
+
query_list = arguments.get("query_list", [])
|
| 458 |
+
query_tag = arguments.get("query_tag", [])
|
| 459 |
+
# Simulate search results
|
| 460 |
+
return {
|
| 461 |
+
"role": "tool",
|
| 462 |
+
"content": [
|
| 463 |
+
{
|
| 464 |
+
"name": function_name,
|
| 465 |
+
"type": "text",
|
| 466 |
+
"text": f"Search keywords: {query_list}, Category: {query_tag}\nSearch results: Relevant information found"
|
| 467 |
+
}
|
| 468 |
+
]
|
| 469 |
+
}
|
| 470 |
+
|
| 471 |
+
return None
|
| 472 |
+
```
|
| 473 |
+
|
| 474 |
+
### Returning Function Execution Results to the Model
|
| 475 |
+
|
| 476 |
+
After successfully parsing function calls, you should add the function execution results to the conversation history so that the model can access and utilize this information in subsequent interactions. Refer to chat_template.jinja for concatenation format.
|
| 477 |
+
|
| 478 |
+
## References
|
| 479 |
+
|
| 480 |
+
- [MiniMax-M2 Model Repository](https://github.com/MiniMaxAI/MiniMax-M2)
|
| 481 |
+
- [vLLM Project Homepage](https://github.com/vllm-project/vllm)
|
| 482 |
+
- [OpenAI Python SDK](https://github.com/openai/openai-python)
|
docs/vllm_deploy_guide.md
ADDED
|
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MiniMax M2 Model vLLM Deployment Guide
|
| 2 |
+
|
| 3 |
+
We recommend using [vLLM](https://docs.vllm.ai/en/latest/) to deploy the [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2) model. vLLM is a high-performance inference engine with excellent serving throughput, efficient and intelligent memory management, powerful batch request processing capabilities, and deeply optimized underlying performance. We recommend reviewing vLLM's official documentation to check hardware compatibility before deployment.
|
| 4 |
+
|
| 5 |
+
## Applicable Models
|
| 6 |
+
|
| 7 |
+
This guide applies to the following models. Simply modify the model name during deployment. The deployment process is illustrated using MiniMax-M2 as an example.
|
| 8 |
+
|
| 9 |
+
- MiniMaxAI/MiniMax-M2
|
| 10 |
+
|
| 11 |
+
## System Requirements
|
| 12 |
+
|
| 13 |
+
- OS: Linux
|
| 14 |
+
|
| 15 |
+
- Python: 3.9 - 3.12
|
| 16 |
+
|
| 17 |
+
- GPU:
|
| 18 |
+
|
| 19 |
+
- compute capability 7.0 or higher
|
| 20 |
+
|
| 21 |
+
- Memory requirements: 220 GB for weights, 60 GB per 1M context tokens
|
| 22 |
+
|
| 23 |
+
The following are recommended configurations; actual requirements should be adjusted based on your use case:
|
| 24 |
+
|
| 25 |
+
- 4x 96GB GPUs: Supports context input of up to 400K tokens.
|
| 26 |
+
|
| 27 |
+
- 8x 144GB GPUs: Supports context input of up to 3M tokens.
|
| 28 |
+
|
| 29 |
+
## Deployment with Python
|
| 30 |
+
|
| 31 |
+
It is recommended to use a virtual environment (such as venv, conda, or uv) to avoid dependency conflicts. We recommend installing vLLM in a fresh Python environment:
|
| 32 |
+
|
| 33 |
+
```bash
|
| 34 |
+
# Not yet released, please install nightly build
|
| 35 |
+
uv pip install -U vllm \
|
| 36 |
+
--torch-backend=auto \
|
| 37 |
+
--extra-index-url https://wheels.vllm.ai/nightly
|
| 38 |
+
# If released, install using uv
|
| 39 |
+
uv pip install "vllm" --torch-backend=auto
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
Run the following command to start the vLLM server. vLLM will automatically download and cache the MiniMax-M2 model from Hugging Face.
|
| 43 |
+
|
| 44 |
+
4-GPU deployment command:
|
| 45 |
+
|
| 46 |
+
```bash
|
| 47 |
+
SAFETENSORS_FAST_GPU=1 VLLM_USE_V1=0 vllm serve \
|
| 48 |
+
--model MiniMaxAI/MiniMax-M2 \
|
| 49 |
+
--trust-remote-code \
|
| 50 |
+
--enable-expert-parallel --tensor-parallel-size 4 \
|
| 51 |
+
--enable-auto-tool-choice --tool-call-parser minimax_m2 \
|
| 52 |
+
--reasoning-parser minimax_m2
|
| 53 |
+
```
|
| 54 |
+
|
| 55 |
+
## Testing Deployment
|
| 56 |
+
|
| 57 |
+
After startup, you can test the vLLM OpenAI-compatible API with the following command:
|
| 58 |
+
|
| 59 |
+
```bash
|
| 60 |
+
curl http://localhost:8000/v1/chat/completions \
|
| 61 |
+
-H "Content-Type: application/json" \
|
| 62 |
+
-d '{
|
| 63 |
+
"model": "MiniMaxAI/MiniMax-M2",
|
| 64 |
+
"messages": [
|
| 65 |
+
{"role": "system", "content": [{"type": "text", "text": "You are a helpful assistant."}]},
|
| 66 |
+
{"role": "user", "content": [{"type": "text", "text": "Who won the world series in 2020?"}]}
|
| 67 |
+
]
|
| 68 |
+
}'
|
| 69 |
+
```
|
| 70 |
+
|
| 71 |
+
## Common Issues
|
| 72 |
+
|
| 73 |
+
### Hugging Face Network Issues
|
| 74 |
+
|
| 75 |
+
If you encounter network issues, you can set up a proxy before pulling the model.
|
| 76 |
+
|
| 77 |
+
```bash
|
| 78 |
+
export HF_ENDPOINT=https://hf-mirror.com
|
| 79 |
+
```
|
| 80 |
+
|
| 81 |
+
### MiniMax-M2 model is not currently supported
|
| 82 |
+
|
| 83 |
+
This vLLM version is outdated. Please upgrade to the latest version.
|
| 84 |
+
|
| 85 |
+
## Getting Support
|
| 86 |
+
|
| 87 |
+
If you encounter any issues while deploying the MiniMax model:
|
| 88 |
+
|
| 89 |
+
- Contact our technical support team through official channels such as email at api@minimaxi.com
|
| 90 |
+
|
| 91 |
+
- Submit an issue on our [GitHub](https://github.com/MiniMax-AI) repository
|
| 92 |
+
|
| 93 |
+
We continuously optimize the deployment experience for our models. Feedback is welcome!
|
| 94 |
+
|
docs/vllm_deploy_guide_cn.md
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MiniMax M2 模型 vLLM 部署指南
|
| 2 |
+
|
| 3 |
+
我们推荐使用 [vLLM](https://docs.vllm.ai/en/latest/) 来部署 [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2) 模型。vLLM 是一个高性能的推理引擎,其具有卓越的服务吞吐、高效智能的内存管理机制、强大的批量请求处理能力、深度优化的底层性能等特性。我们建议在部署之前查看 vLLM 的官方文档以检查硬件兼容性。
|
| 4 |
+
|
| 5 |
+
## 本文档适用模型
|
| 6 |
+
|
| 7 |
+
本文档适用以下模型,只需在部署时修改模型名称即可。以下以 MiniMax-M2 为例说明部署流程。
|
| 8 |
+
|
| 9 |
+
- MiniMaxAI/MiniMax-M2
|
| 10 |
+
|
| 11 |
+
## 环境要求
|
| 12 |
+
|
| 13 |
+
- OS:Linux
|
| 14 |
+
|
| 15 |
+
- Python:3.9 - 3.12
|
| 16 |
+
|
| 17 |
+
- GPU:
|
| 18 |
+
|
| 19 |
+
- compute capability 7.0 or higher
|
| 20 |
+
|
| 21 |
+
- 显存需求:权重需要 220 GB,每 1M 上下文 token 需要 60 GB
|
| 22 |
+
|
| 23 |
+
以下为推荐配置,实际需求请根据业务场景调整:
|
| 24 |
+
|
| 25 |
+
- 96G x4 GPU:支持 40 万 token 的上下文输入。
|
| 26 |
+
|
| 27 |
+
- 144G x8 GPU:支持长达 300 万 token 的上下文输入。
|
| 28 |
+
|
| 29 |
+
## 使用 Python 部署
|
| 30 |
+
|
| 31 |
+
建议使用虚拟环境(如 venv、conda、uv)以避免依赖冲突。建议在全新的 Python 环境中安装 vLLM:
|
| 32 |
+
```bash
|
| 33 |
+
# 尚未 release,请安装 nightly 构建
|
| 34 |
+
uv pip install -U vllm \
|
| 35 |
+
--torch-backend=auto \
|
| 36 |
+
--extra-index-url https://wheels.vllm.ai/nightly
|
| 37 |
+
# 如果 release,使用 uv 安装
|
| 38 |
+
uv pip install "vllm" --torch-backend=auto
|
| 39 |
+
```
|
| 40 |
+
|
| 41 |
+
运行如下命令启动 vLLM 服务器,vLLM 会自动从 Huggingface 下载并缓存 MiniMax-M2 模型。
|
| 42 |
+
|
| 43 |
+
4 卡部署命令:
|
| 44 |
+
|
| 45 |
+
```bash
|
| 46 |
+
SAFETENSORS_FAST_GPU=1 VLLM_USE_V1=0 vllm serve \
|
| 47 |
+
--model MiniMaxAI/MiniMax-M2 \
|
| 48 |
+
--trust-remote-code \
|
| 49 |
+
--enable-expert-parallel --tensor-parallel-size 4 \
|
| 50 |
+
--enable-auto-tool-choice --tool-call-parser minimax_m2 \
|
| 51 |
+
--reasoning-parser minimax_m2
|
| 52 |
+
```
|
| 53 |
+
|
| 54 |
+
## 测试部署
|
| 55 |
+
|
| 56 |
+
启动后,可以通过如下命令测试 vLLM OpenAI 兼容接口:
|
| 57 |
+
|
| 58 |
+
```bash
|
| 59 |
+
curl http://localhost:8000/v1/chat/completions \
|
| 60 |
+
-H "Content-Type: application/json" \
|
| 61 |
+
-d '{
|
| 62 |
+
"model": "MiniMaxAI/MiniMax-M2",
|
| 63 |
+
"messages": [
|
| 64 |
+
{"role": "system", "content": [{"type": "text", "text": "You are a helpful assistant."}]},
|
| 65 |
+
{"role": "user", "content": [{"type": "text", "text": "Who won the world series in 2020?"}]}
|
| 66 |
+
]
|
| 67 |
+
}'
|
| 68 |
+
```
|
| 69 |
+
|
| 70 |
+
## 常见问题
|
| 71 |
+
|
| 72 |
+
### Huggingface 网络问题
|
| 73 |
+
|
| 74 |
+
如果遇到网络问题,可以设置代理后再进行拉取。
|
| 75 |
+
|
| 76 |
+
```bash
|
| 77 |
+
export HF_ENDPOINT=https://hf-mirror.com
|
| 78 |
+
```
|
| 79 |
+
|
| 80 |
+
### MiniMax-M2 model is not currently supported
|
| 81 |
+
|
| 82 |
+
该 vLLM 版本过旧,请升级到最新版本。
|
| 83 |
+
|
| 84 |
+
## 获取支持
|
| 85 |
+
|
| 86 |
+
如果在部署 MiniMax 模型过程中遇到任何问题:
|
| 87 |
+
|
| 88 |
+
- 通过邮箱 api@minimaxi.com 等官方渠道联系我们的技术支持团队
|
| 89 |
+
|
| 90 |
+
- 在我们的 [GitHub](https://github.com/MiniMax-AI) 仓库提交 Issue
|
| 91 |
+
我们会持续优化模型的部署体验,欢迎反馈!
|