
MiniMax M2.5 HighSpeed
MiniMax M2.5 high-speed variant, optimized for low-latency scenarios with faster inference
Context Window
205K
Input price / 1M tokens
$0.651M tokens
Output price / 1M tokens
$2.581M tokens
Cached input / 1M tokens
$0.031M tokens
Max Completion
33K
Input Modalities
text
Output Modalities
text
reasoningFunction callingChatStreaming
Description
MiniMax M2.5 high-speed variant, optimized for low-latency scenarios with faster inference
Available Providers
AllToken can route requests to the providers below based on route priority and policy.
ProviderContext LengthInput PriceOutput PriceCached / MLatency p50Throughput
Best For
MiniMax M2.5 high-speed variant, optimized for low-latency scenarios with faster inference
How To Use This Model
Use the exact model ID shown below. This is the safest way to avoid call failures, variant mismatches, or incorrect route assumptions.
curl https://api.alltoken.ai/v1/chat/completions \
-H "Authorization: Bearer sk-your-key" \
-H "Content-Type: application/json" \
-d '{
"model": "minimax-m2.5-highspeed",
"messages": [
{"role": "user", "content": "Hello!"}
]
}'Supported Parameters
temperaturetop_pmax_tokenstoolsreasoning_effortAPI Key Setup
Smart Routing
Let the platform choose the best provider path automatically.
Default Model
If a request does not specify a model, default the key to minimax-m2.5-highspeed.
Forced Model
Always override incoming requests and lock the key to minimax-m2.5-highspeed.