We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

DeepInfra raises $107M Series B to scale the inference cloud — read the announcement

XiaomiMiMo/

MiMo-V2.5-Pro

$1.00

in

$3.00

out

$0.20

cached

/ 1M tokens

MiMo-V2.5-Pro is an open-source Mixture-of-Experts (MoE) language model with 1.02T total parameters and 42B active parameters. It utilizes the hybrid attention architecture and 3-layers Multi-Token Prediction (MTP) introduced in [MiMo-V2-Flash](https://github.com/XiaomiMiMo/MiMo-V2-Flash).

Deploy Private Endpoint
Public
fp8
1,048,576
JSON
Function
XiaomiMiMo/MiMo-V2.5-Pro cover image
demoapi

GKZPammL

2026-05-03T02:03:53+00:00