DeepInfra raises $107M Series B to scale the inference cloud — read the announcement
XiaomiMiMo/
$1.00
in
$3.00
out
$0.20
cached
/ 1M tokens
MiMo-V2.5-Pro is an open-source Mixture-of-Experts (MoE) language model with 1.02T total parameters and 42B active parameters. It utilizes the hybrid attention architecture and 3-layers Multi-Token Prediction (MTP) introduced in [MiMo-V2-Flash](https://github.com/XiaomiMiMo/MiMo-V2-Flash).

© 2026 DeepInfra. All rights reserved.