GLM-5.1 - state-of-the-art agentic engineering, now available on DeepInfra!

GLM-5 is the latest open-weights reasoning model released by Z AI (Zhipu AI) in February 2026, characterized by high “thinking token” usage. It is a Mixture of Experts (MoE) model with 744B total parameters and 40B active parameters, scaling up from GLM-4.5’s 355B parameters. The model was pre-trained on 28.5T tokens and features a 200K+ context window.
GLM-5 is purpose-built for complex systems engineering and long-horizon agentic tasks. It integrates DeepSeek Sparse Attention (DSA), substantially reducing deployment costs while preserving long-context capacity. The model was trained using a novel asynchronous RL infrastructure called “Slime” that improves training throughput and efficiency.
GLM-5 achieves best-in-class performance among all open-source models on reasoning, coding, and agentic tasks, closing the gap with frontier models like Claude Opus 4.5. On SWE-bench Verified and Terminal-Bench 2.0, GLM-5 records leading open-model scores of 77.8 and 56.2, respectively.
GLM-5 is now available across multiple inference providers, but they’re not created equal. This analysis breaks down which one delivers the best performance, lowest cost, and fastest response times for your use case.
| Provider | Best For | Blended ($/1M) | Input ($/1M) | Output ($/1M) | Speed (t/s) | Latency (TTFT) | JSON | Func Calling | Why Notable |
|---|---|---|---|---|---|---|---|---|---|
| DeepInfra FP8 (deepinfra.com) | Lowest cost / best cost baseline | $1.24 | $0.80 | $2.56 | 57 | 1.22s | Yes | Yes | Cheapest blended + input/output; competitive latency for interactive apps at budget. |
| Fireworks | Max throughput + lowest latency | $1.55 | — | — | 212.2 | 0.74s | Yes | Yes | Best raw performance (fastest output + lowest TTFT) but higher blended price. |
| Baseten | Strong balance (fast + low cost) | $1.50 | $0.95 | $3.15 | 183.1 | 0.83s | Yes | Yes | Near-top performance with 2nd-lowest blended price. |
| Eigen AI | High speed alternative | $1.55 | — | — | 204.5 | 1.53s | Yes | Yes | Very high output speed; blended price higher than DeepInfra. |
| Novita FP8 | Price runner-up | $1.55 | $1.00 | $3.20 | 50 | 1.45s | Yes | Yes | Among lowest blended prices, but still above DeepInfra on all metrics. |
Based on benchmarks of all the aforementioned providers, DeepInfra is the recommended API for production-scale GLM-5 deployment. It offers the lowest blended price on the market ($1.24 per 1M tokens), which is critical for reasoning models that generate high volumes of internal chain-of-thought tokens. For use cases requiring sub-second latency, Fireworks is the fastest provider tested.
| Use Case | Recommended Provider | Why? |
|---|---|---|
| Chatbots / Real-time | Fireworks | Lowest Latency (0.74s) |
| Batch Processing / RAG | DeepInfra | Lowest Cost ($1.24/1M) |
| Agentic Workflows | Baseten | Balanced Speed & Cost |
While other providers chase raw burst speed, DeepInfra (FP8) secures the top recommendation by dominating the most critical metric for scaling reasoning models: Cost Efficiency.
Reasoning models like GLM-5 generate a high volume of “thinking” tokens before producing a final answer. This drastically inflates output token usage compared to standard LLMs. DeepInfra offers the lowest blended price and, crucially, the most competitive output pricing, making it the optimal choice for high-volume production environments where margins matter.
DeepInfra provides a competitive latency profile (tying for 4th place) while undercutting the market standard price ($1.55) by 20%. For applications requiring heavy reasoning chains, the savings on output tokens make DeepInfra the superior architectural choice.
For applications requiring massive batch processing or long-form content generation where the user is waiting for the stream to finish, Output Tokens Per Second (t/s) is the governing metric.
Technical Note: The gap between the top tier (Fireworks/Eigen) and the mid-tier (Google at 75.7 t/s) is substantial. If your application relies on rapid text generation (e.g., code autocompletion), the premium for Fireworks or Eigen is justified.
TTFT is the critical metric for conversational AI and chatbots. It measures the time between the request and the first visible character.
DeepInfra and Together.ai (FP4) sit in the second tier at 1.22s. While slower than Fireworks, 1.2 seconds is generally acceptable for most asynchronous reasoning tasks where the user expects a brief thinking pause.
Winner: DeepInfra (FP8)
While speed is important, the cost structure is the deciding factor for reasoning models. GLM-5 generates significant “thinking” tokens, inflating output costs.
DeepInfra undercuts the market average by 20%. For high-volume reasoning chains, this price difference is the primary differentiator. The market has largely coalesced around a standard blended price of $1.55 per 1 million tokens. Deviations from this norm highlight the value leaders:
Technical integration is just as important as raw speed.
| Provider | Blended Price (/1M) | Latency (TTFT) | Output Speed (t/s) | Recommendation |
|---|---|---|---|---|
| DeepInfra FP8 | $1.24 | 1.22s | ~57 | Best Overall (Value) |
| Fireworks | $1.55 | 0.74s | 212.2 | Best Performance |
| Baseten | $1.50 | 0.83s | 183.1 | Best Balanced |
| Eigen AI | $1.55 | 1.53s | 204.5 | High Throughput |
| FriendliAI | $1.55 | 0.94s | 73.9 | Low Latency |
| $1.55 | 1.33s | 75.7 | Standard |
For developers integrating GLM-5 (Reasoning) into their stack, the optimal choice depends on the specific bottleneck of the application.
If you are building a consumer-facing chatbot where every millisecond counts, Fireworks is the technically superior option. However, for the vast majority of enterprise use cases — where reasoning models are used for complex data processing, RAG pipelines, or agentic workflows — DeepInfra is the recommended provider. It offers a robust feature set (JSON mode and Function calling), acceptable latency, and a pricing structure that provides a significant long-term competitive advantage.
Kimi K2 0905 API Benchmarks: Latency, Throughput & Cost<p>About Kimi K2 0905 Kimi K2 0905 is a state-of-the-art large language model developed by Moonshot AI, representing a significant advancement in open-weight AI capabilities. This Mixture-of-Experts (MoE) model features 1 trillion total parameters with 32 billion activated parameters per forward pass, making it highly efficient while maintaining frontier-level performance. The model supports a 256k […]</p>
Qwen3.5 27B API Benchmarks: Latency, Throughput & Cost<p>About Qwen3.5 27B (Reasoning) Qwen3.5 27B is part of Alibaba Cloud’s latest-generation foundation model family, released in February 2026. Unlike the Mixture-of-Experts variants in the Qwen3.5 series, the 27B model uses a dense architecture combining Gated Delta Networks and Feed Forward Networks. It achieves strong benchmark scores including MMLU-Pro (86.1%), GPQA Diamond (85.5%), and SWE-bench […]</p>
Building Efficient AI Inference on NVIDIA Blackwell PlatformDeepInfra delivers up to 20x cost reductions on NVIDIA Blackwell by combining MoE architectures, NVFP4 quantization, and inference optimizations — with a Latitude case study.© 2026 Deep Infra. All rights reserved.