NVIDIA Nemotron 3 Super - blazing-fast agentic AI, ready to deploy today!

Today we're excited to share that Deep Infra has raised $18 million in Series A funding, led by Felicis and our earliest believer and advisor Georges Harik.
When we founded Deep Infra in 2022, we saw a clear gap: while enormous resources were being poured into training AI models, the infrastructure needed to run these models in production was lagging behind.
The past two years have been a whirlwind. We've scaled our processing volume by over 8,000x since our seed stage. What started as a bet on AI infrastructure has quickly become a critical service for developers deploying increasingly sophisticated models.
Our growth accelerated following the emergence of "thinking models" like DeepSeek. These open source alternatives demonstrated that the innovation cycle in AI was becoming even more rapid than anticipated, requiring significantly more computation during inference.
The reality of deploying modern AI models is challenging for most organizations. Running these models requires significant compute resources, specialized hardware like GPUs that are difficult to acquire, and deep expertise in infrastructure optimization. Most companies simply can't afford the investment or overcome the supply chain challenges to build this infrastructure themselves.
This challenge has shaped our approach from day one. After years of scaling systems to hundreds of millions of users before founding this company, we've developed a set of core principles that guide how we build Deep Infra:
These principles have guided our approach as we've expanded our computing capacity, recently receiving a large shipment of NVIDIA Blackwell GPUs with more on order to support our rapid growth. You can see how this funding injection will be put to good use.
To our customers who have trusted us with their production workloads: thank you. We're just getting started as we continue building the infrastructure that powers the next generation of AI applications.
Follow us on X (formerly Twitter) and LinkedIn to stay updated on our journey. We look forward to sharing more exciting developments in the coming months.

Kimi K2 0905 API from Deepinfra: Practical Speed, Predictable Costs, Built for Devs - Deep Infra<p>Kimi K2 0905 is Moonshot’s long-context Mixture-of-Experts update designed for agentic and coding workflows. With a context window up to ~256K tokens, it can ingest large codebases, multi-file documents, or long conversations and still deliver structured, high-quality outputs. But real-world performance isn’t defined by the model alone—it’s determined by the inference provider that serves it: […]</p>
Kimi K2 0905 API Benchmarks: Latency, Throughput & Cost<p>About Kimi K2 0905 Kimi K2 0905 is a state-of-the-art large language model developed by Moonshot AI, representing a significant advancement in open-weight AI capabilities. This Mixture-of-Experts (MoE) model features 1 trillion total parameters with 32 billion activated parameters per forward pass, making it highly efficient while maintaining frontier-level performance. The model supports a 256k […]</p>
Qwen3.5 35B A3B API Benchmarks: Latency, Throughput & Cost<p>About Qwen3.5 35B A3B Qwen3.5 35B A3B is a native vision-language model released by Alibaba Cloud in February 2026. It uses a hybrid architecture that integrates Gated Delta Networks with a sparse Mixture-of-Experts model, achieving higher inference efficiency. With 35 billion total parameters and only 3 billion activated per token through 256 experts (8 routed […]</p>
© 2026 Deep Infra. All rights reserved.