We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

Nemotron 3 Nano Omni — the first multimodal model in the Nemotron 3 family, now on DeepInfra!

ByteDance logo

ByteDance/

Seed-2.0-code

Partner

$0.50

in

$3.00

out

$0.10

cached

/ 1M tokens

*

A coding model optimized for real-world development environments, with reliable tool use in common IDEs such as Claude Code. It delivers strong front-end performance and supports Skills.

Public
256,000
JSON
Function
Multimodal
ByteDance/Seed-2.0-code cover image
ByteDance/Seed-2.0-code cover image
Seed-2.0-code

Ask me anything

0.00s

Settings

Model Information

Seed-2.0-Code is optimized for enterprise-grade coding scenarios. Building on the strong agentic and VLM capabilities of Seed 2.0, it further strengthens code generation and software engineering performance. It delivers particularly strong front-end results and is also специально optimized for the multilingual coding needs commonly found in enterprise environments, making it well suited for integration with a wide range of AI coding tools.

New Features

  • Built on a powerful foundation model: Based on the Seed 2.0 foundation model, Dola-Seed-2.0-Code provides strong agentic capabilities, including support for skill invocation and custom tool use. It also offers robust VLM capabilities, enabling image-grounded tool use such as Browser Use.
  • Optimized for programming tasks: The model is specifically enhanced for working with a broad set of programming languages and developer tools, and performs strongly on long-horizon, multi-turn coding tasks.
  • Strong complex instruction following: It can reliably understand long and complex programming instructions and consistently deliver outputs that match user expectations. -Implicit caching support: Caching is enabled by default with no additional configuration required, making it suitable for agentic workflows and other complex, long-context scenarios. The model may automatically cache shared prefixes of requests; however, cache hits are not guaranteed. In general, a cache hit is only triggered once at least 1,024 tokens have accumulated. Cached-hit segments are billed at the list price for cached input tokens.