We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

BAAI/

bge-m3-multi

BGE-M3 is a multilingual text embedding model developed by BAAI, distinguished by its Multi-Linguality (supporting 100+ languages), Multi-Functionality (unified dense, multi-vector, and sparse retrieval), and Multi-Granularity (handling inputs from short queries to 8192-token documents). It achieves state-of-the-art retrieval performance across diverse benchmarks while maintaining a single model for multiple retrieval modes.

Public
$0.010 / Mtoken
8,192
ProjectPaperLicense
BAAI/bge-m3-multi cover image

Input

inputs
You can add more items with the button on the right

You need to login to use this model

Login

Settings

Dense is a set of low-dimensional vectors where every token in the input is represented by a fully populated embedding derived from a neural model. 2

Sparse is a collection of high-dimensional vectors where each word in the input is assigned a lexical weight, with most values being zero. 2

Colbert is a system of contextualized vectors where every token in the input is represented by its own BERT-derived embedding. 2

whether to normalize the computed embeddings 2

Output

MultiModalEmbeddingsOut
{
  "input_tokens": 42,
  "embeddings": [
    [
      0,
      0.5,
      1
    ],
    [
      1,
      0.5,
      0
    ]
  ],
  "sparse": [
    [
      0,
      0,
      1
    ],
    [
      0,
      0.6,
      0
    ]
  ],
  "colbert": [
    [
      [
        0.5,
        0.1,
        1
      ],
      [
        0.3,
        0.6,
        0.5
      ]
    ],
    [
      [
        0.3,
        0.6,
        0.5
      ]
    ]
  ],
  "embedding_jsons": [
    "[0.0, 0.5, 1.0]",
    "[1.0, 0.5, 0.0]"
  ]
}

Model Information

Unlock the most affordable AI hosting

Run models at scale with our fully managed GPU infrastructure, delivering enterprise-grade uptime at the industry's best rates.