bge-m3

Model Description

BGE-M3 stands out for its Multi-Functionality (simultaneous dense, sparse, and multi-vector retrieval), Multi-Linguality (100+ languages), and Multi-Granularity (up to 8,192-token documents). It enhances retrieval pipelines by enabling hybrid retrieval (e.g., combining dense embeddings with BM25-like sparse weights) and re-ranking for higher accuracy. The model integrates seamlessly with tools like Vespa and Milvus, and its unified fine-tuning supports diverse retrieval methods. Recent updates include improved MIRACL benchmark performance and multilingual long-document datasets (MLDR).

🔔How to Use

graph LR A("Purchase Now") --> B["Start Chat on Homepage"] A --> D["Read API Documentation"] B --> C["Register / Login"] C --> E["Enter Key"] D --> F["Enter Endpoint & Key"] E --> G("Start Using") F --> G style A fill:#f9f9f9,stroke:#333,stroke-width:1px style B fill:#f9f9f9,stroke:#333,stroke-width:1px style C fill:#f9f9f9,stroke:#333,stroke-width:1px style D fill:#f9f9f9,stroke:#333,stroke-width:1px style E fill:#f9f9f9,stroke:#333,stroke-width:1px style F fill:#f9f9f9,stroke:#333,stroke-width:1px style G fill:#f9f9f9,stroke:#333,stroke-width:1px

Purchase Now

Start Chat on Homepage

Register / Login

Enter Key

Read API Documentation

Enter Endpoint & Key

Start Using

Description Ends

Recommend Models

gemini-2.5-flash-lite-preview-06-17

A Gemini 2.5 Flash model optimized for cost efficiency and low latency.

gpt-5

Our smartest, fastest, and most useful model yet, with thinking built in. Available to everyone.

gpt-4.1-nano

GPT-4.1 nano is the fastest, most cost-effective GPT-4.1 model.