Model Filter

Model Type

Features

Context Windown

Maxmium Output

Provider

Recommend

We have launched the Basic Series economy models, offering higher discounts. Click to view the model comparison >>

1MT: One million tokens. This pricing is based on the conversion rate of ¥2 = $1. If your purchase rate is ¥3.5 = $1, the price should be multiplied by 1.75 accordingly.

grok-4-0709

Our latest and greatest flagship model, offering unparalleled performance in natural language, math and reasoning - the perfect jack of all trades.

claudecode/claude-sonnet-4-20250514

The Claude model series offered by the Claude Code has moderate stability and is extremely low-priced, making it more suitable for data batch processing tasks where strict stability requirements are not particularly stringent.

az/claude-sonnet-4-20250514

The Claude model series offered by the Microsoft Azure platform has moderate stability and is extremely low-priced, making it more suitable for data batch processing tasks where strict stability requirements are not particularly stringent.

sora_image

Reverse-engineered version of the official GPT-Image-1, featuring stable performance, high cost-effectiveness, compatibility with traditional OpenAI formats, and support for direct image generation through conversation.

gemini-2.5-pro

Gemini 2.5 Pro is Google's most advanced AI model designed for coding and complex tasks, featuring enhanced reasoning capabilities, native multimodal support, and a 1-million token context window.

gemini-2.5-flash

Gemini 2.5 Flash is Google's most efficient multimodal AI model designed for fast, cost-effective performance on everyday tasks with native audio capabilities and a 1-million token context window.

gemini-2.5-flash-lite-preview-06-17

A Gemini 2.5 Flash model optimized for cost efficiency and low latency.

o3-pro

The o-series of models are trained with reinforcement learning to think before they answer and perform complex reasoning. The o3-pro model uses more compute to think harder and provide consistently better answers. o3-pro is available in the Responses API only to enable support for multi-turn model interactions before responding to API requests, and other advanced API features in the future. Since o3-pro is designed to tackle tough problems, some requests may take several minutes to finish. To avoid timeouts, try using background mode.

gemini-2.5-pro-preview-06-05

Google has released an upgraded preview of Gemini 2.5 Pro (06-05) that significantly improves coding performance, mathematical reasoning, and response formatting while addressing previous performance concerns.

DeepSeek-R1-0528

The DeepSeek-R1-0528 is an upgraded AI model with enhanced programming, design, and inference efficiency, delivering high-quality outputs for complex tasks.

gemini-2.5-flash-preview-05-20

A comprehensive overview of Google Gemini 2.5 Flash (gemini-2.5-flash-preview-05-20), focusing on its hybrid reasoning architecture, multimodal capabilities, optimized performance, API pricing, application scenarios, and future developments in the AI field.

basic/gemini-2.5-flash-preview-05-20

A comprehensive overview of Google Gemini 2.5 Flash (gemini-2.5-flash-preview-05-20), focusing on its hybrid reasoning architecture, multimodal capabilities, optimized performance, API pricing, application scenarios, and future developments in the AI field.

claude-opus-4-20250514

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

claude-opus-4-20250514-thinking

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

claude-sonnet-4-20250514

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

claude-sonnet-4-20250514-thinking

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

basic2/claude-opus-4-20250514

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

basic2/claude-sonnet-4-20250514

Comprehensive introduction to Anthropic's newly released Claude 4 models, Opus 4 and Sonnet 4, highlighting their features, performance benchmarks, application scenarios, pricing, and availability. This report summarizes key differences between the models and discusses their integration with major platforms such as GitHub Copilot, emphasizing their advantages in coding, advanced reasoning, and ethical AI responses.

gemini-2.5-pro-preview-05-06

Gemini 2.5 Pro is our state-of-the-art thinking model, capable of reasoning over complex problems in code, math, and STEM, as well as analyzing large datasets, codebases, and documents using long context.

qwen3-235b-a22b

Qwen3-235B-A22B, part of the cutting-edge Qwen3 series, introduces the next generation of large language models with both dense and mixture-of-experts (MoE) architectures. This advanced model enables seamless switching between thinking mode (for complex reasoning, math, and coding) and non-thinking mode (for fast, general-purpose dialogue) within a single model—delivering optimal performance for a wide range of applications.