GLM-4-32B-0414

Model Description

GLM-4-32B-0414 is a new generation open-source model in the GLM series, with 32 billion parameters. The model’s performance is comparable to OpenAI’s GPT series and DeepSeek’s V3/R1 series, and it supports very user-friendly local deployment features. GLM-4-32B-Base-0414 is pre-trained on 15T of high-quality data, including a large amount of synthetic data for various reasoning types, laying the foundation for subsequent reinforcement learning extensions. In the post-training phase, in addition to aligning human preferences in dialogue scenarios, the research team also enhanced the model’s performance in instruction following, engineering code, and function calls using techniques such as rejection sampling and reinforcement learning, strengthening the atomic capabilities required for agent tasks. GLM-4-32B-0414 has achieved good results in areas such as engineering code, artifact generation, function calls, search-based question answering, and report generation, with some benchmark metrics approaching or even surpassing the levels of larger models like GPT-4o and DeepSeek-V3-0324 (671B).

🔔How to Use

graph LR A("Purchase Now") --> B["Start Chat on Homepage"] A --> D["Read API Documentation"] B --> C["Register / Login"] C --> E["Enter Key"] D --> F["Enter Endpoint & Key"] E --> G("Start Using") F --> G style A fill:#f9f9f9,stroke:#333,stroke-width:1px style B fill:#f9f9f9,stroke:#333,stroke-width:1px style C fill:#f9f9f9,stroke:#333,stroke-width:1px style D fill:#f9f9f9,stroke:#333,stroke-width:1px style E fill:#f9f9f9,stroke:#333,stroke-width:1px style F fill:#f9f9f9,stroke:#333,stroke-width:1px style G fill:#f9f9f9,stroke:#333,stroke-width:1px

Purchase Now

Start Chat on Homepage

Register / Login

Enter Key

Read API Documentation

Enter Endpoint & Key

Start Using

Description Ends

Recommend Models

o3-mini

o3-mini is our newest small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini supports key developer features, like Structured Outputs, function calling, and Batch API.

o3-pro

The o-series of models are trained with reinforcement learning to think before they answer and perform complex reasoning. The o3-pro model uses more compute to think harder and provide consistently better answers. o3-pro is available in the Responses API only to enable support for multi-turn model interactions before responding to API requests, and other advanced API features in the future. Since o3-pro is designed to tackle tough problems, some requests may take several minutes to finish. To avoid timeouts, try using background mode.

grok-4-fast-reasoning

We're excited to release grok-4-fast, xAI latest advancement in cost-efficient reasoning models. Contains two latest models, codenamed: grok-4-fast-reasoning and grok-4-fast-noreasoning.