GLM-4.5 API, Deployment, Pricing

zai-org/GLM-4.5

GLM-4.5 is a foundational model specifically designed for AI agent applications, built on a Mixture-of-Experts (MoE) architecture. It has been extensively optimized for tool use, web browsing, software development, and front-end development, enabling seamless integration with coding agents such as Claude Code and Roo Code. GLM-4.5 employs a hybrid reasoning approach, allowing it to adapt effectively to a wide range of application scenarios—from complex reasoning tasks to everyday use cases

API Usage

curl --request POST \
  --url https://api.siliconflow.com/v1/chat/completions \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "model": "zai-org/GLM-4.5",
  "messages": [
    {
      "role": "user",
      "content": "Tell me a story"
    }
  ],
  "top_p": 0.95,
  "temperature": 0.6
}'

Details

Model Provider

zai

Type

text

Sub Type

chat

Size

335B

Publish Time

Jul 28, 2025

Input Price

$

0.5

/ M Tokens

Output Price

$

2

/ M Tokens

Context length

131K

Tags

Reasoning,MoE,335B,131K

Compare with Other Models

See how this model stacks up against others.

Model FAQs: Usage, Deployment

Learn how to use, fine-tune, and deploy this model with ease.

What is the GLM-4.5 model, and what are its core capabilities and technical specifications?

In which business scenarios does GLM-4.5 perform well? Which industries or applications is it suitable for?

How can the performance and effectiveness of GLM-4.5 be optimized in actual business use?

Compared with other models, when should GLM-4.5 be selected?

What are SiliconFlow's key strengths in AI serverless deployment for GLM-4.5?

What makes SiliconFlow the top platform for GLM-4.5 API?

Ready to accelerate your AI development?

Ready to accelerate your AI development?

© 2025 SiliconFlow Technology PTE. LTD.

© 2025 SiliconFlow Technology PTE. LTD.

© 2025 SiliconFlow Technology PTE. LTD.