GLM-4.1V-9B-Thinking
About GLM-4.1V-9B-Thinking
GLM-4.1V-9B-Thinking is an open-source Vision-Language Model (VLM) jointly released by Zhipu AI and Tsinghua University's KEG lab, designed to advance general-purpose multimodal reasoning. Built upon the GLM-4-9B-0414 foundation model, it introduces a 'thinking paradigm' and leverages Reinforcement Learning with Curriculum Sampling (RLCS) to significantly enhance its capabilities in complex tasks. As a 9B-parameter model, it achieves state-of-the-art performance among models of a similar size, and its performance is comparable to or even surpasses the much larger 72B-parameter Qwen-2.5-VL-72B on 18 different benchmarks. The model excels in a diverse range of tasks, including STEM problem-solving, video understanding, and long document understanding, and it can handle images with resolutions up to 4K and arbitrary aspect ratios
Available Serverless
Run queries immediately, pay only for usage
$
0.035
/
$
0.14
Per 1M Tokens (input/output)
Metadata
Specification
State
Available
Architecture
Calibrated
No
Mixture of Experts
No
Total Parameters
9B
Activated Parameters
9B
Reasoning
No
Precision
FP8
Context length
66K
Max Tokens
66K
Supported Functionality
Serverless
Supported
Serverless LoRA
Not supported
Fine-tuning
Not supported
Embeddings
Not supported
Rerankers
Not supported
Support image input
Supported
JSON Mode
Not supported
Structured Outputs
Not supported
Tools
Not supported
Fim Completion
Not supported
Chat Prefix Completion
Not supported
SiliconFlow Service
Comprehensive solutions to deploy and scale your AI applications with maximum flexibility
60%
lower latency
2x
higher throughput
65%
cost savings
Compare with Other Models
See how this model stacks up against others.

Z.ai
chat
GLM-4.6
Release on: Oct 4, 2025
Total Context:
205K
Max output:
205K
Input:
$
0.5
/ M Tokens
Output:
$
1.9
/ M Tokens

Z.ai
chat
GLM-4.5
Release on: Jul 28, 2025
Total Context:
131K
Max output:
131K
Input:
$
0.4
/ M Tokens
Output:
$
2.0
/ M Tokens

Z.ai
chat
GLM-4.5-Air
Release on: Jul 28, 2025
Total Context:
131K
Max output:
131K
Input:
$
0.14
/ M Tokens
Output:
$
0.86
/ M Tokens

Z.ai
chat
GLM-4.5V
Release on: Aug 13, 2025
Total Context:
66K
Max output:
66K
Input:
$
0.14
/ M Tokens
Output:
$
0.86
/ M Tokens

Z.ai
chat
GLM-4.1V-9B-Thinking
Release on: Jul 4, 2025
Total Context:
66K
Max output:
66K
Input:
$
0.035
/ M Tokens
Output:
$
0.14
/ M Tokens

Z.ai
chat
GLM-Z1-32B-0414
Release on: Apr 18, 2025
Total Context:
131K
Max output:
131K
Input:
$
0.14
/ M Tokens
Output:
$
0.57
/ M Tokens

Z.ai
chat
GLM-4-32B-0414
Release on: Apr 18, 2025
Total Context:
33K
Max output:
33K
Input:
$
0.27
/ M Tokens
Output:
$
0.27
/ M Tokens

Z.ai
chat
GLM-Z1-9B-0414
Release on: Apr 18, 2025
Total Context:
131K
Max output:
131K
Input:
$
0.086
/ M Tokens
Output:
$
0.086
/ M Tokens

Z.ai
chat
GLM-4-9B-0414
Release on: Apr 18, 2025
Total Context:
33K
Max output:
33K
Input:
$
0.086
/ M Tokens
Output:
$
0.086
/ M Tokens
Model FAQs: Usage, Deployment
Learn how to use, fine-tune, and deploy this model with ease.
