Home/Models/GLM 4 32B
z-ai logo

GLM 4 32B

by z-ai · Released Jul 2025

37.8
avg score
Rank #156
Compare
Better than 33% of all models
Context
128K tokens (~64 books)
Input $/1M
$0.10
Output $/1M
$0.10
Type
text
License
Proprietary
Benchmarks
6 tested
Data updated today
About

GLM 4 32B is a cost-effective foundation language model. It can efficiently perform complex tasks and has significantly enhanced capabilities in tool use, online search, and code-related intelligent tasks. It...

Tested on 6 benchmarks with 18.0% average. Top scores: BBH (HuggingFace) (35.8%), MMLU-PRO (34.9%), IFEval (14.3%).

Looking for similar performance at lower cost?
Mistral Nemo scores 37.4 (99% as good) at $0.02/1M input · 80% cheaper
Capabilities
reasoning
14.2
#114 globally
math
0.0
#220 globally
knowledge
21.9
#193 globally
general
35.8
#25 globally
language
14.3
#149 globally
Benchmark Scores
Compare All
Tested on 6 benchmarks · Ranked across 5 categories
Score Distribution (all 233 models)
0255075100
▲ You are here
MUSR

HuggingFace MuSR (Multi-Step Reasoning). Tests multi-hop reasoning requiring chaining multiple facts together.

14.2
MATH Level 5

HuggingFace evaluation of MATH Level 5 problems. Competition math requiring advanced reasoning and proof construction.

0.0
MMLU-PRO

HuggingFace MMLU-Pro. Harder version of MMLU with 10 answer choices instead of 4 and more challenging questions.

34.9
GPQA

HuggingFace evaluation of GPQA (Graduate-Level Google-Proof Q&A). PhD-level science questions that cannot be easily searched.

8.8
Excellent (85+) Good (70-85) Average (50-70) Below (<50)
Links
Documentation
Community
BenchGecko API
glm-4-32b
Specifications
  • Typetext
  • Context128K tokens (~64 books)
  • ReleasedJul 2025
  • LicenseProprietary
  • StatusActive
  • Cost / Message~$0.000
Available On
z-ai logoz-ai$0.10
Share & Export
Tweet
GLM 4 32B is a proprietary text AI model by z-ai, released in July 2025. It has an average benchmark score of 37.8. Context window: 128K tokens.