← Back to all models

GLM 4.5

glm-4.5

GLM-4.5 and GLM-4.5-Air are our latest flagship models, purpose-built as foundational models for agent-oriented applications. Both leverage a Mixture-of-Experts (MoE) architecture. GLM-4.5 has a total parameter count of 355B with 32B active parameters per forward pass, while GLM-4.5-Air adopts a more streamlined design with 106B total parameters and 12B active parameters.

Available at 14 Providers

Provider Source Input Price ($/1M) Output Price ($/1M) Description Free
vercel vercel Input: $0.60 Output: $2.20 GLM-4.5 and GLM-4.5-Air are our latest flagship models, purpose-built as foundational models for agent-oriented applications. Both leverage a Mixture-of-Experts (MoE) architecture. GLM-4.5 has a total parameter count of 355B with 32B active parameters per forward pass, while GLM-4.5-Air adopts a more streamlined design with 106B total parameters and 12B active parameters.
poe poe Input: $5,700.00 Output: - The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities to meet the complex demands of intelligent agent applications. Technical Specifications File Support: PDF and Markdown files Context window: 128k tokens
nebius models-dev Input: $0.60 Output: $2.20 Provider: Nebius Token Factory, Context: 131072, Output Limit: 8192
siliconflowcn models-dev Input: $0.40 Output: $2.00 Provider: SiliconFlow (China), Context: 131000, Output Limit: 131000
siliconflow models-dev Input: $0.40 Output: $2.00 Provider: SiliconFlow, Context: 131000, Output Limit: 131000
huggingface models-dev Input: $0.60 Output: $2.20 Provider: Hugging Face, Context: 131072, Output Limit: 98304
zenmux models-dev Input: $0.35 Output: $1.54 Provider: ZenMux, Context: 128000, Output Limit: 64000
synthetic models-dev Input: $0.55 Output: $2.19 Provider: Synthetic, Context: 128000, Output Limit: 96000
deepinfra models-dev Input: $0.60 Output: $2.20 Provider: Deep Infra, Context: 131072, Output Limit: 98304
zhipuai models-dev Input: $0.60 Output: $2.20 Provider: Zhipu AI, Context: 131072, Output Limit: 98304
modelscope models-dev Input: $0.00 Output: $0.00 Provider: ModelScope, Context: 131072, Output Limit: 98304
wandb litellm Input: $55,000.00 Output: $200,000.00 Source: wandb, Context: 131072
openrouter openrouter Input: $0.35 Output: $1.55 GLM-4.5 is our latest flagship foundation model, purpose-built for agent-based applications. It leverages a Mixture-of-Experts (MoE) architecture and supports a context length of up to 128k tokens. GLM-4.5 delivers significantly enhanced capabilities in reasoning, code generation, and agent alignment. It supports a hybrid inference mode with two options, a "thinking mode" designed for complex reasoning and tool use, and a "non-thinking mode" optimized for instant responses. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) Context: 131072
zai zai Input: $0.60 Output: $0.11 -