glm-4.5
GLM-4.5 and GLM-4.5-Air are our latest flagship models, purpose-built as foundational models for agent-oriented applications. Both leverage a Mixture-of-Experts (MoE) architecture. GLM-4.5 has a total parameter count of 355B with 32B active parameters per forward pass, while GLM-4.5-Air adopts a more streamlined design with 106B total parameters and 12B active parameters.
| Provider | Source | Input Price ($/1M) | Output Price ($/1M) | Description | Free |
|---|---|---|---|---|---|
| vercel | vercel | Input: $0.60 | Output: $2.20 | GLM-4.5 and GLM-4.5-Air are our latest flagship models, purpose-built as foundational models for agent-oriented applications. Both leverage a Mixture-of-Experts (MoE) architecture. GLM-4.5 has a total parameter count of 355B with 32B active parameters per forward pass, while GLM-4.5-Air adopts a more streamlined design with 106B total parameters and 12B active parameters. | |
| poe | poe | Input: $5,700.00 | Output: - | The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities to meet the complex demands of intelligent agent applications. Technical Specifications File Support: PDF and Markdown files Context window: 128k tokens | |
| nebius | models-dev | Input: $0.60 | Output: $2.20 | Provider: Nebius Token Factory, Context: 131072, Output Limit: 8192 | |
| siliconflowcn | models-dev | Input: $0.40 | Output: $2.00 | Provider: SiliconFlow (China), Context: 131000, Output Limit: 131000 | |
| siliconflow | models-dev | Input: $0.40 | Output: $2.00 | Provider: SiliconFlow, Context: 131000, Output Limit: 131000 | |
| huggingface | models-dev | Input: $0.60 | Output: $2.20 | Provider: Hugging Face, Context: 131072, Output Limit: 98304 | |
| zenmux | models-dev | Input: $0.35 | Output: $1.54 | Provider: ZenMux, Context: 128000, Output Limit: 64000 | |
| synthetic | models-dev | Input: $0.55 | Output: $2.19 | Provider: Synthetic, Context: 128000, Output Limit: 96000 | |
| deepinfra | models-dev | Input: $0.60 | Output: $2.20 | Provider: Deep Infra, Context: 131072, Output Limit: 98304 | |
| zhipuai | models-dev | Input: $0.60 | Output: $2.20 | Provider: Zhipu AI, Context: 131072, Output Limit: 98304 | |
| modelscope | models-dev | Input: $0.00 | Output: $0.00 | Provider: ModelScope, Context: 131072, Output Limit: 98304 | |
| wandb | litellm | Input: $55,000.00 | Output: $200,000.00 | Source: wandb, Context: 131072 | |
| openrouter | openrouter | Input: $0.35 | Output: $1.55 | GLM-4.5 is our latest flagship foundation model, purpose-built for agent-based applications. It leverages a Mixture-of-Experts (MoE) architecture and supports a context length of up to 128k tokens. GLM-4.5 delivers significantly enhanced capabilities in reasoning, code generation, and agent alignment. It supports a hybrid inference mode with two options, a "thinking mode" designed for complex reasoning and tool use, and a "non-thinking mode" optimized for instant responses. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) Context: 131072 | |
| zai | zai | Input: $0.60 | Output: $0.11 | - |