glm-4.6
As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications.
| Provider | Source | Input Price ($/1M) | Output Price ($/1M) | Description | Free |
|---|---|---|---|---|---|
| vercel | vercel | Input: $0.45 | Output: $1.80 | As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications. | |
| poe | poe | Input: $6,600.00 | Output: - | As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications. Use `--enable_thinking false` to disable thinking about the response before giving a final answer. This is enabled by default. Bot does not support media (video and audio file) attachments. Technical Specifications File Support: Text, Markdown and PDF files Context window: 200k tokens | |
| siliconflowcn | models-dev | Input: $0.50 | Output: $1.90 | Provider: SiliconFlow (China), Context: 205000, Output Limit: 205000 | |
| togetherai | models-dev | Input: $0.60 | Output: $2.20 | Provider: Together AI, Context: 200000, Output Limit: 32768 | |
| baseten | models-dev | Input: $0.60 | Output: $2.20 | Provider: Baseten, Context: 200000, Output Limit: 200000 | |
| siliconflow | models-dev | Input: $0.50 | Output: $1.90 | Provider: SiliconFlow, Context: 205000, Output Limit: 205000 | |
| helicone | models-dev | Input: $0.45 | Output: $1.50 | Provider: Helicone, Context: 204800, Output Limit: 131072 | |
| huggingface | models-dev | Input: $0.60 | Output: $2.20 | Provider: Hugging Face, Context: 200000, Output Limit: 128000 | |
| opencode | models-dev | Input: $0.60 | Output: $2.20 | Provider: OpenCode Zen, Context: 204800, Output Limit: 131072 | |
| zenmux | models-dev | Input: $0.35 | Output: $1.54 | Provider: ZenMux, Context: 200000, Output Limit: 128000 | |
| iflowcn | models-dev | Input: $0.00 | Output: $0.00 | Provider: iFlow, Context: 200000, Output Limit: 128000 | |
| synthetic | models-dev | Input: $0.55 | Output: $2.19 | Provider: Synthetic, Context: 200000, Output Limit: 64000 | |
| zhipuai | models-dev | Input: $0.60 | Output: $2.20 | Provider: Zhipu AI, Context: 204800, Output Limit: 131072 | |
| nanogpt | models-dev | Input: $1.00 | Output: $2.00 | Provider: NanoGPT, Context: 200000, Output Limit: 8192 | |
| friendli | models-dev | Input: - | Output: - | Provider: Friendli, Context: 131072, Output Limit: 131072 | |
| ionet | models-dev | Input: $0.40 | Output: $1.75 | Provider: IO.NET, Context: 200000, Output Limit: 4096 | |
| modelscope | models-dev | Input: $0.00 | Output: $0.00 | Provider: ModelScope, Context: 202752, Output Limit: 98304 | |
| openrouter | openrouter | Input: $0.35 | Output: $1.50 | Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks. Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Code、Cline、Roo Code and Kilo Code, including improvements in generating visually polished front-end pages. Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability. More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks. Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios. Context: 202752 | |
| factoryai | factoryai | Input: - | Output: - | - | |
| zai | zai | Input: $0.60 | Output: $0.11 | - |