← Back to all models

GLM 4.6

glm-4.6

As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications.

Available at 20 Providers

Provider Source Input Price ($/1M) Output Price ($/1M) Description Free
vercel vercel Input: $0.45 Output: $1.80 As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications.
poe poe Input: $6,600.00 Output: - As the latest iteration in the GLM series, GLM-4.6 achieves comprehensive enhancements across multiple domains, including real-world coding, long-context processing, reasoning, searching, writing, and agentic applications. Use `--enable_thinking false` to disable thinking about the response before giving a final answer. This is enabled by default. Bot does not support media (video and audio file) attachments. Technical Specifications File Support: Text, Markdown and PDF files Context window: 200k tokens
siliconflowcn models-dev Input: $0.50 Output: $1.90 Provider: SiliconFlow (China), Context: 205000, Output Limit: 205000
togetherai models-dev Input: $0.60 Output: $2.20 Provider: Together AI, Context: 200000, Output Limit: 32768
baseten models-dev Input: $0.60 Output: $2.20 Provider: Baseten, Context: 200000, Output Limit: 200000
siliconflow models-dev Input: $0.50 Output: $1.90 Provider: SiliconFlow, Context: 205000, Output Limit: 205000
helicone models-dev Input: $0.45 Output: $1.50 Provider: Helicone, Context: 204800, Output Limit: 131072
huggingface models-dev Input: $0.60 Output: $2.20 Provider: Hugging Face, Context: 200000, Output Limit: 128000
opencode models-dev Input: $0.60 Output: $2.20 Provider: OpenCode Zen, Context: 204800, Output Limit: 131072
zenmux models-dev Input: $0.35 Output: $1.54 Provider: ZenMux, Context: 200000, Output Limit: 128000
iflowcn models-dev Input: $0.00 Output: $0.00 Provider: iFlow, Context: 200000, Output Limit: 128000
synthetic models-dev Input: $0.55 Output: $2.19 Provider: Synthetic, Context: 200000, Output Limit: 64000
zhipuai models-dev Input: $0.60 Output: $2.20 Provider: Zhipu AI, Context: 204800, Output Limit: 131072
nanogpt models-dev Input: $1.00 Output: $2.00 Provider: NanoGPT, Context: 200000, Output Limit: 8192
friendli models-dev Input: - Output: - Provider: Friendli, Context: 131072, Output Limit: 131072
ionet models-dev Input: $0.40 Output: $1.75 Provider: IO.NET, Context: 200000, Output Limit: 4096
modelscope models-dev Input: $0.00 Output: $0.00 Provider: ModelScope, Context: 202752, Output Limit: 98304
openrouter openrouter Input: $0.35 Output: $1.50 Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks. Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Code、Cline、Roo Code and Kilo Code, including improvements in generating visually polished front-end pages. Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability. More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks. Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios. Context: 202752
factoryai factoryai Input: - Output: - -
zai zai Input: $0.60 Output: $0.11 -