qwen3-next-80b
The Qwen3-Next-80B is the next-generation foundation model released by Qwen, optimized for extreme context length and large-scale parameter efficiency, also known as "Qwen3-Next-80B-A3B." Despite its ultra-efficiency, it outperforms Qwen3-32B on downstream tasks - while requiring less than 1/10 of the training cost. Moreover, it delivers over 10x higher inference throughput than Qwen3-32B when handling contexts longer than 32k tokens. Use `--enable_thinking false` to disable thinking mode before giving an answer. This is the non-thinking version of https://poe.com/Qwen3-Next-80B-Think; supports 65k tokens of context.
| Provider | Source | Input Price ($/1M) | Output Price ($/1M) | Description | Free |
|---|---|---|---|---|---|
| poe | poe | Input: $2,400.00 | Output: - | The Qwen3-Next-80B is the next-generation foundation model released by Qwen, optimized for extreme context length and large-scale parameter efficiency, also known as "Qwen3-Next-80B-A3B." Despite its ultra-efficiency, it outperforms Qwen3-32B on downstream tasks - while requiring less than 1/10 of the training cost. Moreover, it delivers over 10x higher inference throughput than Qwen3-32B when handling contexts longer than 32k tokens. Use `--enable_thinking false` to disable thinking mode before giving an answer. This is the non-thinking version of https://poe.com/Qwen3-Next-80B-Think; supports 65k tokens of context. | |
| venice | models-dev | Input: $0.35 | Output: $1.90 | Provider: Venice AI, Context: 262144, Output Limit: 65536 |