gpt-4.1-nano
GPT-4.1 nano is the fastest, most cost-effective GPT 4.1 model.
| Provider | Source | Input Price ($/1M) | Output Price ($/1M) | Description | Free |
|---|---|---|---|---|---|
| vercel | vercel | Input: $0.10 | Output: $0.40 | GPT-4.1 nano is the fastest, most cost-effective GPT 4.1 model. | |
| poe | poe | Input: $0.09 | Output: $0.36 | GPT-4.1 nano is an extremely fast and cheap model, ideal for text/vision summarization/categorization tasks. Supports native vision and 1M input tokens of context. Check out the newest version of this bot here: https://poe.com/GPT-5-nano. | |
| abacus | models-dev | Input: $0.10 | Output: $0.40 | Provider: Abacus, Context: 1047576, Output Limit: 32768 | |
| githubmodels | models-dev | Input: $0.00 | Output: $0.00 | Provider: GitHub Models, Context: 128000, Output Limit: 16384 | |
| azure | models-dev | Input: $0.10 | Output: $0.40 | Provider: Azure, Context: 1047576, Output Limit: 32768 | |
| helicone | models-dev | Input: $0.10 | Output: $0.40 | Provider: Helicone, Context: 1047576, Output Limit: 32768 | |
| openai | models-dev | Input: $0.10 | Output: $0.40 | Provider: OpenAI, Context: 1047576, Output Limit: 32768 | |
| aihubmix | models-dev | Input: $0.10 | Output: $0.40 | Provider: AIHubMix, Context: 1047576, Output Limit: 32768 | |
| azurecognitiveservices | models-dev | Input: $0.10 | Output: $0.40 | Provider: Azure Cognitive Services, Context: 1047576, Output Limit: 32768 | |
| openrouter | openrouter | Input: $0.10 | Output: $0.40 | For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion. Context: 1047576 |