← Back to all models

GPT-4.1 nano

gpt-4.1-nano

GPT-4.1 nano is the fastest, most cost-effective GPT 4.1 model.

Available at 10 Providers

Provider Source Input Price ($/1M) Output Price ($/1M) Description Free
vercel vercel Input: $0.10 Output: $0.40 GPT-4.1 nano is the fastest, most cost-effective GPT 4.1 model.
poe poe Input: $0.09 Output: $0.36 GPT-4.1 nano is an extremely fast and cheap model, ideal for text/vision summarization/categorization tasks. Supports native vision and 1M input tokens of context. Check out the newest version of this bot here: https://poe.com/GPT-5-nano.
abacus models-dev Input: $0.10 Output: $0.40 Provider: Abacus, Context: 1047576, Output Limit: 32768
githubmodels models-dev Input: $0.00 Output: $0.00 Provider: GitHub Models, Context: 128000, Output Limit: 16384
azure models-dev Input: $0.10 Output: $0.40 Provider: Azure, Context: 1047576, Output Limit: 32768
helicone models-dev Input: $0.10 Output: $0.40 Provider: Helicone, Context: 1047576, Output Limit: 32768
openai models-dev Input: $0.10 Output: $0.40 Provider: OpenAI, Context: 1047576, Output Limit: 32768
aihubmix models-dev Input: $0.10 Output: $0.40 Provider: AIHubMix, Context: 1047576, Output Limit: 32768
azurecognitiveservices models-dev Input: $0.10 Output: $0.40 Provider: Azure Cognitive Services, Context: 1047576, Output Limit: 32768
openrouter openrouter Input: $0.10 Output: $0.40 For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion. Context: 1047576