gpt-4.1-mini
GPT 4.1 mini provides a balance between intelligence, speed, and cost that makes it an attractive model for many use cases.
| Provider | Source | Input Price ($/1M) | Output Price ($/1M) | Description | Free |
|---|---|---|---|---|---|
| vercel | vercel | Input: $0.40 | Output: $1.60 | GPT 4.1 mini provides a balance between intelligence, speed, and cost that makes it an attractive model for many use cases. | |
| poe | poe | Input: $0.36 | Output: $1.40 | GPT-4.1 mini is a small, fast & affordable model that matches or beats GPT-4o in many intelligence and vision-related tasks. Supports 1M tokens of context. Check out the newest version of this bot here: https://poe.com/GPT-5-mini. | |
| abacus | models-dev | Input: $0.40 | Output: $1.60 | Provider: Abacus, Context: 1047576, Output Limit: 32768 | |
| githubmodels | models-dev | Input: $0.00 | Output: $0.00 | Provider: GitHub Models, Context: 128000, Output Limit: 16384 | |
| azure | models-dev | Input: $0.40 | Output: $1.60 | Provider: Azure, Context: 1047576, Output Limit: 32768 | |
| helicone | models-dev | Input: $0.40 | Output: $1.60 | Provider: Helicone, Context: 1047576, Output Limit: 32768 | |
| openai | models-dev | Input: $0.40 | Output: $1.60 | Provider: OpenAI, Context: 1047576, Output Limit: 32768 | |
| requesty | models-dev | Input: $0.40 | Output: $1.60 | Provider: Requesty, Context: 1047576, Output Limit: 32768 | |
| aihubmix | models-dev | Input: $0.40 | Output: $1.60 | Provider: AIHubMix, Context: 1047576, Output Limit: 32768 | |
| azurecognitiveservices | models-dev | Input: $0.40 | Output: $1.60 | Provider: Azure Cognitive Services, Context: 1047576, Output Limit: 32768 | |
| openrouter | openrouter | Input: $0.40 | Output: $1.60 | GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints. Context: 1047576 |