openai

GPT-4o mini

GPT-4o mini is OpenAI’s most cost-efficient small model, offering GPT-4 class intelligence with superior speed and multimodal vision for high-volume tasks.

Small ModelCost-EfficientVision-CapableFast AIMultimodal
openai logoopenaiGPT-4July 18, 2024
Context
128Ktokens
Max Output
16Ktokens
Input Price
$0.15/ 1M
Output Price
$0.60/ 1M
Modality:TextImage
Capabilities:VisionToolsStreaming
Benchmarks
GPQA
40.2%
GPQA: Graduate-Level Science Q&A. A rigorous benchmark with 448 multiple-choice questions in biology, physics, and chemistry created by domain experts. PhD experts only achieve 65-74% accuracy, while non-experts score just 34% even with unlimited web access (hence 'Google-proof'). GPT-4o mini scored 40.2% on this benchmark.
HLE
2.3%
HLE: High-Level Expertise Reasoning. Tests a model's ability to demonstrate expert-level reasoning across specialized domains. Evaluates deep understanding of complex topics that require professional-level knowledge. GPT-4o mini scored 2.3% on this benchmark.
MMLU
82%
MMLU: Massive Multitask Language Understanding. A comprehensive benchmark with 16,000 multiple-choice questions across 57 academic subjects including math, philosophy, law, and medicine. Tests broad knowledge and reasoning capabilities. GPT-4o mini scored 82% on this benchmark.
MMLU Pro
57%
MMLU Pro: MMLU Professional Edition. An enhanced version of MMLU with 12,032 questions using a harder 10-option multiple choice format. Covers Math, Physics, Chemistry, Law, Engineering, Economics, Health, Psychology, Business, Biology, Philosophy, and Computer Science. GPT-4o mini scored 57% on this benchmark.
SimpleQA
8.6%
SimpleQA: Factual Accuracy Benchmark. Tests a model's ability to provide accurate, factual responses to straightforward questions. Measures reliability and reduces hallucinations in knowledge retrieval tasks. GPT-4o mini scored 8.6% on this benchmark.
IFEval
76.5%
IFEval: Instruction Following Evaluation. Measures how well a model follows specific instructions and constraints. Tests the ability to adhere to formatting rules, length limits, and other explicit requirements. GPT-4o mini scored 76.5% on this benchmark.
AIME 2025
5%
AIME 2025: American Invitational Math Exam. Competition-level mathematics problems from the prestigious AIME exam designed for talented high school students. Tests advanced mathematical problem-solving requiring abstract reasoning, not just pattern matching. GPT-4o mini scored 5% on this benchmark.
MATH
50.2%
MATH: Mathematical Problem Solving. A comprehensive math benchmark testing problem-solving across algebra, geometry, calculus, and other mathematical domains. Requires multi-step reasoning and formal mathematical knowledge. GPT-4o mini scored 50.2% on this benchmark.
GSM8k
91.3%
GSM8k: Grade School Math 8K. 8,500 grade school-level math word problems requiring multi-step reasoning. Tests basic arithmetic and logical thinking through real-world scenarios like shopping or time calculations. GPT-4o mini scored 91.3% on this benchmark.
MGSM
87%
MGSM: Multilingual Grade School Math. The GSM8k benchmark translated into 10 languages including Spanish, French, German, Russian, Chinese, and Japanese. Tests mathematical reasoning across different languages. GPT-4o mini scored 87% on this benchmark.
MathVista
42%
MathVista: Mathematical Visual Reasoning. Tests the ability to solve math problems that involve visual elements like charts, graphs, geometry diagrams, and scientific figures. Combines visual understanding with mathematical reasoning. GPT-4o mini scored 42% on this benchmark.
SWE-Bench
8.7%
SWE-Bench: Software Engineering Benchmark. AI models attempt to resolve real GitHub issues in open-source Python projects with human verification. Tests practical software engineering skills on production codebases. Top models went from 4.4% in 2023 to over 70% in 2024. GPT-4o mini scored 8.7% on this benchmark.
HumanEval
87.2%
HumanEval: Python Programming Problems. 164 hand-written programming problems where models must generate correct Python function implementations. Each solution is verified against unit tests. Top models now achieve 90%+ accuracy. GPT-4o mini scored 87.2% on this benchmark.
LiveCodeBench
32.5%
LiveCodeBench: Live Coding Benchmark. Tests coding abilities on continuously updated, real-world programming challenges. Unlike static benchmarks, uses fresh problems to prevent data contamination and measure true coding skills. GPT-4o mini scored 32.5% on this benchmark.
MMMU
59.4%
MMMU: Multimodal Understanding. Massive Multi-discipline Multimodal Understanding benchmark testing vision-language models on college-level problems across 30 subjects requiring both image understanding and expert knowledge. GPT-4o mini scored 59.4% on this benchmark.
MMMU Pro
34%
MMMU Pro: MMMU Professional Edition. Enhanced version of MMMU with more challenging questions and stricter evaluation. Tests advanced multimodal reasoning at professional and expert levels. GPT-4o mini scored 34% on this benchmark.
ChartQA
78%
ChartQA: Chart Question Answering. Tests the ability to understand and reason about information presented in charts and graphs. Requires extracting data, comparing values, and performing calculations from visual data representations. GPT-4o mini scored 78% on this benchmark.
DocVQA
88%
DocVQA: Document Visual Q&A. Document Visual Question Answering benchmark testing the ability to extract and reason about information from document images including forms, reports, and scanned text. GPT-4o mini scored 88% on this benchmark.
Terminal-Bench
40%
Terminal-Bench: Terminal/CLI Tasks. Tests the ability to perform command-line operations, write shell scripts, and navigate terminal environments. Measures practical system administration and development workflow skills. GPT-4o mini scored 40% on this benchmark.
ARC-AGI
8%
ARC-AGI: Abstraction & Reasoning. Abstraction and Reasoning Corpus for AGI - tests fluid intelligence through novel pattern recognition puzzles. Each task requires discovering the underlying rule from examples, measuring general reasoning ability rather than memorization. GPT-4o mini scored 8% on this benchmark.

About GPT-4o mini

Learn about GPT-4o mini's capabilities, features, and how it can help you achieve better results.

A New Standard for Small Models

GPT-4o mini represents a significant leap in AI efficiency, designed to replace GPT-3.5 Turbo as the go-to model for developers. Built with a native multimodal architecture, it delivers GPT-4 class performance at a fraction of the cost and latency. It features a massive 128,000 token context window and supports complex outputs of up to 16,384 tokens, making it ideal for processing long-form documents and high-volume data streams.

Intelligence Meets Affordability

Unlike previous small models that sacrificed intelligence for speed, GPT-4o mini maintains high reasoning capabilities across text and vision tasks. It is 60% cheaper than GPT-3.5 Turbo and significantly more capable, scoring 82% on the MMLU benchmark. This model is specifically optimized for applications where low latency and high reliability are paramount, such as real-time customer assistants and large-scale data classification engines.

GPT-4o mini

Use Cases for GPT-4o mini

Discover the different ways you can use GPT-4o mini to achieve great results.

Customer Support Automation

Deploy low-latency, cost-effective chatbots that can handle thousands of routine inquiries with high accuracy.

Structured Data Extraction

Convert large volumes of unstructured text or image data into clean, formatted JSON schemas for databases.

Real-Time Translation

Power messaging apps with near-instant translation across dozens of languages at minimal operational cost.

Document Summarization

Summarize long-form reports and meeting transcripts using the expansive 128k context window.

Visual Analysis & OCR

Extract text from images, analyze charts, and describe visual content for accessibility purposes.

Rapid Prototyping

Quickly build and iterate on AI applications with minimal API spend before scaling to larger models.

Strengths

Limitations

Unmatched Price-to-Performance: At $0.15/1M input tokens, it offers a 99% cost reduction compared to GPT-4 while maintaining high logic.
Limited Reasoning Depth: Does not possess the advanced chain-of-thought capabilities found in models like OpenAI o1.
Low-Latency Response Time: Engineered for speed, providing near-instant responses critical for real-time conversational agents.
Output Token Constraint: The 16,384 output limit is lower than specialized models that offer full 128k output parity.
Native Multimodal Vision: Unlike legacy small models, it processes images natively with high scores on ChartQA (78.0).
Complex Coding Gaps: Struggles with high-level architectural design and large multi-file repository debugging.
Strong Instruction Following: Achieves a 76.5% on IFEval, demonstrating superior capability in adhering to complex formatting rules.
No Native Audio Input: Does not support direct audio-to-audio processing through the standard text/vision API endpoint.

API Quick Start

openai/gpt-4o-mini

View Documentation
openai SDK
import OpenAI from "openai";

const openai = new OpenAI();

async function main() {
  const completion = await openai.chat.completions.create({
    model: "gpt-4o-mini",
    messages: [
      { role: "system", content: "You are a helpful assistant." },
      { role: "user", content: "Analyze this data and return JSON." },
    ],
    response_format: { type: "json_object" },
  });

  console.log(completion.choices[0].message.content);
}

main();

Install the SDK and start making API calls in minutes.

What People Are Saying About GPT-4o mini

See what the community thinks about GPT-4o mini

GPT-4o mini is basically the death knell for GPT-3.5. It's faster, cheaper, and smarter. There's no reason to use the old model anymore.
rowqi
reddit
OpenAI really changed the game with the pricing. 15 cents for a million tokens is insane for this level of logic.
peterm1598
hackernews
I'm seeing near-instant responses for my chatbot. The latency on 4o-mini is impressive for the price point.
WesternCat5211
twitter
GPT-4o mini launched 4 days ago. already processing more than 200B tokens per day!
Sam Altman
twitter
The vision performance for something this cheap is the real story. It handles charts better than some flagship models did a year ago.
VisionDev99
reddit
It follows system instructions perfectly for JSON extraction. A massive upgrade over 3.5 Turbo's reliability.
LLM_Master
hackernews

Videos About GPT-4o mini

Watch tutorials, reviews, and discussions about GPT-4o mini

The era of the small model is here

It is priced at 15 cents per million input tokens... that is insanely cheap

GPT-4o mini is very very fast... unbelievable fast

This model is basically replacing GPT-3.5 Turbo and it's 60% cheaper

Look at how fast this is generating text, it's almost instantaneous

GPT-4o mini it's free it's fast and it's actually pretty good

GPT-3.5 is completely gone... it brings you down to GPT-4o mini

Look at the speed of that way faster right

It is a multimodal model meaning it can see images

It outperforms GPT-3.5 Turbo and even GPT-4 on some benchmarks

OpenAI has basically dropped a competitor which seems to be beating a lot of these models

It's the model that they've optimized for charging the least amount

Having this being able to output 16,000 output tokens at a time really is a useful kind of thing

The context window is 128k, which is great for RAG applications

The safety measures on this model are built in from the start

More than just prompts

Supercharge your workflow with AI Automation

Automatio combines the power of AI agents, web automation, and smart integrations to help you accomplish more in less time.

AI Agents
Web Automation
Smart Workflows

Pro Tips for GPT-4o mini

Expert tips to help you get the most out of GPT-4o mini and achieve better results.

Direct System Instructions

Provide concise system prompts to steer the model towards specific output formats like JSON or Markdown.

Batching Context

Utilize the 128k window by grouping related small tasks into a single prompt to reduce overhead costs.

Vision Labeling

Ask for specific visual attributes when processing images to improve the precision of metadata extraction.

Streaming for UX

Always enable the stream parameter to provide immediate token feedback for end-users.

Temperature Tuning

Set temperature to 0 for structured data tasks to ensure deterministic and consistent outputs.

Testimonials

What Our Users Say

Join thousands of satisfied users who have transformed their workflow

Jonathan Kogan

Jonathan Kogan

Co-Founder/CEO, rpatools.io

Automatio is one of the most used for RPA Tools both internally and externally. It saves us countless hours of work and we realized this could do the same for other startups and so we choose Automatio for most of our automation needs.

Mohammed Ibrahim

Mohammed Ibrahim

CEO, qannas.pro

I have used many tools over the past 5 years, Automatio is the Jack of All trades.. !! it could be your scraping bot in the morning and then it becomes your VA by the noon and in the evening it does your automations.. its amazing!

Ben Bressington

Ben Bressington

CTO, AiChatSolutions

Automatio is fantastic and simple to use to extract data from any website. This allowed me to replace a developer and do tasks myself as they only take a few minutes to setup and forget about it. Automatio is a game changer!

Sarah Chen

Sarah Chen

Head of Growth, ScaleUp Labs

We've tried dozens of automation tools, but Automatio stands out for its flexibility and ease of use. Our team productivity increased by 40% within the first month of adoption.

David Park

David Park

Founder, DataDriven.io

The AI-powered features in Automatio are incredible. It understands context and adapts to changes in websites automatically. No more broken scrapers!

Emily Rodriguez

Emily Rodriguez

Marketing Director, GrowthMetrics

Automatio transformed our lead generation process. What used to take our team days now happens automatically in minutes. The ROI is incredible.

Jonathan Kogan

Jonathan Kogan

Co-Founder/CEO, rpatools.io

Automatio is one of the most used for RPA Tools both internally and externally. It saves us countless hours of work and we realized this could do the same for other startups and so we choose Automatio for most of our automation needs.

Mohammed Ibrahim

Mohammed Ibrahim

CEO, qannas.pro

I have used many tools over the past 5 years, Automatio is the Jack of All trades.. !! it could be your scraping bot in the morning and then it becomes your VA by the noon and in the evening it does your automations.. its amazing!

Ben Bressington

Ben Bressington

CTO, AiChatSolutions

Automatio is fantastic and simple to use to extract data from any website. This allowed me to replace a developer and do tasks myself as they only take a few minutes to setup and forget about it. Automatio is a game changer!

Sarah Chen

Sarah Chen

Head of Growth, ScaleUp Labs

We've tried dozens of automation tools, but Automatio stands out for its flexibility and ease of use. Our team productivity increased by 40% within the first month of adoption.

David Park

David Park

Founder, DataDriven.io

The AI-powered features in Automatio are incredible. It understands context and adapts to changes in websites automatically. No more broken scrapers!

Emily Rodriguez

Emily Rodriguez

Marketing Director, GrowthMetrics

Automatio transformed our lead generation process. What used to take our team days now happens automatically in minutes. The ROI is incredible.

Related AI Models

alibaba

Qwen3-Coder-Next

alibaba

Qwen3-Coder-Next is Alibaba Cloud's elite Apache 2.0 coding model, featuring an 80B MoE architecture and 256k context window for advanced local development.

256K context
$0.14/$0.42/1M
openai

GPT-5.4

OpenAI

GPT-5.4 is OpenAI's frontier model featuring a 1.05M context window and Extreme Reasoning. It excels at autonomous UI interaction and long-form data analysis.

1M context
$2.50/$15.00/1M
google

Gemini 3.1 Flash-Lite

Google

Gemini 3.1 Flash-Lite is Google's fastest, most cost-efficient model. Features 1M context, native multimodality, and 363 tokens/sec speed for scale.

1M context
$0.25/$1.50/1M
openai

GPT-5.3 Instant

OpenAI

Explore GPT-5.3 Instant, OpenAI's "Anti-Cringe" model. Features a 128K context window, 26.8% fewer hallucinations, and a natural, helpful tone for everyday...

128K context
$1.75/$14.00/1M
google

Gemini 3.1 Pro

Google

Gemini 3.1 Pro is Google's elite multimodal model featuring the DeepThink reasoning engine, a 1M+ context window, and industry-leading ARC-AGI logic scores.

1M context
$2.50/$15.00/1M
anthropic

Claude Sonnet 4.6

Anthropic

Claude Sonnet 4.6 offers frontier performance for coding and computer use with a massive 1M token context window for only $3/1M tokens.

1M context
$3.00/$15.00/1M
alibaba

Qwen3.5-397B-A17B

alibaba

Qwen3.5-397B-A17B is Alibaba's flagship open-weight MoE model. It features native multimodal reasoning, a 1M context window, and a 19x decoding throughput...

1M context
$0.60/$3.60/1M
minimax

MiniMax M2.5

minimax

MiniMax M2.5 is a SOTA MoE model featuring a 1M context window and elite agentic coding capabilities at disruptive pricing for autonomous agents.

1M context
$0.30/$1.20/1M

Frequently Asked Questions About GPT-4o mini

Find answers to common questions about GPT-4o mini