OpenAI logo

OpenAI

o3 Mini

OpenAI o3-mini is a cost-efficient language model optimized for STEM reasoning tasks, particularly excelling in science, mathematics, and coding. This model supports the `reasoning_effort` parameter, which can be set to "high", "medium", or "low" to control the thinking time of the model. The default is "medium". OpenRouter also offers the model slug `openai/o3-mini-high` to default the parameter to "high". The model features three adjustable reasoning effort levels and supports key developer capabilities including function calling, structured outputs, and streaming, though it does not include vision processing capabilities. The model demonstrates significant improvements over its predecessor, with expert testers preferring its responses 56% of the time and noting a 39% reduction in major errors on complex questions. With medium reasoning effort settings, o3-mini matches the performance of the larger o1 model on challenging reasoning evaluations like AIME and GPQA, while maintaining lower latency and cost.

Input / 1M tokens
$1.10
Output / 1M tokens
$4.40
Context window
200K tokens
Provider
OpenAI
Cached input / 1M
$0.550
Knowledge cutoff
2023-10-31

Performance

Median streaming throughput and first-token latency measured by Artificial Analysis.

Output tokens / sec
164 t/s
Time to first token
8.46s

Benchmarks

Intelligence, coding, and math indexes plus the underlying evaluation scores.

Intelligence Index
26
Coding Index
18
Math Index
MMLU-Pro
79.1%
GPQA
74.8%
HLE
8.7%
LiveCodeBench
71.7%
SciCode
39.9%
MATH-500
97.3%
AIME
77.0%

Benchmarks via Artificial Analysis