OpenAI
o3 Mini High
OpenAI o3-mini-high is the same model as [o3-mini](/openai/o3-mini) with reasoning_effort set to high. o3-mini is a cost-efficient language model optimized for STEM reasoning tasks, particularly excelling in science, mathematics, and coding. The model features three adjustable reasoning effort levels and supports key developer capabilities including function calling, structured outputs, and streaming, though it does not include vision processing capabilities. The model demonstrates significant improvements over its predecessor, with expert testers preferring its responses 56% of the time and noting a 39% reduction in major errors on complex questions. With medium reasoning effort settings, o3-mini matches the performance of the larger o1 model on challenging reasoning evaluations like AIME and GPQA, while maintaining lower latency and cost.
- Input / 1M tokens
- $1.10
- Output / 1M tokens
- $4.40
- Context window
- 200K tokens
- Provider
- OpenAI
- Cached input / 1M
- $0.550
- Knowledge cutoff
- 2023-10-31
Performance
Median streaming throughput and first-token latency measured by Artificial Analysis.
- Output tokens / sec
- 150 t/s
- Time to first token
- 31.39s
Benchmarks
Intelligence, coding, and math indexes plus the underlying evaluation scores.
- Intelligence Index
- 25
- Coding Index
- 17
- Math Index
- —
- MMLU-Pro
- 80.2%
- GPQA
- 77.3%
- HLE
- 12.3%
- LiveCodeBench
- 73.4%
- SciCode
- 39.8%
- MATH-500
- 98.5%
- AIME
- 86.0%
Benchmarks via Artificial Analysis