A

Allenai

Olmo 3.1 32B Instruct

Olmo 3.1 32B Instruct is a large-scale, 32-billion-parameter instruction-tuned language model engineered for high-performance conversational AI, multi-turn dialogue, and practical instruction following. As part of the Olmo 3.1 family, this variant emphasizes responsiveness to complex user directions and robust chat interactions while retaining strong capabilities on reasoning and coding benchmarks. Developed by Ai2 under the Apache 2.0 license, Olmo 3.1 32B Instruct reflects the Olmo initiative’s commitment to openness and transparency.

Input / 1M tokens
$0.200
Output / 1M tokens
$0.600
Context window
66K tokens
Provider
Allenai

Performance

Median streaming throughput and first-token latency measured by Artificial Analysis.

Output tokens / sec
54 t/s
Time to first token
0.29s

Benchmarks

Intelligence, coding, and math indexes plus the underlying evaluation scores.

Intelligence Index
12
Coding Index
6
Math Index
MMLU-Pro
GPQA
53.9%
HLE
4.9%
LiveCodeBench
SciCode
16.7%
MATH-500
AIME

Benchmarks via Artificial Analysis