Mistral logo

Mistral

Mistral Nemo

A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It supports function calling and is released under the Apache 2.0 license.

Input / 1M tokens
$0.020
Output / 1M tokens
$0.040
Context window
131K tokens
Provider
Mistral
Knowledge cutoff
2024-04-30

Performance

Median streaming throughput and first-token latency measured by Artificial Analysis.

Output tokens / sec
Time to first token