High performance in a 24b open-source model

Mistral Small 3

Mistral Small 3 is the most efficient and versatile model of Mistral. Pre-trained and instructed version, Apache 2.0, 24B, 81% MMLU, 150 token/s. No synthetic data so great base for anything reasoning.

Discover more from NextBigWhat

Subscribe now to keep reading and get access to the full archive.

Continue reading