Mistral AI's 7B parameter model with sliding window attention and strong performance across benchmarks.
Sparse mixture-of-experts model with 8 expert groups of 7B parameters each for efficient high-quality generation.
Mistral Small 24B optimized for low-latency workloads with strong instruction following and multilingual support.