Mistral Saba is a relatively small model with 24 billion parameters. As a reminder, fewer parameters generally leads to better performance with lower latency. But more parameters usually means smarter answers, even though it’s not a linear correlation.
Mistral Saba is comparable in size to Mistral Small 3, its general purpose small model. But, according to Mistral’s own tests, Mistral Saba performs much better than Mistral Small 3 when handling Arabic content.