Cohere Tiny Aya: 3B-Parameter Multilingual Model Outperforms Larger Competitors in 46 of 61 Languages

Cohere’s Tiny Aya is a 3.35B-parameter model family covering 70 languages. With 4-bit quantization it fits in 2.14 GB, runs at 10 tokens per second on an iPhone 13, and beats Gemma3-4B in translation quality for 46 of 61 languages on WMT24++.
artificial-intelligence
Author

Kabui, Charles

Published

2026-02-22

Keywords

cohere, tiny-aya, multilingual-ai, small-language-model, on-device-ai, fusion-of-n, simmerge, edge-computing, quantization