Have you ever tried to use a state-of-the-art AI model in a language other than English, only to find the performance lackluster or the latency unbearable? It’s a common frustration, especially if you aren’t running enterprise-grade hardware. But on February 17, 2026, Cohere took a significant swing at fixing this disparity with the launch of ‘Tiny Aya.’
This isn’t just another massive model release designed for a server farm. Cohere For AI, the company’s non-profit research arm, has unveiled a family of 3.35 billion parameter open weights models. The headline feature? They support over 70 languages. This release marks a strategic pivot for the lab, moving from the depth-focused ‘Aya 23’ released back in May 2024 to a breadth-focused approach that tries to squeeze high-performance multilingual capabilities into a package small enough to run on consumer hardware.
What makes the Tiny Aya family different from previous releases?
If you’ve been following Cohere’s trajectory, you might remember the ‘Aya Expanse’ launch in late 2024, which featured larger 8B and 32B models. Tiny Aya goes in the opposite direction. By shrinking the parameter count to 3.35 billion, Cohere is explicitly targeting the ‘Small Language Model’ (SLM) market.
The release isn’t a monolith; it’s a family of four distinct variants:
Tiny Aya Global
Tiny Aya Earth
Tiny Aya Fire
Tiny Aya Water
While the specific technical nuances between ‘Fire’ and ‘Water’ are part of the deeper documentation, the overarching goal is clear: optimization. These models are designed for edge deployment and low-resource environments. They feature an 8k context window, which is substantial for models of this size, allowing them to handle decent chunks of text without needing a massive cloud tether.
This is a direct evolution of the work started by Sara Hooker, the former lead who initiated the Aya project, and is now being carried forward by Marzieh Fadaee and CEO Aidan Gomez. They are essentially trying to prove that you don’t need a trillion parameters to speak Hindi, Arabic, or Yoruba fluently.
How does Tiny Aya stack up against the competition?
The timing of this release places Tiny Aya in a crowded and competitive arena. The industry has been obsessed with SLMs lately, with major players realizing that efficiency is just as important as raw intelligence. Tiny Aya is stepping into the ring against heavyweights like Google’s Gemma 2B and Meta’s Llama 3.2 (specifically the 1B and 3B variants).
However, Cohere has a specific edge: multilingualism. While models like Llama 3.2 are incredibly capable, they have historically leaned heavily toward English and major European languages. By optimizing Tiny Aya for over 70 languages, Cohere is positioning this family as the go-to choice for developers in the Global South and for applications requiring broad linguistic coverage on edge devices.
According to Cohere For AI, the models are ‘optimized for efficient, strong, and balanced multilingual representation across 70+ languages.’ This focus on “balanced” representation is key—it suggests an attempt to mitigate the quality drop-off usually seen when smaller models try to learn less-resourced languages.
What are the licensing and usage constraints?
For developers eager to download these weights, there is an important caveat regarding usage rights. Tiny Aya is released under a CC-BY-NC (Creative Commons Non-Commercial) license. This means researchers, hobbyists, and non-profits can use the models freely, provided they adhere to Cohere’s Acceptable Use Policy.
This licensing structure aligns with the ‘Cohere For AI’ mission as a research lab. It allows for widespread experimentation and academic work without giving away the commercial farm to competitors. It also follows the precedent set by previous Aya releases, ensuring that the community contributing to these open science projects sees the benefits, while maintaining a boundary for enterprise commercialization.
Between the Lines
Analysis: This release signals a maturation in the ‘Small Language Model’ market. By releasing four distinct 3B parameter variants, Cohere is acknowledging that one size does not fit all, even at the small scale. The winners here are developers in regions with expensive data costs or limited hardware access, who finally get a modern, multilingual model that runs locally. The losers are the hyperscalers who profit from API calls; as models like Tiny Aya become capable enough for edge devices, the need to round-trip every query to a massive data center diminishes. Cohere is effectively democratizing the ‘inference’ layer of AI, betting that accessibility will drive the next wave of adoption.