Aya 101
modelLandmark multilingual model covering 101 languages (50+ previously underserved), built through a global collaboration of 3,000+ researchers from 119 countries. 13B parameters, fine-tuned on mT5. Published at ACL 2024.
Accompanied by the Aya Dataset: 204K human-annotated instruction pairs and 513M augmented instances across 114 languages. Apache 2.0. Pioneered the open, community-driven approach to multilingual AI that defined Cohere Labs' identity.
Model Details
Architecture DENSE
Parameters 13B
Paper
arXiv: 2402.07827
Venue: ACL 2024