Sarashina2.2
modelCompute-efficient Sarashina iteration focused on small, high-quality Japanese models. Dense Llama-style Transformers at 0.5B / 1B / 3B with both base and instruction-tuned variants. MIT licensed. Sarashina2.2-3B-Instruct scores 3.75 on Elyza-tasks-100 (vs 2.99 for Qwen2.5-3B-Instruct), 6.51 on Japanese MT-Bench, and 7.71 on English MT-Bench, demonstrating that a small Japanese-specialized model can outperform larger Japanese-capable models on Japanese tasks.
The 2.2 generation was extended into multimodal variants: Sarashina2.2-Vision-3B (image→text, November 2025) and Sarashina2.2-OCR (4B, ~March 2026, Japanese-document OCR).
Outputs 3
Sarashina2.2 (0.5B / 1B / 3B)
modelSmall Japanese LLM family with base + instruct variants at 0.5B, 1B, 3B. MIT.
Variants
| Name | Parameters | Notes |
|---|---|---|
| Sarashina2.2-0.5B | 0.5B | — |
| Sarashina2.2-1B | 1B | — |
| Sarashina2.2-3B | 3B | — |
Sarashina2.2-Vision-3B
modelVision-language variant of Sarashina2.2-3B. MIT.
Sarashina2.2-OCR
model4B OCR-specialized variant of the Sarashina2.2 family for Japanese documents.