PFN's first LLM. 13B dense LLaMA-based Transformer trained on 1.5T tokens (1.32T English, 0.18T Japanese). 4K context. Apache 2.0. Established PFN's LLM capabilities and launched the PLaMo series.

Model Details

Architecture DENSE
Parameters 13B
Context window 4,096
open-weightmultilingual

Related