PLaMo-13B
modelPFN's first LLM. 13B dense LLaMA-based Transformer trained on 1.5T tokens (1.32T English, 0.18T Japanese). 4K context. Apache 2.0. Established PFN's LLM capabilities and launched the PLaMo series.
Model Details
Architecture DENSE
Parameters 13B
Context window 4,096