ERNIE 5.0
model paperMost advanced ERNIE model to date. 2.4 trillion total parameters with ultra-sparse MoE (<3% activation ratio) for extreme inference efficiency. Native full-modality unified modeling (text, image, audio, video) in a single autoregressive framework.
Outputs 2
ERNIE 5.0
modelMost advanced model to date. 2.4 trillion total parameters with ultra-sparse MoE (<3% activation ratio) for extreme inference efficiency. Native full-modality unified modeling (text, image, audio, video) in a single autoregressive framework.
Architecture MOE
Parameters 2.4T
Previewed at Baidu World 2025 on November 13, 2025. Official release January 22, 2026.
ERNIE 5.0 Technical Report
paperTrillion-parameter unified autoregressive multimodal model with elastic MoE training. 437 co-authors.
arXiv: 2602.04705