百度:2026年文心5.0技术报告(英文版)
百度:2026年文心5.0技术报告(英文版).pdf |
下载文档 |
资源简介
In this report, we introduce ERNIE 5.0, a natively autoregressive foundation model desinged for unified multimodal understanding and generation across text, image, video, and audio. All modalities are trained from scratch under a unified next-groupof-tokens prediction objective, based on an ultra-sparse mixture-of-experts (MoE) architecture with modality-agnostic expert routing. To address practical challenges in large-scale deployment under diverse resource constraints, ERNIE 5.0 adopts a no
本文档仅能预览20页


