Workflow
Llama 3.1 405B大语言模型
icon
Search documents
云端算力竞赛新突破:创纪录英伟达GB200参与MLPerf测试,性能提升超两倍
硬AI· 2025-06-05 10:32
Core Viewpoint - The collaboration between CoreWeave, NVIDIA, and IBM has achieved the largest scale MLPerf Training v5.0 test result in history, utilizing 2496 GB200 Grace Blackwell chips, showcasing the capabilities of CoreWeave's AI-optimized cloud platform [2][3] Group 1: MLPerf Training v5.0 Test Results - The GB200 NVL72 cluster completed the training of the largest and most complex Llama 3.1 405B model in just 27.3 minutes, achieving over two times the training performance compared to similar scale clusters [3] - This test result highlights the significant performance leap provided by the GB200 NVL72 architecture and CoreWeave's strong infrastructure for consistent and high-performance AI workloads [3] Group 2: Industry Participation and Growth - The MLPerf Training v5.0 test received a record total of 201 performance test submissions from 20 organizations, indicating a new high in industry participation [6] - The introduction of the new Llama 3.1 405B large language model (LLM) pre-training benchmark has replaced the previous GPT-3 based benchmarks, reflecting the growing importance of large-scale training [5][6] - The participation of new organizations such as AMD, IBM, MangoBoost, Nebius, and SCITIX in the MLPerf Training tests has been particularly welcomed, emphasizing the increasing focus on energy efficiency in AI training systems [6]