AI大模型训练

Search documents
下一只“寒王”呼之欲出!算力+机器人共振,英伟达核心伙伴潜力股
Xin Lang Cai Jing· 2025-10-08 04:16
当寒武纪的崛起还让人记忆犹新时, 《全球数智化指数2025》报告震撼发布:到2035年,全社会算力总量将增长10万倍!这一预测正在科技 圈和金融圈引发巨震。 可以说算力就是AI时代的生产力核心 2025年,中国智能算力规模将达到1037.3EFLOPS,较2024年增长43%;2026年,这一数字将达 1460.3EFLOPS,实现两年翻倍增长! 并且,全球主要经济体均将算力视为战略资源。 欧盟启动《欧洲芯片法案》计划2030年占全球20%市场份额。 算力竞争已成为国际科技博弈的关键领域。 需求呈指数级爆发,多领域同步引爆。 从AI大模型训练到自动驾驶,从智慧城市到工业机器人,甚至军工领域,算力正成为制胜关键: 例如,智能汽车:单车算力需求年复合增长超50%; 卫星互联网:全球低轨卫星数据处理需求激增; 第5家:紫光股份 旗下紫光晓通是NVIDIA企业级产品的总代理,以NVIDIA企业级产品为核心,为客户提供"计算+网络 +存储+安全+备份+AI软件"的全栈解决方案 第4家:英维克 向IDC公司提供机房的液冷系统,同时提供服务器内部的液冷模组,代表客户包括华为、英伟达等。 第3家:工业富联 工业4.0:智能制 ...
微信WeChat-YATT横空出世,腾讯强化学习布局剑指何方
Sou Hu Cai Jing· 2025-09-24 09:56
Core Insights - Tencent's open-sourcing of WeChat-YATT training library signifies a strategic move in the competitive landscape of AI model training, particularly as OpenAI's GPT-5 approaches release [1][2] - WeChat-YATT is designed with a focus on reinforcement learning and multimodal models, differentiating itself from mainstream frameworks like TensorFlow and PyTorch [2] Group 1: WeChat-YATT's Innovations - WeChat-YATT achieves significant breakthroughs in three areas: optimized parameter update efficiency for reinforcement learning, flexible multimodal data fusion interfaces, and a modular design that lowers the barriers for distributed training [2][4] - The library's emphasis on "ease of extensibility" reflects Tencent's recognition of the need for rapid iteration in large model training [4] Group 2: Competitive Positioning - Compared to Meta's PyTorch, WeChat-YATT excels in reinforcement learning support; against Google's JAX, it shows advantages in Chinese language scenarios and multimodal processing [4] - WeChat-YATT's deep integration with the WeChat ecosystem sets it apart from similar reinforcement learning frameworks like Ray RLlib [4] Group 3: Strategic Implications - The release of WeChat-YATT aligns with Tencent's broader AI strategy, which includes trademark applications for "WeChat AI Service Platform" and the deployment of the mixed Yuan model in business scenarios [7] - Tencent aims to create a closed-loop AI ecosystem through foundational technology breakthroughs and application deployment, with WeChat-YATT serving as a critical component in this strategy [7] - The focus on reinforcement learning indicates Tencent's commitment to key areas such as gaming, recommendation systems, and autonomous driving, positioning itself for future AI applications [7] Group 4: Long-term Vision - The naming of WeChat-YATT, "Yet Another Transformer Trainer," reflects both a sense of humor and Tencent's long-term investment in AI infrastructure [6] - The competition in the era of large models is fundamentally a competition for infrastructure, with WeChat-YATT representing a piece of Tencent's broader AI blueprint [7]
提升大模型通信性能30% DeepSeek致谢腾讯大模型网络提速技术方案贡献
Shen Zhen Shang Bao· 2025-05-11 22:32
Core Insights - Tencent's technical team has optimized the DeepEP communication framework, achieving significant performance improvements in various network environments, with a 100% enhancement in RoCE and a 30% enhancement in IB networks, facilitating more efficient AI large model training solutions [2][3] - The optimization addresses key bottlenecks in the original DeepEP framework, particularly in bandwidth utilization and CPU control delays, which were limiting its broader application [2][3] Group 1 - The optimization includes intelligent bandwidth allocation through topology-aware multi-QP chaining technology, ensuring full utilization of dual-port network card bandwidth and preventing bandwidth waste [3] - Tencent has resolved CPU control bottlenecks in GPU communication by optimizing the control plane operations to bypass CPU intermediaries, reducing latency and energy consumption [3] - A new "QP internal sequencing lock" mechanism has been introduced to ensure accurate and sequential data transmission among multiple GPUs, even when handling over 1,000 simultaneous data transfer tasks [3] Group 2 - The optimized DeepEP framework has been fully open-sourced and successfully applied in Tencent's mixed Yuan large model training and inference projects, demonstrating excellent versatility in high-performance environments built with Tencent's Xingmai and H20 servers [3]
DeepSeek致谢腾讯技术团队:对DeepEP的优化,是一次“huge speedup”代码贡献
Xin Lang Ke Ji· 2025-05-07 11:12
Core Insights - Tencent's technical team has optimized the DeepEP communication framework, achieving significant performance improvements across various network environments, with a 100% performance increase in RoCE networks and a 30% increase in IB networks, enhancing AI large model training efficiency [1][2] Group 1: Technical Enhancements - The optimization involved replacing IBRC with IBGDA and utilizing distinct Queue Pairs (QPs) per channel for parallel data transmission, which improved the robustness and communication performance of the normal kernels [1] - The algorithm bandwidth for the optimized framework reached 58 GB/s in RDMA scenarios, with physical bandwidth calculated at 43.5 GB/s [1] Group 2: Industry Impact - Since the open-sourcing of DeepSeek, including DeepEP, in February, the framework has demonstrated a 300% increase in communication efficiency, addressing the dependency on NVIDIA NCCL for MoE architecture large models [2] - The optimizations have been successfully applied in Tencent's mixed Yuan model projects, showcasing excellent versatility in high-performance environments built with Tencent's Starry Network and H20 servers [2]
技术驱动与绿色转型双轮并进,润泽科技一季报稳健增长
Zheng Quan Shi Bao Wang· 2025-04-29 04:08
Core Insights - The company reported a revenue of 1.198 billion yuan and a net profit of 430 million yuan for Q1 2025, indicating healthy financial metrics [1] - As a leading provider of intelligent computing infrastructure in China, the company is leveraging technological innovation and green development to build a future-oriented computing foundation [1] - The company has established seven AIDC intelligent computing clusters across key economic regions, with all delivered and upcoming computing centers having secured production orders, expected to be operational by 2025 [1] Technological Developments - The company is deepening the commercialization of liquid cooling technology, having delivered the industry's first fully liquid-cooled green computing center in 2023 [1] - The Power Usage Effectiveness (PUE) of the liquid-cooled computing centers has been reduced to approximately 1.15, showcasing significant energy efficiency [1] - The company is enhancing energy-saving renovations in existing computing centers and has achieved industry-leading PUE levels in its Langfang park, supporting AI model training with reliable and efficient computing infrastructure [1] Green Development Strategy - The company is actively promoting a "low-carbon green" process for its computing centers, with its A-7 and A-18 centers recognized as national green data centers due to their excellent energy-saving performance [2] - In 2024, the company completed a total of 800 million kilowatt-hours in green electricity transactions, emphasizing its commitment to energy-saving technology research and green transformation [2] Strategic Expansion - The company's strategic layout in Hainan Free Trade Port aligns with national policies, as the State Council approved the establishment of cross-border e-commerce comprehensive pilot zones in Hainan and other cities [3] - The company is constructing an intelligent computing infrastructure cluster in Danzhou, Hainan, with a planned capacity of approximately 30,000 cabinets, aimed at enhancing cross-border operations [3] - This initiative supports the digital economy development directive outlined in the Hainan Free Trade Port construction plan and lays the groundwork for the company to expand into overseas markets [3]