Workflow
IcePop
icon
Search documents
蚂蚁Ring-1T正式登场,万亿参数思考模型,数学能力对标IMO银牌
机器之心· 2025-10-14 06:33
Core Insights - Ant Group has launched the Ling-1T and Ring-1T models, marking significant advancements in open-source AI with capabilities comparable to closed-source giants [3][6][19] - The Ring-1T model is the first open-source trillion-parameter reasoning model, showcasing exceptional performance in various benchmarks and tasks [6][9][19] Model Launch and Performance - Ant Group announced the Ling-1T model on October 9, which is their largest language model to date, achieving over a thousand downloads within four days of its release [3][5] - Following this, the Ring-1T model was officially launched on October 14, demonstrating superior reasoning abilities and achieving notable results in international mathematics competitions [6][19] Benchmark Testing - The Ring-1T model underwent rigorous testing across eight critical benchmarks, including mathematics competitions, code generation, and logical reasoning [12][14] - Results indicate that Ring-1T significantly outperformed its preview version, achieving state-of-the-art (SOTA) performance in multiple dimensions, particularly in complex reasoning tasks [9][14][16] Competitive Analysis - In logical reasoning tasks, Ring-1T surpassed the performance of leading closed-source models like Gemini-2.5-Pro, showcasing its competitive edge [16] - The model's performance in the Arena-Hard-v2.0 comprehensive ability test was just slightly behind GPT-5-Thinking, placing it among the top-tier models in the industry [16] Practical Applications - Ring-1T demonstrated its coding capabilities by generating functional game code for simple games like Flappy Bird and Snake, showcasing its practical application in software development [20][23] - The model also excelled in creative writing, producing engaging narratives and scripts that incorporate historical facts and storytelling techniques [40][43] Technical Innovations - The development of Ring-1T involved advanced reinforcement learning techniques, particularly the IcePop algorithm, which mitigates training inconsistencies and enhances model stability [45][46] - Ant Group's self-developed RL framework, ASystem, supports the efficient training of large-scale models, addressing hardware resource challenges and improving training consistency [50][52]