Core Insights - DeepSeek has released its latest mathematical model, DeepSeek Math-V2, which has generated significant excitement in the AI community due to its self-verifying capabilities in deep reasoning, particularly in mathematics [1][2]. Model Performance - Math-V2 demonstrates strong theorem-proving abilities, distinguishing itself from previous models that merely solved problems without rigorous reasoning [2]. - The model achieved gold medal-level results in the IMO 2025 and CMO 2024 competitions, and scored 118 out of 120 in the Putnam 2024 competition, showcasing its superior performance [2]. Benchmarking Results - In the IMO-Proof Bench evaluation, Math-V2 scored 99%, outperforming Google's Gemini Deep Think (89%) and GPT-5 (59%) [3]. - In advanced testing, Math-V2 scored 61.9%, just behind Gemini Deep Think's 65.7% [3]. Community Impact - The release of Math-V2 has sparked discussions across social media platforms and communities, highlighting its potential to automate verification-heavy tasks in programming languages [5][8]. - Experts in the AI field have praised DeepSeek's return and the significance of Math-V2, indicating a shift from "chatbot" to "reasoner" era in AI development [8][9].
不只是“做题家”!DeepSeek最新模型打破数学推理局限,部分性能超越Gemini DeepThink