Workflow
Seek .(SKLTY)
icon
Search documents
DeepSeek发布V3.2正式版
Xin Jing Bao· 2025-12-01 15:01
Core Insights - DeepSeek announced the release of two official model versions: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale [1] Model Overview - DeepSeek-V3.2 aims to balance reasoning capability and output length, making it suitable for everyday use, such as Q&A scenarios and general agent tasks [1] - In benchmark tests for reasoning, DeepSeek-V3.2 achieved performance comparable to GPT-5, slightly below Gemini-3.0-Pro [1] - Compared to Kimi-K2-Thinking, V3.2 significantly reduced output length, leading to lower computational costs and reduced user wait times [1] Special Features - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, exploring the boundaries of model performance [1] - This version is an enhanced long-thinking variant of DeepSeek-V3.2, incorporating theorem-proving capabilities from DeepSeek-Math-V2 [1] - The model exhibits excellent instruction-following, rigorous mathematical proof, and logical verification abilities, performing comparably to Gemini-3.0-Pro in mainstream reasoning benchmark tests [1]
DeepSeek,上新
Core Insights - DeepSeek has released two new models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aimed at enhancing reasoning capabilities and output length for various applications [1][2]. Model Performance - DeepSeek-V3.2 achieved performance comparable to GPT-5 and slightly below Gemini-3.0-Pro in public reasoning benchmarks, while significantly reducing output length compared to Kimi-K2-Thinking, thus lowering computational costs and user wait times [1][3]. - The DeepSeek-V3.2-Speciale model demonstrated exceptional instruction-following, rigorous mathematical proof, and logical validation capabilities, achieving gold medal-level performance in major competitions such as IMO 2025 and ICPC World Finals 2025 [2][3]. Benchmark Comparisons - In various benchmark tests, DeepSeek-V3.2-Speciale outperformed standard versions and other models, with notable scores in AIME 2025 (96.0) and HMMT Feb 2025 (99.2), while also achieving high rankings in IMOAnswerBench and LiveCodeBench [3]. - The performance of DeepSeek-V3.2-Speciale in complex tasks was significantly better than the standard version, although it required more tokens, indicating higher operational costs [3]. Model Features - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation, enhancing its versatility [4]. - The model has improved generalization capabilities through a novel large-scale agent training data synthesis method, allowing it to perform well in real-world applications [4].
DeepSeek 重要发布
Core Insights - DeepSeek has officially released two models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, with updates available on the official website, app, and API [1] - DeepSeek-V3.2 aims to balance reasoning capabilities and output length, making it suitable for everyday use cases such as Q&A and general agent tasks [1] - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, enhancing long-thinking abilities and incorporating theorem-proving capabilities from DeepSeek-Math-V2 [1] Model Performance - The V3.2-Speciale model exhibits excellent instruction-following, rigorous mathematical proof, and logical verification capabilities, performing comparably to leading international models on mainstream reasoning benchmarks [1] - Notably, the V3.2-Speciale model has achieved gold medals in several prestigious competitions, including IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025 [1] - In the ICPC and IOI competitions, the model's performance reached the level of the second and tenth place among human competitors, respectively [1]
DeepSeek,又有大动作!
Core Insights - DeepSeek has launched two new models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aiming to enhance reasoning capabilities and output length for various applications [1][2][3] Model Features - DeepSeek-V3.2 is designed for everyday use, balancing reasoning ability and output length, and has achieved performance comparable to GPT-5 in benchmark tests [2][3] - DeepSeek-V3.2-Speciale enhances long reasoning capabilities and incorporates theorem proving abilities from DeepSeek-Math-V2, excelling in complex tasks but requiring more tokens and higher costs [3][4] Technological Advancements - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation, significantly improving generalization capabilities [4] - The model has been trained on over 1,800 environments and 85,000 complex instructions, narrowing the performance gap between open-source and closed-source models [4] Market Outlook - The AI industry is experiencing a resonance period, with rapid expansion in AI infrastructure and commercialization of downstream applications, expected to continue thriving through 2026 [5][6][7] - Investment opportunities are identified in domestic AI chains, overseas AI hardware markets, and innovative applications in the domestic edge AI industry [7]
DeepSeek又上新!模型硬刚谷歌,承认开源与闭源差距拉大
Di Yi Cai Jing· 2025-12-01 13:31
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which are leading in reasoning capabilities globally [1][3]. Model Overview - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for everyday use such as Q&A and general intelligence tasks. It has reached the level of GPT-5 in public reasoning tests, slightly below Google's Gemini3 Pro [3]. - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the extreme, combining features from DeepSeek-Math-V2 for theorem proving, and excels in instruction following and logical verification [3][4]. Performance Metrics - Speciale has surpassed Google's Gemini3 Pro in several reasoning benchmark tests, including the American Mathematics Invitational, Harvard MIT Mathematics Competition, and International Mathematical Olympiad [4]. - In various benchmarks, DeepSeek's performance is competitive, with specific scores noted in a comparative table against GPT-5 and Gemini-3.0 [5]. Technical Limitations - Despite achievements, DeepSeek acknowledges limitations compared to proprietary models like Gemini3 Pro, particularly in knowledge breadth and token efficiency [6]. - The company plans to enhance pre-training computation and optimize reasoning chains to improve model efficiency and capabilities [6][7]. Mechanism Innovations - DeepSeek introduced a Sparse Attention Mechanism (DSA) to reduce computational complexity, which has proven effective in enhancing performance without sacrificing long-context capabilities [7][8]. - Both new models incorporate this mechanism, making DeepSeek-V3.2 a cost-effective alternative that narrows the performance gap with proprietary models [8]. Community Reception - The release has been positively received in the community, with users noting that DeepSeek's models are now comparable to GPT-5 and Gemini3 Pro, marking a significant achievement in open-source model development [8].
AI进化速递 | DeepSeek发布新模型
Di Yi Cai Jing· 2025-12-01 12:48
Group 1 - DeepSeek V3.2 has been officially released, enhancing Agent capabilities and integrating reasoning and thinking [1] - Doubao Mobile Assistant has released a technical preview version [1] - Tsinghua University has established a research institute for embodied intelligence and robotics [1] Group 2 - Didi's autonomous driving service is undergoing trial operations in Guangzhou, offering all-weather, fully unmanned Robotaxi services [2] - HSBC has formed a strategic partnership with Mistral AI to enhance the application of generative AI in banking operations [2]
DeepSeek V3.2 正式版发布:性能比肩GPT-5 ,略低于 Gemini-3.0-Pro
Xin Lang Ke Ji· 2025-12-01 11:23
Core Insights - DeepSeek has officially released two models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, following the experimental version launched two months ago [1] - DeepSeek-V3.2 aims to balance reasoning capability and output length, making it suitable for everyday use, such as Q&A scenarios and general agent tasks [1] - The performance of DeepSeek-V3.2 in benchmark tests is comparable to GPT-5 and slightly lower than Gemini-3.0-Pro, with significantly reduced output length compared to Kimi-K2-Thinking, leading to lower computational costs and reduced user wait times [1] Model Specifications - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, serving as an enhanced version of DeepSeek-V3.2 with theorem-proving abilities from DeepSeek-Math-V2 [2] - This model excels in instruction following, rigorous mathematical proofs, and logical validation, achieving performance on par with Gemini-3.0-Pro in mainstream reasoning benchmarks [2] - DeepSeek-V3.2-Speciale has won gold medals in prestigious competitions such as IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025, with ICPC and IOI scores reaching the second and tenth positions among human competitors, respectively [2] - While the Speciale model significantly outperforms the standard version in complex tasks, it consumes more tokens and incurs higher costs [2] - Currently, DeepSeek-V3.2-Speciale is available only for research purposes and does not support tool invocation, nor has it been optimized for everyday conversation and writing tasks [2]
DeepSeek发布两个正式版模型
Core Insights - DeepSeek has released two official model versions: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale [1] - The main goal of DeepSeek-V3.2 is to balance reasoning capability with output length, making it suitable for everyday use cases such as Q&A and general agent tasks [1] - The DeepSeek-V3.2-Speciale version aims to push the reasoning capabilities of open-source models to the extreme, exploring the boundaries of model capabilities [1] Summary by Categories - **Product Launch** - DeepSeek has updated its official website, app, and API to the official version of DeepSeek-V3.2 [1] - The Speciale version is currently available only as a temporary API service for community evaluation and research [1] - **Model Objectives** - DeepSeek-V3.2 is designed for daily applications, focusing on practical scenarios like Q&A and general agent tasks [1] - DeepSeek-V3.2-Speciale is focused on maximizing the reasoning capabilities of the model, aiming to explore its limits [1]
DeepSeekV3.2正式版发布 强化Agent能力 融入思考推理
Hua Er Jie Jian Wen· 2025-12-01 11:11
风险提示及免责条款 市场有风险,投资需谨慎。本文不构成个人投资建议,也未考虑到个别用户特殊的投资目标、财务状况或需要。用户应考虑本文中的任何 意见、观点或结论是否符合其特定状况。据此投资,责任自负。 DeepSeek宣布同时发布两个正式版模型:DeepSeek-V3.2 和 DeepSeek-V3.2-Speciale。DeepSeek-V3.2 的 目标是平衡推理能力与输出长度,适合日常使用,例如问答场景和通用 Agent 任务场景。DeepSeek- V3.2-Speciale 的目标是将开源模型的推理能力推向极致,探索模型能力的边界。 ...
AI周报 | DeepSeek开源奥数金牌水平模型;前OpenAI 联创称规模扩展时代已终结
Di Yi Cai Jing· 2025-11-30 00:48
Group 1: DeepSeek's New Model - DeepSeek has open-sourced a new model, DeepSeek-Math-V2, which is the first open-source model to reach IMO gold medal level in mathematics [1] - The performance of Math-V2 surpasses that of Google's Gemini DeepThink in certain aspects, as demonstrated in the IMO-ProofBench benchmark and recent math competitions [1] Group 2: AI Scaling Era Conclusion - Ilya Sutskever, CEO of Safe Superintelligence, claims that the era of AI scaling has ended, indicating a shift back to research paradigms rather than mere expansion [2] - He emphasizes that the current computational power cannot continuously yield better scaling, blurring the line between scaling and waste [2] Group 3: Baidu's AI Department Restructuring - Baidu has established two new AI departments: the Basic Model R&D Department and the Application Model R&D Department, both reporting directly to CEO Li Yanhong [3] - The restructuring reflects Baidu's commitment to enhancing its R&D capabilities in large models, with leadership from internally cultivated talents [3] Group 4: Nvidia's Response to Short Selling - Nvidia responded to Michael Burry's claims about the minimal real demand for AI products, clarifying that its strategic investments represent a small portion of its revenue [4] - Following a significant drop in Nvidia's stock price, the company aims to prove the sustained strength of AI demand [4] Group 5: Google's AI Glasses Project - Google is accelerating its new AI glasses project, with hardware manufacturing by Foxconn and chip supply from Qualcomm, expected to enter small-scale production [6] - The project is independent of the previously announced AR glasses and is led by a key figure from Google Labs [6] Group 6: HSBC's Warning on OpenAI's Profitability - HSBC forecasts that OpenAI will face severe financial pressure over the next decade, predicting it will struggle to achieve profitability even with a projected revenue of $213 billion by 2030 [7] - The analysis highlights the significant cash flow deficit OpenAI may encounter, amounting to $207 billion [7] Group 7: Industrial Fulian's Performance Clarification - Industrial Fulian clarified rumors regarding a downward adjustment of its Q4 performance targets, stating that operations are proceeding as planned [8] - The company's stock experienced fluctuations, reflecting market concerns about its relationship with Nvidia [8] Group 8: Denial of Google Order by Tianfu Communication - Tianfu Communication denied rumors of securing a $3 billion order from Google, amidst speculation about its role as a supplier [9] - The stock prices of related companies fluctuated based on market interest in optical module stocks [9] Group 9: Meta's Interest in Google's TPU - Meta is reportedly considering a multi-billion dollar purchase of Google's TPU for its data center development, which could mark the first external sale of Google's TPU [10] - This potential shift could impact Nvidia, as Meta is currently its largest GPU customer [10] Group 10: AI's Water Consumption - A Morgan Stanley report highlights that AI not only consumes significant electricity but also requires substantial water resources for data center operations [11] - The report points out the challenges of water resource allocation for AI data centers, particularly in regions facing water supply issues [12]