DeepSeek
Search documents
DeepSeek 重要发布
Shang Hai Zheng Quan Bao· 2025-12-01 13:57
Core Insights - DeepSeek has officially released two models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, with updates available on the official website, app, and API [1] - DeepSeek-V3.2 aims to balance reasoning capabilities and output length, making it suitable for everyday use cases such as Q&A and general agent tasks [1] - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, enhancing long-thinking abilities and incorporating theorem-proving capabilities from DeepSeek-Math-V2 [1] Model Performance - The V3.2-Speciale model exhibits excellent instruction-following, rigorous mathematical proof, and logical verification capabilities, performing comparably to leading international models on mainstream reasoning benchmarks [1] - Notably, the V3.2-Speciale model has achieved gold medals in several prestigious competitions, including IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025 [1] - In the ICPC and IOI competitions, the model's performance reached the level of the second and tenth place among human competitors, respectively [1]
DeepSeek,又有大动作!
Bei Jing Ri Bao Ke Hu Duan· 2025-12-01 13:32
Core Insights - DeepSeek has launched two new models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aiming to enhance reasoning capabilities and output length for various applications [1][2][3] Model Features - DeepSeek-V3.2 is designed for everyday use, balancing reasoning ability and output length, and has achieved performance comparable to GPT-5 in benchmark tests [2][3] - DeepSeek-V3.2-Speciale enhances long reasoning capabilities and incorporates theorem proving abilities from DeepSeek-Math-V2, excelling in complex tasks but requiring more tokens and higher costs [3][4] Technological Advancements - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation, significantly improving generalization capabilities [4] - The model has been trained on over 1,800 environments and 85,000 complex instructions, narrowing the performance gap between open-source and closed-source models [4] Market Outlook - The AI industry is experiencing a resonance period, with rapid expansion in AI infrastructure and commercialization of downstream applications, expected to continue thriving through 2026 [5][6][7] - Investment opportunities are identified in domestic AI chains, overseas AI hardware markets, and innovative applications in the domestic edge AI industry [7]
DeepSeek又上新!模型硬刚谷歌,承认开源与闭源差距拉大
Di Yi Cai Jing· 2025-12-01 13:31
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which are leading in reasoning capabilities globally [1][3]. Model Overview - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for everyday use such as Q&A and general intelligence tasks. It has reached the level of GPT-5 in public reasoning tests, slightly below Google's Gemini3 Pro [3]. - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the extreme, combining features from DeepSeek-Math-V2 for theorem proving, and excels in instruction following and logical verification [3][4]. Performance Metrics - Speciale has surpassed Google's Gemini3 Pro in several reasoning benchmark tests, including the American Mathematics Invitational, Harvard MIT Mathematics Competition, and International Mathematical Olympiad [4]. - In various benchmarks, DeepSeek's performance is competitive, with specific scores noted in a comparative table against GPT-5 and Gemini-3.0 [5]. Technical Limitations - Despite achievements, DeepSeek acknowledges limitations compared to proprietary models like Gemini3 Pro, particularly in knowledge breadth and token efficiency [6]. - The company plans to enhance pre-training computation and optimize reasoning chains to improve model efficiency and capabilities [6][7]. Mechanism Innovations - DeepSeek introduced a Sparse Attention Mechanism (DSA) to reduce computational complexity, which has proven effective in enhancing performance without sacrificing long-context capabilities [7][8]. - Both new models incorporate this mechanism, making DeepSeek-V3.2 a cost-effective alternative that narrows the performance gap with proprietary models [8]. Community Reception - The release has been positively received in the community, with users noting that DeepSeek's models are now comparable to GPT-5 and Gemini3 Pro, marking a significant achievement in open-source model development [8].
DeepSeek V3.2 正式版发布,V4 还没来,但已经是开源模型里 Agent 能力最强了
Founder Park· 2025-12-01 13:14
Core Insights - DeepSeek has released the official version of its V3.2 model, which significantly enhances reasoning and agent capabilities compared to previous versions [2][9] - The V3.2-Speciale version is an open-source model that performs comparably to Gemini-3.0-Pro on mainstream reasoning benchmarks and has achieved gold medal levels in several prestigious competitions [3][11] - The integration of the DeepSeek Sparse Attention (DSA) technology in V3.2 improves long text processing efficiency and reduces costs by over 50% [3][10] Model Development - The V3 series has been iterated over the past year, with V3.2 being the latest release, focusing on unifying thinking and non-thinking models, a trend seen in other closed-source models like Gemini and GPT-5 [6][9] - The release timeline for DeepSeek models in 2025 includes various versions, each with specific enhancements, such as the introduction of DSA in V3.2 for stability and reasoning improvements [7][8] Performance Metrics - DeepSeek-V3.2 has achieved reasoning capabilities on par with GPT-5 and has shown significant improvements in output length and computational efficiency compared to Kimi-K2-Thinking [10][14] - The V3.2-Speciale version excels in complex tasks, achieving high scores in various academic competitions, including IMO 2025 and ICPC 2025, with notable rankings among human competitors [11][14] Tool Utilization - A key advancement in V3.2 is the incorporation of thinking processes into tool calls, allowing the model to support both thinking and non-thinking modes in its operations [15][18] - DeepSeek has developed a large-scale agent training data synthesis method that enhances the model's generalization capabilities by creating numerous "hard-to-answer, easy-to-verify" tasks [16][18]
DeepSeek,重大突发!
券商中国· 2025-12-01 13:01
Core Viewpoint - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aiming to enhance reasoning capabilities and application in various scenarios [1][2]. Model Features - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for daily use such as Q&A and general agent tasks. It has achieved performance comparable to GPT-5 in benchmark tests, slightly below Gemini-3.0-Pro [2]. - DeepSeek-V3.2-Speciale is an enhanced version of V3.2, integrating theorem proving capabilities from DeepSeek-Math-V2, excelling in instruction following and logical verification. It has won gold medals in several prestigious competitions, including IMO 2025 and ICPC World Finals 2025 [3]. Technological Advancements - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation. The model has been trained on over 85,000 complex instructions across 1,800 environments, significantly improving its generalization ability [4]. Market Outlook - The AI industry is entering a period of resonance, with rapid expansion in AI infrastructure and commercialization of downstream applications. Analysts predict continued prosperity in the AI sector, with a focus on domestic chips, servers, and AI applications [5][7]. - On December 1, AI-related stocks showed strong performance in the secondary market, with significant gains in AI mobile devices and AI computing power sectors [6].
AI进化速递 | DeepSeek发布新模型
Di Yi Cai Jing· 2025-12-01 12:48
Group 1 - DeepSeek V3.2 has been officially released, enhancing Agent capabilities and integrating reasoning and thinking [1] - Doubao Mobile Assistant has released a technical preview version [1] - Tsinghua University has established a research institute for embodied intelligence and robotics [1] Group 2 - Didi's autonomous driving service is undergoing trial operations in Guangzhou, offering all-weather, fully unmanned Robotaxi services [2] - HSBC has formed a strategic partnership with Mistral AI to enhance the application of generative AI in banking operations [2]
DeepSeek-V3.2系列开源,性能直接对标Gemini-3.0-Pro
量子位· 2025-12-01 12:13
衡宇 发自 奥特赛德 量子位 | 公众号 QbitAI 突袭! ChatGPT发布三周年,DeepSeek嚯一下发出两个模型: 前者聚焦平衡实用 ,适用于日常问答、通用Agent任务、真实应用场景下的工具调用。 推理达GPT-5水平,略低于Gemini-3.0-Pro。 下图展示的是DeepSeek-V3.2与其他模型在各类Agent工具调用评测集上的得分 ——特别强调,DeepSeek-V3.2并没有针对这些测试集的工具做特殊训练。 划重点,ICPC达到人类选手第二、IOI人类选手第十名水平。 具体来说,DeepSeek-V3.2侧重于平衡推理能力与输出长度,降低计算开销。 DeepSeek官微推文中写道,"DeepSeek-V3.2模型在Agent评测中达到了当前开源模型的最高水平"。 该模型其他情况如下: DeepSeek-V3.2 DeepSeek-V3.2-Speciale 推理能力比肩GPT-5; 相比Kimi-K2-Thinking大幅缩短输出长度,减少用户等待时间; DeepSeek旗下首个"思考融入工具调用" 的模型,支持思考/非思考双模式工具调用; 基于1800+环境、85000+复杂指令 ...
DeepSeek发布两款新模型
第一财经· 2025-12-01 11:36
Core Insights - DeepSeek has released two official model versions: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale [1][4] Group 1: Model Features - DeepSeek-V3.2 aims to balance reasoning capability and output length, making it suitable for everyday use. In benchmark tests, it reached the level of GPT-5, slightly below Gemini-3.0-Pro, while significantly reducing output length compared to Kimi-K2-Thinking, thus decreasing computational overhead and user wait time [3] - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the extreme, exploring the boundaries of model capabilities. It is an enhanced version of DeepSeek-V3.2 with long-thinking capabilities and incorporates theorem-proving abilities from DeepSeek-Math-V2 [3] Group 2: Achievements - The V3.2-Speciale model has successfully won gold medals at several prestigious competitions, including IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025 [3] Group 3: Availability - The official web version, app, and API have all been updated to the official version of DeepSeek-V3.2. The Speciale version is currently available only as a temporary API service for community evaluation and research [4]
DeepSeek发布新模型
财联社· 2025-12-01 11:35
今日,DeepSeek发布两个正式版模型: DeepSeek-V3.2 和 DeepSeek-V3.2-Speciale。 DeepSeek-V3.2强化Agent能力,官方网页端、App 和 API 均已更新为正式版 DeepSeek-V3.2。 Speciale 版本目前仅以临时 API 服务形式开放,以供社区评测与研究。 ...
DeepSeek V3.2 正式版发布:性能比肩GPT-5 ,略低于 Gemini-3.0-Pro
Xin Lang Ke Ji· 2025-12-01 11:23
Core Insights - DeepSeek has officially released two models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, following the experimental version launched two months ago [1] - DeepSeek-V3.2 aims to balance reasoning capability and output length, making it suitable for everyday use, such as Q&A scenarios and general agent tasks [1] - The performance of DeepSeek-V3.2 in benchmark tests is comparable to GPT-5 and slightly lower than Gemini-3.0-Pro, with significantly reduced output length compared to Kimi-K2-Thinking, leading to lower computational costs and reduced user wait times [1] Model Specifications - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, serving as an enhanced version of DeepSeek-V3.2 with theorem-proving abilities from DeepSeek-Math-V2 [2] - This model excels in instruction following, rigorous mathematical proofs, and logical validation, achieving performance on par with Gemini-3.0-Pro in mainstream reasoning benchmarks [2] - DeepSeek-V3.2-Speciale has won gold medals in prestigious competitions such as IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025, with ICPC and IOI scores reaching the second and tenth positions among human competitors, respectively [2] - While the Speciale model significantly outperforms the standard version in complex tasks, it consumes more tokens and incurs higher costs [2] - Currently, DeepSeek-V3.2-Speciale is available only for research purposes and does not support tool invocation, nor has it been optimized for everyday conversation and writing tasks [2]