Workflow
DeepSeek
icon
Search documents
DeepSeek发布最强开源新品,瞄向全能Agent,给GPT-5与Gemini 3下战书
Tai Mei Ti A P P· 2025-12-01 15:03
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, marking a significant advancement in AI capabilities, particularly in reasoning and output efficiency [2][3] - The V3.2 model is positioned as the strongest open-source large model, outperforming competitors in various benchmarks while significantly reducing output length and computational costs [3][4] - The V3.2 model integrates a new sparse attention mechanism (DSA) to enhance performance in long-context scenarios, while also improving the model's ability to follow instructions and generalize in complex environments [8][9] Model Performance - In benchmark tests, DeepSeek-V3.2 achieved competitive scores against models like GPT-5, Claude 4.5, and Gemini 3 Pro, with notable strengths in specific areas [4][5] - The V3.2 model demonstrated superior performance in question-and-answer scenarios, providing detailed and accurate travel recommendations through advanced tool usage [5][6] - The V3.2 Speciale model focuses on maximizing reasoning capabilities, achieving results comparable to Gemini 3.0 Pro in mainstream reasoning benchmarks, although it requires a higher token cost and is not designed for everyday use [9][10] Development Focus - DeepSeek emphasizes practical usability and generalization in its models, aiming to overcome common pitfalls in AI interactions, such as making basic common-sense errors [6][8] - The company is committed to enhancing the reasoning abilities of its models, as evidenced by the integration of advanced mathematical reasoning capabilities from the recently released DeepSeek-Math-V2 [9][10] - The competitive landscape for large models is intensifying, with major players like GPT-5 and Gemini 3 pushing the boundaries of AI capabilities, suggesting a dynamic future for AI development [10]
DeepSeek发布V3.2正式版
Xin Jing Bao· 2025-12-01 15:01
Core Insights - DeepSeek announced the release of two official model versions: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale [1] Model Overview - DeepSeek-V3.2 aims to balance reasoning capability and output length, making it suitable for everyday use, such as Q&A scenarios and general agent tasks [1] - In benchmark tests for reasoning, DeepSeek-V3.2 achieved performance comparable to GPT-5, slightly below Gemini-3.0-Pro [1] - Compared to Kimi-K2-Thinking, V3.2 significantly reduced output length, leading to lower computational costs and reduced user wait times [1] Special Features - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, exploring the boundaries of model performance [1] - This version is an enhanced long-thinking variant of DeepSeek-V3.2, incorporating theorem-proving capabilities from DeepSeek-Math-V2 [1] - The model exhibits excellent instruction-following, rigorous mathematical proof, and logical verification abilities, performing comparably to Gemini-3.0-Pro in mainstream reasoning benchmark tests [1]
DeepSeek,上新
Core Insights - DeepSeek has released two new models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aimed at enhancing reasoning capabilities and output length for various applications [1][2]. Model Performance - DeepSeek-V3.2 achieved performance comparable to GPT-5 and slightly below Gemini-3.0-Pro in public reasoning benchmarks, while significantly reducing output length compared to Kimi-K2-Thinking, thus lowering computational costs and user wait times [1][3]. - The DeepSeek-V3.2-Speciale model demonstrated exceptional instruction-following, rigorous mathematical proof, and logical validation capabilities, achieving gold medal-level performance in major competitions such as IMO 2025 and ICPC World Finals 2025 [2][3]. Benchmark Comparisons - In various benchmark tests, DeepSeek-V3.2-Speciale outperformed standard versions and other models, with notable scores in AIME 2025 (96.0) and HMMT Feb 2025 (99.2), while also achieving high rankings in IMOAnswerBench and LiveCodeBench [3]. - The performance of DeepSeek-V3.2-Speciale in complex tasks was significantly better than the standard version, although it required more tokens, indicating higher operational costs [3]. Model Features - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation, enhancing its versatility [4]. - The model has improved generalization capabilities through a novel large-scale agent training data synthesis method, allowing it to perform well in real-world applications [4].
DeepSeek,重大发布
证券时报· 2025-12-01 14:16
Core Insights - DeepSeek has released two official model versions: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, with the former being available on the official website, app, and API, while the latter is currently accessible only through a temporary API for community evaluation [2][4]. Model Performance - DeepSeek-V3.2 aims to balance reasoning capability and output length, achieving performance comparable to GPT-5 and slightly below Gemini-3.0-Pro in benchmark tests. It significantly reduces output length compared to Kimi-K2-Thinking, leading to lower computational costs and reduced user wait times [4]. - The Speciale version is designed to push the limits of reasoning capabilities, combining features from DeepSeek-V3.2 and DeepSeek-Math-V2, and has shown performance on par with Gemini-3.0-Pro in mainstream reasoning benchmarks [4]. Benchmark Results - In various benchmark tests, DeepSeek-V3.2-Speciale achieved notable results, including: - AIME 2025: 96.0 (23k) - HMMT Feb 2025: 99.2 (27k) - HMMT Nov 2025: 94.4 (25k) - IMOAnswerBench: 84.5 (45k) - CodeForces: 2701 (77k) - HILE: 30.6 (35k) [5]. Cost Efficiency - The introduction of the new attention mechanism DSA in DeepSeek-V3.2-Exp has led to significant improvements in training efficiency and a reduction in model costs, enhancing the model's cost-effectiveness and potential for broader application [5].
DeepSeek又上新!模型硬刚谷歌
第一财经· 2025-12-01 14:05
Core Viewpoint - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which are leading in reasoning capabilities globally [3]. Model Overview - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for everyday use such as Q&A and general intelligence tasks. It has reached the level of GPT-5 in public reasoning tests, slightly below Google's Gemini3 Pro [5]. - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the extreme, combining long-thinking enhancements and theorem-proving abilities from DeepSeek-Math-V2 [5]. Performance Metrics - Speciale has surpassed Google's Gemini3 Pro in several reasoning benchmark tests, including the American Mathematics Invitational, Harvard MIT Mathematics Competition, and International Mathematical Olympiad [6]. - In the AIME 2025 benchmark, Speciale scored 96.0, while Gemini-3.0 scored 95.0 [7]. - Speciale achieved gold medals in IMO, ICPC World Finals, and IOI, with ICPC and IOI scores reaching the levels of the second and tenth human competitors, respectively [6]. Limitations and Future Plans - DeepSeek acknowledges limitations compared to proprietary models like Gemini3 Pro, including a narrower breadth of world knowledge and lower token efficiency [8]. - The company plans to increase pre-training computational resources and optimize model reasoning chains to improve efficiency and fill knowledge gaps [8]. Industry Context - The gap between open-source and closed-source models is widening, with proprietary systems showing stronger performance in complex tasks [10]. - DeepSeek has introduced a sparse attention mechanism (DSA) to reduce computational complexity without sacrificing long-context performance, which has been effective in improving model performance [11]. Community Reception - The release of DeepSeek's models has been positively received in overseas social media, with comments highlighting the achievement of matching GPT-5 and Gemini3 Pro with an open-source model [11].
DeepSeek 重要发布
Core Insights - DeepSeek has officially released two models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, with updates available on the official website, app, and API [1] - DeepSeek-V3.2 aims to balance reasoning capabilities and output length, making it suitable for everyday use cases such as Q&A and general agent tasks [1] - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the limit, enhancing long-thinking abilities and incorporating theorem-proving capabilities from DeepSeek-Math-V2 [1] Model Performance - The V3.2-Speciale model exhibits excellent instruction-following, rigorous mathematical proof, and logical verification capabilities, performing comparably to leading international models on mainstream reasoning benchmarks [1] - Notably, the V3.2-Speciale model has achieved gold medals in several prestigious competitions, including IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025 [1] - In the ICPC and IOI competitions, the model's performance reached the level of the second and tenth place among human competitors, respectively [1]
DeepSeek,又有大动作!
Core Insights - DeepSeek has launched two new models: DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aiming to enhance reasoning capabilities and output length for various applications [1][2][3] Model Features - DeepSeek-V3.2 is designed for everyday use, balancing reasoning ability and output length, and has achieved performance comparable to GPT-5 in benchmark tests [2][3] - DeepSeek-V3.2-Speciale enhances long reasoning capabilities and incorporates theorem proving abilities from DeepSeek-Math-V2, excelling in complex tasks but requiring more tokens and higher costs [3][4] Technological Advancements - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation, significantly improving generalization capabilities [4] - The model has been trained on over 1,800 environments and 85,000 complex instructions, narrowing the performance gap between open-source and closed-source models [4] Market Outlook - The AI industry is experiencing a resonance period, with rapid expansion in AI infrastructure and commercialization of downstream applications, expected to continue thriving through 2026 [5][6][7] - Investment opportunities are identified in domestic AI chains, overseas AI hardware markets, and innovative applications in the domestic edge AI industry [7]
DeepSeek又上新!模型硬刚谷歌,承认开源与闭源差距拉大
Di Yi Cai Jing· 2025-12-01 13:31
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which are leading in reasoning capabilities globally [1][3]. Model Overview - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for everyday use such as Q&A and general intelligence tasks. It has reached the level of GPT-5 in public reasoning tests, slightly below Google's Gemini3 Pro [3]. - DeepSeek-V3.2-Speciale is designed to push the reasoning capabilities of open-source models to the extreme, combining features from DeepSeek-Math-V2 for theorem proving, and excels in instruction following and logical verification [3][4]. Performance Metrics - Speciale has surpassed Google's Gemini3 Pro in several reasoning benchmark tests, including the American Mathematics Invitational, Harvard MIT Mathematics Competition, and International Mathematical Olympiad [4]. - In various benchmarks, DeepSeek's performance is competitive, with specific scores noted in a comparative table against GPT-5 and Gemini-3.0 [5]. Technical Limitations - Despite achievements, DeepSeek acknowledges limitations compared to proprietary models like Gemini3 Pro, particularly in knowledge breadth and token efficiency [6]. - The company plans to enhance pre-training computation and optimize reasoning chains to improve model efficiency and capabilities [6][7]. Mechanism Innovations - DeepSeek introduced a Sparse Attention Mechanism (DSA) to reduce computational complexity, which has proven effective in enhancing performance without sacrificing long-context capabilities [7][8]. - Both new models incorporate this mechanism, making DeepSeek-V3.2 a cost-effective alternative that narrows the performance gap with proprietary models [8]. Community Reception - The release has been positively received in the community, with users noting that DeepSeek's models are now comparable to GPT-5 and Gemini3 Pro, marking a significant achievement in open-source model development [8].
DeepSeek V3.2 正式版发布,V4 还没来,但已经是开源模型里 Agent 能力最强了
Founder Park· 2025-12-01 13:14
Core Insights - DeepSeek has released the official version of its V3.2 model, which significantly enhances reasoning and agent capabilities compared to previous versions [2][9] - The V3.2-Speciale version is an open-source model that performs comparably to Gemini-3.0-Pro on mainstream reasoning benchmarks and has achieved gold medal levels in several prestigious competitions [3][11] - The integration of the DeepSeek Sparse Attention (DSA) technology in V3.2 improves long text processing efficiency and reduces costs by over 50% [3][10] Model Development - The V3 series has been iterated over the past year, with V3.2 being the latest release, focusing on unifying thinking and non-thinking models, a trend seen in other closed-source models like Gemini and GPT-5 [6][9] - The release timeline for DeepSeek models in 2025 includes various versions, each with specific enhancements, such as the introduction of DSA in V3.2 for stability and reasoning improvements [7][8] Performance Metrics - DeepSeek-V3.2 has achieved reasoning capabilities on par with GPT-5 and has shown significant improvements in output length and computational efficiency compared to Kimi-K2-Thinking [10][14] - The V3.2-Speciale version excels in complex tasks, achieving high scores in various academic competitions, including IMO 2025 and ICPC 2025, with notable rankings among human competitors [11][14] Tool Utilization - A key advancement in V3.2 is the incorporation of thinking processes into tool calls, allowing the model to support both thinking and non-thinking modes in its operations [15][18] - DeepSeek has developed a large-scale agent training data synthesis method that enhances the model's generalization capabilities by creating numerous "hard-to-answer, easy-to-verify" tasks [16][18]
DeepSeek,重大突发!
券商中国· 2025-12-01 13:01
Core Viewpoint - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, aiming to enhance reasoning capabilities and application in various scenarios [1][2]. Model Features - DeepSeek-V3.2 aims to balance reasoning ability and output length, suitable for daily use such as Q&A and general agent tasks. It has achieved performance comparable to GPT-5 in benchmark tests, slightly below Gemini-3.0-Pro [2]. - DeepSeek-V3.2-Speciale is an enhanced version of V3.2, integrating theorem proving capabilities from DeepSeek-Math-V2, excelling in instruction following and logical verification. It has won gold medals in several prestigious competitions, including IMO 2025 and ICPC World Finals 2025 [3]. Technological Advancements - DeepSeek-V3.2 is the first model to integrate reasoning with tool usage, supporting both reasoning and non-reasoning modes for tool invocation. The model has been trained on over 85,000 complex instructions across 1,800 environments, significantly improving its generalization ability [4]. Market Outlook - The AI industry is entering a period of resonance, with rapid expansion in AI infrastructure and commercialization of downstream applications. Analysts predict continued prosperity in the AI sector, with a focus on domestic chips, servers, and AI applications [5][7]. - On December 1, AI-related stocks showed strong performance in the secondary market, with significant gains in AI mobile devices and AI computing power sectors [6].