Workflow
DeepSeek
icon
Search documents
DeepSeek大模型V3.2亮相!华为、寒武纪芯片同步适配开源,首次自研DSA注意力机制,API价格砍半
Hua Er Jie Jian Wen· 2025-09-29 13:53
Core Insights - DeepSeek has officially released and open-sourced the DeepSeek-V3.2-Exp model on the Hugging Face platform, marking a significant step towards the next generation architecture [1] - The new model introduces the DeepSeek Sparse Attention (DSA) mechanism, which aims to optimize training and inference efficiency for long texts while reducing computational resource consumption [1] - The model supports a maximum context length of 160K, with successful adaptations completed by Huawei and Cambricon [1] Technical Breakthroughs - The DeepSeek Sparse Attention (DSA) mechanism achieves fine-grained sparse attention, significantly enhancing training and inference efficiency for long text scenarios without compromising output quality [1][3] - The training settings for DeepSeek-V3.2-Exp were strictly aligned with the previous version, V3.1-Terminus, showing comparable performance across major public evaluation datasets [3] Benchmark Performance - Performance comparison between DeepSeek-V3.1-Terminus and DeepSeek-V3.2-Exp across various benchmarks shows: - MMLU-Pro: 85.0 (both versions) - GPQA-Diamond: 80.7 (V3.1) vs 79.9 (V3.2) - Humanity's Last Exam: 21.7 (V3.1) vs 19.8 (V3.2) - BrowseComp: 38.5 (V3.1) vs 40.1 (V3.2) - SimpleQA: 96.8 (V3.1) vs 97.1 (V3.2) - Codeforces-Div1: 2046 (V3.1) vs 2121 (V3.2) - AIME 2025: 88.4 (V3.1) vs 89.3 (V3.2) [4] Cost Reduction - The introduction of the new model has led to a significant reduction in API service costs, with a price drop of over 50%, effective immediately [4] Open Source and Community Support - DeepSeek has fully open-sourced the DeepSeek-V3.2-Exp model on Hugging Face and ModelScope, along with related research papers [6] - The company has retained API access for the V3.1-Terminus version for comparison purposes until October 15, 2025, with pricing aligned to V3.2-Exp [6] - To support community research, DeepSeek has also open-sourced GPU operators designed for the new model, recommending the use of the TileLang version for ease of debugging and rapid iteration [6] Industry Collaboration - Cambricon has announced the completion of adaptation for the new model and has open-sourced the vLLM-MLU inference engine source code, allowing developers to experience the new model's features on their hardware platform [6][7]
DeepSeek发布新模型V3.2-Exp并再度降价
Xin Jing Bao· 2025-09-29 13:28
Core Insights - DeepSeek has released an experimental version of its model, DeepSeek-V3.2-Exp, which introduces Sparse Attention for improved training and inference efficiency on long texts [1] Group 1: Model Development - The new version, V3.2-Exp, is a step towards a next-generation architecture, building on the previous V3.1-Terminus [1] - The Sparse Attention mechanism is aimed at optimizing the model's performance for long text processing [1] Group 2: Pricing and Accessibility - The API pricing has been significantly reduced, with costs now at 0.2 yuan per million tokens for cache hits, 2 yuan for cache misses, and 3 yuan for output [1] - This pricing represents a reduction of over 50% compared to previous costs for developers using the DeepSeek API [1]
DeepSeek-V3.2-Exp发布,训练推理提效,API成本降50%以上
Sou Hu Cai Jing· 2025-09-29 13:18
Core Insights - DeepSeek has released the DeepSeek-V3.2-Exp model, which is an experimental version aimed at transitioning to a new generation architecture [1] - The new model introduces DeepSeek Sparse Attention, focusing on optimizing training and inference efficiency for long texts [1] - The official app, web version, and mini-program have all been updated to DeepSeek-V3.2-Exp, and the API costs have been significantly reduced by over 50% for developers [1] - The performance of DeepSeek-V3.2-Exp on various public evaluation sets is comparable to that of V3.1-Terminus [1]
Deepseek API大降价,开发者成本可降超50%
Core Insights - DeepSeek has officially released and open-sourced the DeepSeek-V3.2-Exp model on the Hugging Face platform, with updates across its official app, web, and mini-programs [1] - The new model incorporates a sparse Attention architecture, which reduces computational resource consumption and enhances inference efficiency [1] - The introduction of the new model has led to a significant reduction in service costs, with API prices dropping by over 50% [1] Company Developments - On September 22, DeepSeek updated the DeepSeek-V3.1 model to the DeepSeek-V3.1-Terminus version, achieving a maximum single-item performance improvement of over 36% [3] - The API pricing for the previous version remained unchanged, with costs set at 0.5 yuan for 1 million tokens input and 12 yuan for output [3] - The rapid updates and enhancements in model performance are seen as crucial for retaining developers in a competitive open-source model landscape [3] Industry Impact - Analysts from Guosheng Securities highlight that DeepSeek's high performance, low cost, and innovative algorithms position it competitively against global standards, reshaping investor perceptions of Chinese tech companies' innovation capabilities [3] - The growth in AI model training and inference is expected to drive demand for AI computing power, benefiting various segments of the computing industry, including AI chips and servers [4]
Deepseek API大降价,开发者成本可降超50%
21世纪经济报道· 2025-09-29 12:35
Core Viewpoint - DeepSeek has released its new model DeepSeek-V3.2-Exp, which significantly reduces service costs and enhances model inference efficiency through a sparse Attention architecture [1][5]. Group 1: Model Updates - The DeepSeek-V3.2-Exp model was officially launched on September 29, with updates across its app, web, and mini-program platforms [1]. - The previous version, DeepSeek-V3.1, was updated to DeepSeek-V3.1-Terminus on September 22, showing a performance improvement of over 36% [5]. - The API pricing for DeepSeek has been reduced by over 50% due to the new model's cost efficiency [1]. Group 2: Market Impact - Following the release of DeepSeek-V3.1, there was a notable surge in stock prices for several domestic chip industry companies [5]. - Analysts highlight that the rapid updates and bug fixes in open-source models are crucial for retaining developers in a competitive landscape [5]. Group 3: Industry Insights - The AI computing demand is expected to grow due to large model training and inference, benefiting various segments of the computing industry, including AI chips and servers [6]. - DeepSeek's innovations and performance position it competitively against global leaders, reshaping investor perceptions of Chinese tech companies' capabilities [5].
DeepSeek-V3.2-Exp模型发布并开源,API价格大幅下调
3 6 Ke· 2025-09-29 12:12
Core Insights - DeepSeek-V3.2-Exp model has been officially released and open-sourced, featuring significant updates in architecture and efficiency [1][4] - The introduction of DeepSeek Sparse Attention (DSA) aims to enhance training and inference efficiency for long texts without compromising output quality [1][5] - The API costs for developers have been reduced by over 50% due to the new model's service cost decrease [4] Group 1: Model Features - DeepSeek-V3.2-Exp is an experimental version that builds on V3.1-Terminus, incorporating a sparse attention mechanism [1] - The model achieves fine-grained sparse attention, significantly improving long text training and inference efficiency [1] - The new model's performance is comparable to V3.1-Terminus across various public evaluation datasets [5] Group 2: Development and Implementation - The development of the new model required the design and implementation of numerous new GPU operators, utilizing TileLang for rapid prototyping [2] - The open-sourced operators include both TileLang and CUDA versions, with a recommendation for the community to use the TileLang version for easier debugging [2] Group 3: Previous Versions and Improvements - DeepSeek-V3.1 was released on August 21, featuring a mixed inference architecture and improved efficiency compared to DeepSeek-R1-0528 [4] - The subsequent update to DeepSeek-V3.1-Terminus on September 22 addressed user feedback, enhancing language consistency and agent capabilities [4]
X @Bloomberg
Bloomberg· 2025-09-29 12:08
DeepSeek updated an experimental AI model Monday in what it called a step toward next-generation artificial intelligence https://t.co/jGu6RIkgj9 ...
降价!DeepSeek,大消息!
Core Insights - DeepSeek has officially released the DeepSeek-V3.2-Exp model, which introduces the DeepSeek Sparse Attention mechanism to enhance training and inference efficiency for long texts [1][3] - The performance of DeepSeek-V3.2-Exp is comparable to its predecessor, DeepSeek-V3.1-Terminus, across various benchmark datasets [3][4] - The official app, web version, and mini-program have been updated to DeepSeek-V3.2-Exp, with a significant reduction in API costs by over 50% for developers [4] Model Performance - DeepSeek-V3.2-Exp maintains similar performance levels to DeepSeek-V3.1-Terminus in several benchmarks, such as MMLU-Pro (85.0), GPQA-Diamond (79.9), and SimpleQA (97.1) [4] - Notable improvements were observed in the BrowseComp and Codeforces-Div1 benchmarks, with scores of 40.1 and 2121 respectively for V3.2-Exp [4] Recent Developments - DeepSeek has been active recently, with the release of DeepSeek-V3.1 on August 21, which marked a step towards the "Agent era" with enhanced reasoning capabilities and efficiency [8] - A research paper on the DeepSeek-R1 reasoning model was featured on the cover of the prestigious journal Nature, highlighting significant advancements in AI technology from China [8][9] - Nature's editorial praised DeepSeek for breaking the gap in independent peer review for mainstream large models, marking a milestone for Chinese AI research [9]
降价!DeepSeek,大消息!
证券时报· 2025-09-29 11:55
Core Viewpoint - DeepSeek has launched the DeepSeek-V3.2-Exp model, which introduces a Sparse Attention mechanism to enhance training and inference efficiency for long texts, while maintaining output quality similar to its predecessor, V3.1-Terminus [2][4]. Group 1: Model Performance - The DeepSeek-V3.2-Exp model shows comparable performance to the V3.1-Terminus across various benchmark datasets, with specific scores indicating slight variations in certain areas [5]. - In the MMLU-Pro benchmark, both models scored 85.0, while in the General GPQA-Diamond benchmark, V3.1 scored 80.7 and V3.2-Exp scored 79.9 [5]. - The Sparse Attention mechanism has led to significant improvements in training and inference efficiency without compromising model output [4]. Group 2: Recent Developments - DeepSeek has been active recently, with the V3.1-Terminus model being released on August 21, which introduced a hybrid reasoning architecture and improved efficiency and agent capabilities [8]. - A research paper on the DeepSeek-R1 reasoning model was published in the prestigious journal Nature, marking a significant achievement for Chinese AI research [8][9]. - The new pricing policy for the DeepSeek API has reduced costs for developers by over 50%, making it more accessible [4].
“价格屠夫”DeepSeek上线,新模型成本下降超50%
Di Yi Cai Jing· 2025-09-29 11:50
Core Insights - DeepSeek, known as the "price butcher," has significantly reduced its pricing for the newly released DeepSeek-V3.2-Exp model, with output prices dropping by 75% and overall API costs for developers decreasing by over 50% [1][3]. Pricing Changes - Input pricing for DeepSeek-V3.2-Exp has been adjusted: - Cache hit price decreased from 0.5 yuan per million tokens to 0.2 yuan per million tokens - Cache miss price reduced from 4 yuan per million tokens to 2 yuan per million tokens - Output pricing has been slashed from 12 yuan per million tokens to 3 yuan per million tokens [3]. Model Performance and Features - The V3.2-Exp model is an experimental version that introduces DeepSeek Sparse Attention, enhancing training and inference efficiency for long texts without compromising output quality [3][6]. - Performance evaluations show that DeepSeek-V3.2-Exp maintains comparable results to the previous V3.1-Terminus model across various public benchmark datasets [3][4][5]. Community Support and Open Source - DeepSeek has open-sourced GPU operators designed for the new model, including TileLang and CUDA versions, encouraging community research and experimentation [6]. - The model is now available on platforms like Huggingface and has been updated across official applications and APIs [5][6]. Industry Context - Following the recent release of DeepSeek-V3.1-Terminus, there is speculation about the future of the V4 and R2 versions, with industry voices expressing anticipation for major updates [6].