Workflow
Seek .(SKLTY)
icon
Search documents
实测低调上线的DeepSeek新模型:编程比Claude 4还能打,写作...还是算了吧
3 6 Ke· 2025-08-20 12:14
Core Insights - DeepSeek has officially launched and open-sourced its new model, DeepSeek-V3.1-Base, following the release of GPT-5, despite not having released R2 yet [1] - The new model features 685 billion parameters and supports multiple tensor types, with significant optimizations in inference efficiency and an expanded context window of 128k [1] Model Performance - Initial tests show that DeepSeek V3.1 achieved a score of 71.6% on the Aider Polyglot programming benchmark, outperforming other open-source models, including Claude 4 Opus [5] - The model successfully processed a long text and provided relevant literary recommendations, demonstrating its capability in handling complex queries [4] - In programming tasks, DeepSeek V3.1 generated code that effectively handled collision detection and included realistic physical properties, showcasing its advanced programming capabilities [8] Community and Market Response - Hugging Face CEO Clément Delangue noted that DeepSeek V3.1 quickly climbed to the fourth position on the trends chart, later reaching second place, indicating strong market interest [79] - The update removed the "R1" label from the deep thinking mode and introduced native "search token" support, enhancing the search functionality [79][80] Future Developments - The company plans to discontinue the mixed thinking mode in favor of training separate Instruct and Thinking models to ensure higher quality outputs [80] - As of the latest update, the model card for DeepSeek-V3.1-Base has not yet been released, but further technical details are anticipated [81]
DeepSeek V3.1发布后,投资者该思考这四个决定未来的问题
3 6 Ke· 2025-08-20 10:51
Core Insights - DeepSeek has quietly launched its new V3.1 model, which has generated significant buzz in both the tech and investment communities due to its impressive performance metrics [1][2][5] - The V3.1 model outperformed the previously dominant Claude Opus 4 in programming capabilities, achieving a score of 71.6% in the Aider programming benchmark [2] - The cost efficiency of V3.1 is notable, with a complete programming task costing approximately $1.01, making it 68 times cheaper than Claude Opus 4 [5] Group 1: Performance and Cost Advantages - The V3.1 model's programming capabilities have surpassed those of Claude Opus 4, marking a significant achievement in the open-source model landscape [2] - The cost to complete a programming task with V3.1 is only about $1.01, which is a drastic reduction compared to competitors, indicating a strong cost advantage [5] Group 2: Industry Implications - The emergence of V3.1 raises questions about the future dynamics between open-source and closed-source models, particularly regarding the erosion and reconstruction of competitive advantages [8] - The shift towards a "hybrid model" is becoming prevalent among enterprises, combining private deployments of fine-tuned open-source models with the use of powerful closed-source models for complex tasks [8][9] Group 3: Architectural Innovations - The removal of the "R1" designation and the introduction of new tokens in V3.1 suggest a potential exploration of "hybrid reasoning" or "model routing" architectures, which could have significant commercial implications [11] - The concept of a "hybrid architecture" aims to optimize inference costs by using a lightweight scheduling model to allocate tasks to the most suitable expert models, potentially enhancing unit economics [12] Group 4: Market Dynamics and Business Models - The drastic reduction in inference costs could lead to a transformation in AI application business models, shifting from per-call or token-based billing to more stable subscription models [13] - As foundational models become commoditized due to open-source competition, the profit distribution within the value chain may shift towards application and solution layers, emphasizing the importance of high-quality private data and industry-specific expertise [14] Group 5: Future Competitive Landscape - The next competitive battleground will focus on "enterprise readiness," encompassing stability, predictability, security, and compliance, rather than solely on performance metrics [15] - Companies that can provide comprehensive solutions, including models, toolchains, and compliance frameworks, will likely dominate the trillion-dollar enterprise market [15]
奥尔特曼:DeepSeek和Kimi是OpenAI开源的重要原因
Huan Qiu Wang Zi Xun· 2025-08-20 08:21
Core Viewpoint - OpenAI's founder Sam Altman believes that the U.S. is underestimating the threat posed by China's next-generation artificial intelligence, and that chip regulations alone are not an effective solution [1][3] Group 1: AI Competition - Altman stated that China can develop faster in reasoning capabilities and has strengths in research and product development [3] - The AI competition between the U.S. and China is deeply intertwined, going beyond simple rankings [3] Group 2: OpenAI's Strategic Shift - OpenAI recently released its first open-weight models, gpt-oss-120b and gpt-oss-20b, marking a significant strategic shift from its long-standing closed-source approach [3] - The decision to release open-weight models was influenced by competition from Chinese models, such as DeepSeek and Kimi K2 [3] - Altman emphasized that if OpenAI did not act, Chinese open-source models would gain widespread adoption, making this a significant factor in their decision [3]
早盘消息0820| T 链 Gen3 技术路线重塑供应链、DeepSeek 模型升级到V3.1…
Xin Lang Cai Jing· 2025-08-20 05:17
Group 1: Photovoltaic Industry - The Ministry of Industry and Information Technology (MIIT) is actively coordinating between power generation companies and local industries to enhance price transmission from manufacturing to power stations, emphasizing a market-oriented and legal approach to eliminate outdated production capacity [1] - The average bidding price for components from China Resources and China Huadian has increased by 5-8% month-on-month, while silicon material companies have proactively limited production, leading to a 10% decrease in silicon wafer inventory over two weeks [1] - The investment sequence indicates a tight supply of silicon materials in Q3, a premium for BC battery technology in Q4, and a simultaneous increase in both volume and price of auxiliary materials such as glass and adhesive films [1][2] Group 2: Solid-State Battery Technology - A breakthrough in solid-state battery technology has been achieved with the introduction of 5μm vapor-deposited lithium anodes, significantly reducing dendrite risk and achieving over 500 cycles with a capacity retention rate above 90% [3] - The cost of 5μm vapor-deposited lithium is projected to drop to 2 million yuan per GWh, compared to 4 million yuan for 20μm rolled lithium foil, indicating a substantial cost reduction in the industry [3] - The solid-state battery market could reach 50-100 billion yuan by 2030, driven by the demand for 100GWh of global solid-state battery production [3] Group 3: Robotics Industry - The T-Link Gen3 technology is reshaping the supply chain with a focus on lightweight materials, energy efficiency, and sensor integration, leading to a re-tendering of motors, reducers, and lead screws [4] - The use of PEEK materials has reduced costs by 30% compared to imports, and the new harmonic magnetic field motors have achieved a 50% reduction in size while doubling power density [5] - The 3D vision solution from Orbbec has a single machine value of 200 USD, and the company has passed factory audits [6] Group 4: Semiconductor and AI Models - The DeepSeek model has been upgraded to V3.1, expanding the context length from 64K to 128K, which is expected to increase demand for GPU memory and HBM [7] - The need for larger training clusters is anticipated to rise by 30%, benefiting semiconductor and storage manufacturers such as Cambricon, Haiguang, and Lanke [7] Group 5: Pharmaceutical Industry - Rongchang Biotech has licensed its ophthalmic drug RC28-E to Japan's Santen Pharmaceutical, marking a shift in domestic innovative drug licensing from popular fields like oncology to specialized areas with differentiated advantages [8] - This collaboration model provides a clear path for value realization in less popular biotech sectors through upfront payments, milestones, and sales sharing, enhancing cash flow and leveraging established commercialization channels [8] Group 6: High-Speed Rail Industry - The China National Railway Group has initiated its second batch of high-speed train tenders for the year, with 210 sets, marking a recent high and exceeding market expectations [9] - This move reinforces the trend of sustained railway investment recovery, with new construction and maintenance peaks positively impacting the performance certainty of core companies in the industry [9]
DeepSeek 开源新模型 V3.1:上下文长度拓展至 128K
Huan Qiu Wang Zi Xun· 2025-08-20 04:54
来源:环球网 【环球网科技综合报道】8月20日消息,DeepSeek日前在Hugging Face上开源了新模型 V3.1-Base。 此外,日前DeepSeek 还发布通知称,线上模型版本已升级至 V3.1,上下文长度拓展至 128k,可通过官 方网页、App、小程序测试,API 接口调用方式保持不变。 就在8月14日,DeepSeek App发布了1.3.0版本,此次更新在修复已知问题、优化文本操作体验的基础 上,首次引入"对话内容生成分享图"功能,为用户提供更便捷、个性化的内容传播方式。(思瀚) ...
DeepSeek V3.1 Base突袭上线,击败Claude 4编程爆表,全网在蹲R2和V4
3 6 Ke· 2025-08-20 03:52
Core Insights - The newly released DeepSeek V3.1 model features 685 billion parameters and supports various precision formats, from BF16 to FP8 [1] - The model demonstrates exceptional programming capabilities, achieving a score of 71.6% in the Aider programming benchmark, surpassing Claude Opus 4 [1][11] - V3.1 introduces native search token support, enhancing search functionalities [1] - The architecture has been innovated by removing the "R1" designation, indicating a potential shift towards a hybrid architecture in future models [1][10] - The cost for a complete programming task is only $1.01, significantly lower than proprietary systems, which are 60 times more expensive [1][13][16] Performance Metrics - DeepSeek V3.1 has 671 billion parameters activated with a context length of 128K tokens, ranking fourth on Hugging Face's trend list even before the model card was released [2] - The model's programming performance is 1% higher than Claude 4, with a cost reduction of 68 times [16] - In the SVGBench benchmark, V3.1 ranks just below GPT-4.1-mini, outperforming its predecessor, DeepSeek R1 [17] User Engagement - The DeepSeek community has grown to over 80,000 followers, indicating strong interest and anticipation for future releases [4] - Users have reported significant improvements in understanding and output speed, particularly in context length tests [21][25]
DeepSeek有点含蓄了,实测V3.1有进步,编程等个别场景硬刚GPT-5
3 6 Ke· 2025-08-20 03:03
没等到Deepseek R2,DeepSeek悄悄更新了V 3.1。 官方群放出的消息就提了一点,上下文长度拓展至128K。128K也是GPT-4o这一代模型的处理Token的长度。因此一开始,鲸哥以为从V3升级到V 3.1,以 为是不大的升级,鲸哥体验下来还有惊喜。 代码能力与前端审美提升 从开源社区Huggingface上传的模型版本看,模型尺寸达685B,支持 BF16、F8_E4M3、F32 等张量类型,平衡模型的计算精度和效率。 最惊喜的是代码能力提升明显,前端审美也有大幅度提升。我们先看V3.1在代码测试中的变现。 请设计并开发一款结合日历和待办事项(To-Do)的产品,其核心功能应包括: 任务分类与颜色标记:用户能够创建不同类别的任务,并为每个类别分配独特的颜色。当任务被归类后,其在日历视图上应以相应的颜色进行标记,以便 快速识别。短期任务管理:*完成标记: 对于计划在特定日期完成的任务,用户应能将其标记为"已完成"。已完成的任务应在界面上以视觉方式(例如, 划掉、变灰或显示完成图标)清晰区分。*逾期处理: 如果任务未在计划日期完成,系统应提供明确的视觉提示(例如,颜色变化、闪烁或标记为逾 期) ...
刚刚,DeepSeek新模型开源,五大能力变化明显,附一手体验
3 6 Ke· 2025-08-20 00:14
Core Insights - DeepSeek has upgraded its online model to DeepSeek V3.1, expanding the context window from 64k to 128k, available across web, app, and mini-program platforms [3][21] - The new model has shown improvements in various capabilities, including programming, understanding physical laws, creative writing, and mathematical problem-solving [4][19] Model Upgrade - The model is now open-sourced on Hugging Face, with only the Base version available for download, showing no significant changes in parameters or tensor types compared to DeepSeek-V3-0324 [2] - Initial experiences with DeepSeek V3.1 indicate enhanced performance in web development, with more complex and aesthetically pleasing outputs compared to the previous version [4][6] Performance Enhancements - In web development tasks, DeepSeek V3.1 produced longer code with improved completion and aesthetics, demonstrating better layout and content planning [4][6] - The model successfully recreated a simple game similar to the Chrome dinosaur game, although some aspects of the game were not accurately rendered [8] Question Answering and Interaction - DeepSeek V3.1 provided more detailed and factually accurate responses to niche historical questions, showing a reduction in "hallucination" compared to its predecessor [10][12] - The model's tone has shifted to be more conversational and nuanced, using conditional statements and emphasizing complexity in its answers [13] Creative Outputs - The model demonstrated its creative capabilities by generating poetry and engaging in playful comparisons between notable figures in AI, showcasing a balanced approach in its responses [17][16] Mathematical Abilities - DeepSeek V3.1 displayed a mixed performance in basic arithmetic, initially providing incorrect answers before correcting itself [18] User Engagement - Users have quickly adopted the new model, with feedback highlighting improvements in physical simulations and creative outputs [19][21]
AI进化速递 | DeepSeek线上模型版本升级
Di Yi Cai Jing· 2025-08-19 13:19
Group 1 - DeepSeek has upgraded its online model to version 3.1, expanding the context length to 128k [1] - Alibaba's Tongyi Qianwen has launched an image editing model called Qwen-Image-Edit [1] - Nvidia has released a small language model named Nemotron-Nano-9B-v2 [1] Group 2 - Xiaopeng Motors' chairman He Xiaopeng announced that humanoid robots and L4-supported vehicles are expected to be mass-produced by 2026 [3] - Nvidia is collaborating with Foxconn to develop a humanoid robot, which is expected to debut in November [1] - Figure's founder indicated that Helix is set to undergo a significant upgrade [1] - Arm has hired Amazon's AI chip head Rami Shino to support its self-developed chip initiatives [1]
DeepSeek线上模型版本升级
Di Yi Cai Jing· 2025-08-19 13:17
Core Viewpoint - The online model version of DeepSeek has been upgraded to V3.1, expanding the context length to 128K, while maintaining the same access methods through the official website, app, mini-program, and API calls [1] Group 1 - DeepSeek's online model has been upgraded to version 3.1 [1] - The context length has been expanded to 128K [1] - Access methods remain unchanged, including the official website, app, mini-program, and API [1]