Seek .(SKLTY)
Search documents
DeepSeek线上模型升级至V3.1-Terminus!算力与应用板块或迎价值重估(附概念股)
Zhi Tong Cai Jing· 2025-09-22 23:37
Core Insights - DeepSeek has officially upgraded its model to DeepSeek-V3.1-Terminus, enhancing performance based on user feedback, particularly in language consistency and agent capabilities [1][2] - The new model shows improved stability in output, with benchmark results indicating significant performance gains across various assessments [1] - The release of DeepSeek-V3.1 is seen as a breakthrough for domestic large models and chip ecosystems, addressing compatibility issues with NVIDIA's FP8 standard [2][3] Model Performance - The benchmark results for DeepSeek-V3.1-Terminus compared to its predecessor are as follows: - MMLU-Pro: 85.0 (up from 84.8) - GPQA-Diamond: 80.7 (up from 80.1) - Humanity's Last Exam: 21.7 (up from 15.9) - BrowseComp: 38.5 (up from 30.0) - SimpleQA: 96.8 (up from 93.4) - SWE Verified: 68.4 (up from 66.0) [1] Industry Impact - The launch of DeepSeek V3.1 has significantly boosted the domestic computing industry, with expectations for increased applications of domestic AI chips in training and inference [3][4] - The success of DeepSeek is viewed as a victory for open-source models, prompting other Chinese companies to adopt similar open-source strategies [3] - The AI computing demand is projected to grow, benefiting various segments of the computing supply chain, including AI chips and servers [4] Related Developments - DeepSeek's research paper on the R1 reasoning model has been featured on the cover of the prestigious journal Nature, marking a significant achievement in the field [2] - Other companies in the industry, such as Baidu and Alibaba, are also advancing their models, with Baidu's Wenxin model showing a 34.8% improvement in factual accuracy [6] and Alibaba launching its Qwen3-Max-Preview model [6]
上证早知道|央行,再次出手;DeepSeek,最新升级;事关工业园区发展,两部门印发
Shang Hai Zheng Quan Bao· 2025-09-22 23:36
Monetary Policy - The central bank announced a 240.5 billion yuan reverse repurchase operation for 7-day terms and a 300 billion yuan operation for 14-day terms, marking the first 14-day operation in 8 months [2][4]. Industry Development - The Ministry of Industry and Information Technology and the National Development and Reform Commission issued guidelines for the high-quality development of industrial parks, emphasizing the acceleration of green facility construction and the development of new energy infrastructure [2][4]. Sports and Health - The National Sports Administration released guidelines aimed at enhancing public service systems for national fitness and promoting the integration of sports and health [4]. Capital Market Insights - The China Securities Regulatory Commission reported that over 90% of new listed companies in recent years are technology firms, with the market capitalization of the technology sector now exceeding 25% of the total A-share market [7][8]. Semiconductor Industry - Domestic chip stocks surged, with companies like Haiguang Information and Chip Origin seeing gains over 10%. The rapid growth in demand for computing power is expected to drive the GPU and semiconductor industries [10]. Company Performance - Changchuan Technology projected a net profit of 827 million to 877 million yuan for the first three quarters of 2025, representing a year-on-year increase of 131.39% to 145.38% due to strong demand in the semiconductor market [11]. Investment Activities - Zhangjiang Hi-Tech announced an investment of 22.345 million yuan in Shanghai Microelectronics, holding a 10.779% stake [13]. - Huadian International plans to increase its registered capital in a joint investment company by 5 billion yuan, maintaining a 12% stake [15]. - Huahai Pharmaceutical received approval for a new drug, which is expected to have significant therapeutic effects [16]. Market Transactions - Institutional investors bought a net 4.54 billion yuan of Jucheng shares, accounting for 23.62% of the total trading volume [21]. - One institutional investor purchased 2.73 billion yuan of Chip Origin shares, which is part of a strategic acquisition plan to enhance its competitive edge in the RISC-V CPU market [22].
港股概念追踪 | DeepSeek线上模型升级至V3.1-Terminus!算力与应用板块或迎价值重估(附概念股)
智通财经网· 2025-09-22 23:27
Core Insights - DeepSeek has officially upgraded its model to DeepSeek-V3.1-Terminus, enhancing performance based on user feedback and improving language consistency and agent capabilities [1][2] - The new model shows improved stability in output, with benchmark results indicating performance increases in various assessments compared to the previous version [1] - The release of DeepSeek V3.1 is seen as a significant breakthrough for domestic large models and chip ecosystems, reducing reliance on NVIDIA standards and promoting domestic computing power autonomy [2][3] Model Performance - The benchmark results for DeepSeek-V3.1-Terminus show improvements in several areas, including: - MMLU-Pro: 84.8 to 85.0 - Humanity's Last Exam: 15.9 to 21.7 - SimpleQA: 93.4 to 96.8 - BrowseComp: 30.0 to 38.5 [1] - The model's agent capabilities have significantly improved, which is expected to enhance commercial applications of AI agents [3] Industry Impact - The launch of DeepSeek V3.1 has led to a surge in the domestic computing industry, with increased demand for AI chips and related infrastructure [3][4] - The success of DeepSeek is viewed as a victory for open-source models, prompting other Chinese companies to adopt similar open-source strategies [3] - The AI computing demand is projected to grow, benefiting various segments of the computing industry, including AI chips, servers, and related technologies [4] Related Companies - Baidu has released its Wenxin model X1.1, showing significant improvements in performance metrics compared to previous versions and competing models [6] - Alibaba's Tongyi Qianwen has launched the Qwen3-Max-Preview model, marking advancements in the domestic large model sector [6] - SenseTime's new interactive platform integrates with Xiaomi AI glasses, showcasing the application of AI in real-world scenarios [7] - ZTE has introduced several products focused on AI and intelligent computing, facilitating the deployment of DeepSeek models across various industries [7]
DeepSeek-V3.1版本更新
Di Yi Cai Jing· 2025-09-22 13:45
Core Insights - The update maintains the original capabilities of the model while addressing user feedback issues [1] Group 1 - DeepSeek has been updated to version DeepSeek-V3.1-Terminus [1] - Improvements include language consistency, alleviating mixed Chinese and English usage, and occasional abnormal characters [1] - The performance of Code Agent and Search Agent has been further optimized [1]
DeepSeek官宣线上模型升级 版本号DeepSeek-V3.1-Terminus
Xin Lang Ke Ji· 2025-09-22 12:06
Core Insights - DeepSeek has announced an upgrade to its online model, now at version DeepSeek-V3.1-Terminus, which includes both thinking and non-thinking modes [1] - The model supports a context length of 128k, enhancing user experience by allowing for more extensive interactions [1] - Users can now experience the upgraded model online, indicating a focus on accessibility and user engagement [1]
DeepSeek官宣线上模型升级,版本号DeepSeek-V3.1-Terminus
Xin Lang Ke Ji· 2025-09-22 11:59
Core Insights - DeepSeek has announced the upgrade of its online model to version DeepSeek-V3.1-Terminus, which includes both a thinking model and a non-thinking mode [2] Group 1: Model Features - The context length for both models is set at 128k [2] - The non-thinking model has a default output length of 4K and a maximum of 8K, while the thinking model has a default output length of 32K and a maximum of 64K [2] Group 2: Pricing Structure - The cost for inputting one million tokens with cache hit is 0.5 yuan, while the cost for cache miss is 4 yuan [2] - The output cost for one million tokens is set at 12 yuan [2]
这一空白终于被DeepSeek打破
Xin Lang Cai Jing· 2025-09-21 06:26
Core Insights - DeepSeek has achieved a significant milestone by having its research paper on the DeepSeek-R1 inference model published in the prestigious journal Nature, marking a breakthrough in the independent peer review of large models [1] - The paper details the model's training methods and data sources, emphasizing transparency and reproducibility in the AI industry, which has been criticized for its "black box" nature since the rise of ChatGPT [1] - DeepSeek's commitment to open-source technology has contributed to its success, with the model being downloaded over 10.9 million times on the HuggingFace platform since its release [1] Industry Impact - DeepSeek is actively applying its technology in verticals such as medical consultation and industrial quality inspection, showcasing the potential of AI to enhance production and daily life [1] - The company exemplifies China's innovative path, demonstrating that true technological advancement thrives in an open and inclusive ecosystem [1] - Amid rising protectionism and unilateralism globally, China is pursuing its own path in scientific innovation while advocating for open collaboration to keep pace with technological development [1]
金沙江创投朱啸虎:大家低估了DeepSeek的影响力
Xin Lang Ke Ji· 2025-09-20 02:26
Core Insights - The influence of DeepSeek is underestimated, according to Zhu Xiaohu, a managing partner at Jinsha River Venture Capital [1] - The future of AI development will not be controlled by a few privatized companies or models, but will instead be characterized by an open-source and open AI ecosystem, which is crucial for humanity [3] Group 1 - DeepSeek's impact on the AI landscape is significant and should not be overlooked [1] - The evolution of AI will lead to a more democratized and accessible ecosystem, moving away from privatization [3]
DeepSeek首度公开R1模型训练成本仅为29.4万美元,“美国同行开始质疑自己的战略”
Xin Lang Cai Jing· 2025-09-19 13:25
Core Insights - DeepSeek has achieved a significant breakthrough in AI model training costs, with the DeepSeek-R1 model costing only $294,000 to train, which is substantially lower than the costs reported by American competitors [1][2][4] - The model's training utilized 512 NVIDIA H800 chips, and the total training time was 80 hours, marking it as the first mainstream large language model to undergo peer review [2][4] - The cost efficiency of DeepSeek's model has sparked discussions about China's position in the global AI landscape, challenging the notion that only countries with the most advanced chips can dominate the AI race [1][2] Cost Efficiency - The training cost of DeepSeek-R1 is reported at $294,000, while OpenAI's CEO indicated that their foundational model training costs exceed $100 million [2] - DeepSeek's approach emphasizes using a large amount of free data for pre-training and fine-tuning with self-generated data, which has been recognized as a cost-effective strategy [5][6] Response to Criticism - DeepSeek addressed accusations from U.S. officials regarding the alleged illegal acquisition of advanced chips, clarifying that they used legally procured H800 chips and acknowledging prior use of A100 chips for smaller model experiments [4][5] - The company defended its use of "distillation" technology, which is a common practice in AI, asserting that it enhances model performance while reducing costs [5][6] Competitive Landscape - The success of DeepSeek-R1 demonstrates that AI competition is shifting from merely having the most GPUs to achieving more with fewer resources, thus altering the competitive dynamics in the industry [6][7] - Other AI models, such as OpenAI's GPT-4 and Google's Gemini, still hold advantages in certain areas, but DeepSeek's model has set a new standard for cost-effective high-performance AI [6][7]
DeepSeek团队梁文锋论文登上《自然》封面
Zheng Quan Shi Bao Wang· 2025-09-19 04:46
Core Viewpoint - The research paper on the DeepSeek-R1 reasoning model, led by Liang Wenfeng, demonstrates that the reasoning capabilities of large language models (LLMs) can be enhanced through pure reinforcement learning, reducing the need for human input in performance improvement [1] Group 1 - The study indicates that LLMs do not need to rely on human examples or complex instructions, as they can autonomously learn to generate reasoning processes through trial-and-error reinforcement learning [1] - The AI exhibits self-reflection, which is considered a significant indication of artificial intelligence exploring cognitive pathways beyond human thinking [1]