DeepSeek
Search documents
DeepSeek,打破历史!中国AI的“Nature时刻”
Zheng Quan Shi Bao· 2025-09-18 05:24
Core Insights - The DeepSeek-R1 inference model research paper has made history by being the first Chinese large model research to be published on the cover of the prestigious journal Nature, marking a significant recognition of China's AI technology in the international scientific community [1][2] - Nature's editorial highlighted that DeepSeek has broken the gap of independent peer review for mainstream large models, which has been lacking in the industry [2] Group 1: Research and Development - The DeepSeek-R1 model's research paper underwent a rigorous peer review process involving eight external experts over six months, emphasizing the importance of transparency and reproducibility in AI model development [2] - The paper disclosed significant details about the training costs and methodologies, including a total training cost of $294,000 (approximately 2.09 million RMB) for R1, achieved using 512 H800 GPUs over 198 hours [3] Group 2: Model Performance and Criticism - DeepSeek addressed initial criticisms regarding the "distillation" method used in R1, clarifying that all training data was sourced from the internet without intentional use of outputs from proprietary models like OpenAI's [3] - The R1 model has been recognized for its cost-effectiveness compared to other inference models, which often incur training costs in the tens of millions [3] Group 3: Future Developments - There is significant anticipation regarding the release of the R2 model, with speculation that delays may be due to computational limitations [4] - The recent release of DeepSeek-V3.1 has introduced a mixed inference architecture and improved efficiency, indicating a step towards the "Agent" era in AI [4][5] - DeepSeek's emphasis on using UE8M0 FP8 Scale parameter precision in V3.1 suggests a strategic alignment with domestic AI chip development, potentially enhancing the performance of future models [5]
DeepSeek,打破历史!中国AI的“Nature时刻”
证券时报· 2025-09-18 04:51
Core Viewpoint - The article highlights the significant achievement of the DeepSeek-R1 inference model, which has become the first Chinese large model research to be published in the prestigious journal Nature, marking a milestone for China's AI technology on the global stage [1][2]. Group 1: Publication and Recognition - DeepSeek-R1's research paper was published in Nature after a rigorous peer review process involving eight external experts, breaking the trend where major models like those from OpenAI and Google were released without independent validation [2][3]. - Nature's editorial praised DeepSeek for filling the gap in the independent peer review of mainstream large models, emphasizing the importance of transparency and reproducibility in AI research [3]. Group 2: Model Training and Cost - The training of the R1 model utilized 512 H800 GPUs for 198 hours and 80 hours respectively, with a total training cost of $294,000 (approximately 2.09 million RMB), which is significantly lower compared to other models that can cost tens of millions [3][4]. - The paper disclosed detailed training costs and methodologies, addressing previous criticisms regarding data sourcing and the "distillation" process, asserting that all data was sourced from the internet without intentional use of proprietary models [4]. Group 3: Future Developments and Innovations - There is ongoing speculation about the release of the R2 model, with delays attributed to computational limitations, while the recent release of DeepSeek-V3.1 has sparked interest in the advancements leading to R2 [5][6]. - DeepSeek-V3.1 introduces a mixed inference architecture and improved efficiency, indicating a shift towards the "Agent" era in AI, and highlights the use of UE8M0 FP8 Scale parameter precision, which is designed for upcoming domestic chips [6][7]. - The adoption of FP8 parameter precision is seen as a strategic move to enhance the performance of domestic AI chips, potentially revolutionizing the landscape of AI model training and inference in China [7].
DeepSeek首次回应“蒸馏OpenAI”质疑
Di Yi Cai Jing· 2025-09-18 04:34
Core Insights - DeepSeek's research paper, "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning," has been published in the prestigious journal Nature, highlighting significant advancements in AI reasoning capabilities [1][11]. Group 1: Research and Development - The initial version of DeepSeek's paper was released on arXiv in January, and the Nature publication includes more detailed model specifications and reduced anthropomorphism in descriptions [5]. - DeepSeek-R1's training cost was reported to be $294,000, with specific costs for different components outlined, including $202,000 for DeepSeek-R1-Zero training and $82,000 for SFT data creation [9]. - The training utilized A100 GPUs for smaller models and expanded to 660 billion parameters for the R1 model, demonstrating a scalable approach to model development [8][10]. Group 2: Model Performance and Validation - DeepSeek-R1 has become the most popular open-source inference model globally, with over 10.9 million downloads on Hugging Face, marking it as the first mainstream large language model to undergo peer review [11]. - The research emphasizes that significant reasoning capabilities can be achieved through reinforcement learning without relying heavily on supervised fine-tuning, which is a departure from traditional methods [13]. - The model's training involved a reward mechanism that encourages correct reasoning, allowing it to self-validate and improve its performance on complex tasks [13]. Group 3: Industry Implications - The findings from DeepSeek's research could set a precedent for future AI model development, particularly in enhancing reasoning capabilities without extensive data requirements [11][13]. - The independent peer review process adds credibility to the model's performance claims, addressing concerns about potential manipulation in AI benchmarking [11].
刚刚集体爆发,人工智能四大重磅彻底引爆
Zheng Quan Shi Bao· 2025-09-18 04:00
人工智能迎来重磅驱动! 今日早盘,人工智能相关板块一骑绝尘。相关ETF大涨,半导体ETF大涨超4%,科创AI ETF大涨超3%,个股涨停潮再现。不过,市场依然是结构性演 绎,盘面涨少跌多。 分析人士认为,目前市场的特点还是由科技驱动。今天,人工智能有四大重磅集体出现:一是DeepSeek-R1开创历史,梁文锋论文登上《自然》封面;二 是马斯克表示,xAI有机会通过GROK 5达到AGI;三是华为预测,2035年全社会的算力总量将增长10万倍;四是芯片自主可控逻辑被相关消息持续驱动。 集体爆发 在光模块、PCB等受限的背景之下,芯片早盘集中爆发。中芯国际A股早盘大涨超7%,华虹半导体涨幅也一度扩大至7%,寒武纪一度大涨超4%;半导体 ETF一度大涨近6%,科创AI ETF一度涨近4%。 消息面上,由DeepSeek团队共同完成、梁文锋担任通讯作者的DeepSeek-R1推理模型研究论文,登上了国际权威期刊《自然(Nature)》的封面。与今年1 月发布的DeepSeek-R1的初版论文相比,本次论文披露了更多模型训练的细节,并正面回应了模型发布之初的蒸馏质疑。DeepSeek-R1也是全球首个经过 同行评审的主 ...
梁文锋论文登上《自然》封面;李飞飞放出3D AI新成果
2 1 Shi Ji Jing Ji Bao Dao· 2025-09-18 03:00
Group 1: AI and Technology Developments - DeepSeek's research paper on the DeepSeek-R1 reasoning model has been published on the cover of the prestigious journal Nature, marking it as the first mainstream large language model to undergo peer review [2] - Stanford professor Fei-Fei Li's startup World Labs launched a new 3D AI project called Marble, which generates vast 3D environments from photos, although it still faces challenges in commercial application [3] - Microsoft plans to invest $30 billion in the UK by 2028 to build AI infrastructure, including a supercomputer with over 23,000 advanced GPUs, alongside investments from Nvidia, Google, OpenAI, and Salesforce totaling over $40 billion [4] - Huawei released a report predicting that by 2035, AI will drive significant technological advancements, with a projected 100,000-fold increase in total computing power and a 500-fold increase in AI storage capacity [5] - The World Trade Organization forecasts that AI could boost global trade by nearly 40% and GDP by 12-13% by 2040, emphasizing the need for balanced access to AI technology across countries [6] Group 2: Industry Collaborations and Investments - Lyft and Waymo announced a partnership to launch fully autonomous ride-hailing services in Nashville by 2026, utilizing Lyft's Flexdrive for fleet management [8] - Dongshan Precision announced its AI strategy focusing on high-end PCB and optical module development to meet the surging demand for AI-driven optical chips, particularly in the 800G and above market [9] - AI chip startup Groq completed a $750 million funding round, achieving a post-money valuation of $6.9 billion, with participation from various investment firms [10] - Alibaba strategically invested in Hello Robotaxi, aiming to enhance collaboration in smart driving models and computing platforms, accelerating the commercialization of Robotaxi services [11]
X @外汇交易员
外汇交易员· 2025-09-18 02:30
DeepSeek-R1论文登上Nature期刊封面,提到的是DeepSeek今年1月在arxiv发布的论文《DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning》,通讯作者为梁文锋。Nature编辑认为,同行评审模式对AI大语言模型发展有益,因为基准测试是可被操控,将模型的设计、方法论和局限性交由独立的外部专家审视,能够有效“挤水分”,抑制AI行业炒作。🗒️DeepSeek-R1被认为是首个通过权威学术期刊同行评审的大语言模型。 ...
下棋比智商!8 大 AI 模型上演棋盘大战,谁能称王?
AI前线· 2025-09-18 02:28
Core Insights - Kaggle has launched the Kaggle Game Arena in collaboration with Google DeepMind, focusing on evaluating AI models through strategic games [2] - The platform provides a controlled environment for AI models to compete against each other, ensuring fair assessments through an all-play-all format [2][3] - The initial participants include eight prominent AI models from various companies, highlighting the competitive landscape in AI development [2] Group 1 - The Kaggle Game Arena shifts the focus of AI evaluation from language tasks and image classification to decision-making under rules and constraints [3] - This benchmarking approach helps identify strengths and weaknesses of AI systems beyond traditional datasets, although some caution that controlled environments may not fully replicate real-world complexities [3] - The platform aims to expand beyond chess to include card games and digital games, testing AI's strategic reasoning capabilities [5] Group 2 - AI enthusiasts express excitement about the potential of the platform to reveal the true capabilities of top AI models in competitive scenarios [4][5] - The standardized competition mechanism of Kaggle Game Arena establishes a new benchmark for assessing AI models, emphasizing decision-making abilities in competitive environments [5]
梁文锋执笔的R1论文登上Nature封面!首次回应外界三大质疑
AI前线· 2025-09-18 02:28
Core Viewpoint - The article highlights the significant breakthrough of DeepSeek's AI model, DeepSeek-R1, which has successfully passed peer review and is recognized as the first large language model to achieve this milestone, marking a notable advancement for domestic AI research on the global stage [3][8]. Summary by Sections Model Development and Features - DeepSeek-R1 utilizes reinforcement learning (RL) to develop reasoning capabilities without relying on extensive human-annotated data, showcasing a novel approach in AI model training [3][12]. - The model was built on DeepSeek-V3 Base, with a focus on rewarding correct predictions to enhance the generation of longer and more logical responses [3][12]. - The training cost for DeepSeek-R1 was approximately $294,000, significantly lower than competitors that often spend tens of millions [6][12]. Peer Review Process - The peer review process for DeepSeek-R1 involved eight external experts over five months, resulting in a comprehensive review document that was three times the length of the original paper [9][12]. - The review addressed various aspects, including originality, methodology, and robustness, leading to improvements in the final published version [9][12]. Data and Safety Measures - The pre-training data for DeepSeek-V3 Base was sourced entirely from the internet, with a significant effort made to clean the data to avoid contamination, removing around 6 million potentially polluted samples [6][12]. - DeepSeek-R1 has implemented external risk control mechanisms and real-time audits, demonstrating superior safety performance compared to other mainstream models like Claude-3.7-Sonnet and GPT-4o [6][12]. Impact and Future Directions - The innovative use of pure reinforcement learning in DeepSeek-R1 is expected to influence future research in large language models, with many researchers looking to apply similar methods to enhance reasoning capabilities across various domains [12][14]. - Despite some concerns regarding the transparency of training data composition, the model has shown competitive performance in balancing accuracy and cost in scientific task challenges [14][12].
DeepSeek论文登上《自然》封面,R1成为首个严格学术审查大模型
Xin Lang Cai Jing· 2025-09-18 02:23
Core Insights - DeepSeek's R1 model has been recognized as the first major language model to be peer-reviewed and published in the prestigious journal Nature, marking a significant milestone in AI research [1][2] - The R1 model achieved over 10.9 million downloads on Hugging Face, making it the most popular open-source inference model globally [2] - DeepSeek's innovative approach utilizes pure reinforcement learning to enhance reasoning capabilities, diverging from traditional human-imitation methods [2][3] Company Developments - DeepSeek's R1 model was developed with a training cost of only $294,000, significantly lower than the costs associated with training AI models by OpenAI and Google, which can reach millions [2] - The company released an upgraded version, DeepSeek-V3.1, which features a mixed reasoning architecture, improved thinking efficiency, and enhanced agent capabilities [3] - DeepSeek was founded in 2023 in Hangzhou, backed by the quantitative firm Huansquare, with a team composed of experts from top universities and international institutions [3] Industry Context - The publication of DeepSeek's research is seen as a critical step in addressing the rampant speculation and unverified claims within the AI industry, emphasizing the importance of independent peer review [3] - The recognition of DeepSeek's work by Nature highlights China's advancements in foundational research in large models, contributing to the global AI landscape [2]
DeepSeek-R1登上Nature封面:朝着AI透明化迈出的可喜一步
3 6 Ke· 2025-09-18 02:02
开源人工智能(AI)的价值正获得更广泛的认可。 刚刚,DeepSeek-R1 论文以封面文章的形式登上了权威科学期刊 Nature,DeepSeek 创始人兼 CEO 梁文峰为该论文的通讯作者。 论文链接:https://www.nature.com/articles/s41586-025-09422-z 研究团队假设,人类定义的推理模式可能会限制模型的探索,而无限制的强化学习(RL)训练可以更好地激励大语言模型(LLM)中新推理能力的涌 现。 他们通过实验证明,LLM 的推理能力可以通过纯 RL 来提升,从而减少增强性能所需的人类输入工作量,且在数学、编程竞赛和 STEM 领域研究生水平 问题等任务上,比经传统方法训练的 LLM 表现更好。 DeepSeek-R1 推出后,得到了全球开发者的广泛好评,截至发文前,其在 GitHub 上的 star 数已经达到了 91.1k。 在一篇同期发表的观点与评论文章中,卡内基梅隆大学助理教授Daphne Ippolito和他的博士生张益铭(现为 Anthropic 的 LLM 安全和对齐研究员)评价 道: "DeepSeek-R1 已从一个强大但不透明的解决方案寻找者 ...