Workflow
大语言模型
icon
Search documents
DeepSeek打破历史!中国AI的“Nature时刻”
Zheng Quan Shi Bao· 2025-09-18 07:29
Core Insights - The DeepSeek-R1 inference model research paper has made history by being the first Chinese large model research to be published in the prestigious journal Nature, marking a significant recognition of China's AI technology on the global scientific stage [1][2] - Nature's editorial highlighted that DeepSeek has broken the gap of independent peer review for mainstream large models, which has been lacking in the industry [2] Group 1: Research and Development - The DeepSeek-R1 model's research paper underwent a rigorous peer review process involving eight external experts over six months, emphasizing the importance of transparency and reproducibility in AI model development [2] - The paper disclosed significant details about the training costs and methodologies, including a total training cost of $294,000 (approximately 2.09 million RMB) for R1, achieved using 512 H800 GPUs [3] Group 2: Model Performance and Criticism - DeepSeek addressed initial criticisms regarding the "distillation" method used in R1, clarifying that all training data was sourced from the internet without intentional use of outputs from proprietary models like OpenAI's [3] - The R1 model's training duration was 198 hours for R1-Zero and 80 hours for R1, showcasing a cost-effective approach compared to other models that often exceed tens of millions of dollars [3] Group 3: Future Developments - There is significant anticipation regarding the release of the R2 model, with speculation that delays may be due to computational limitations [4] - The recent release of DeepSeek-V3.1 indicates advancements towards the "Agent" era, featuring a mixed inference architecture and improved efficiency, which has sparked interest in the upcoming R2 model [4][5] Group 4: Industry Impact - DeepSeek's adoption of UE8M0 FP8 Scale parameter precision in V3.1 suggests a shift towards utilizing domestic AI chips, potentially accelerating the development of China's computing ecosystem [5] - The collaboration between software and hardware in DeepSeek's models is seen as a new paradigm in the AI wave, with expectations for significant performance improvements in domestic computing chips [5]
DeepSeek首次回应“蒸馏OpenAI”质疑
第一财经· 2025-09-18 05:34
Core Viewpoint - DeepSeek's R1 model has gained significant attention after being published in the prestigious journal "Nature," showcasing its ability to enhance reasoning capabilities through reinforcement learning without relying heavily on supervised data [3][11]. Group 1: Model Development and Training - The training cost for the DeepSeek-R1 model was approximately $294,000, with specific costs for different components detailed as follows: R1-Zero training cost was $202,000, SFT dataset creation cost was $10,000, and R1 training cost was $82,000 [10]. - DeepSeek-R1 utilized 64×8 H800 GPUs for training, taking about 198 hours for R1-Zero and around 80 hours for R1 [10]. - The total training cost, including the earlier V3 model, remains significantly lower than competitors, totaling around $6 million for V3 and $294,000 for R1 [10]. Group 2: Model Performance and Validation - DeepSeek's approach allows for significant performance improvements in reasoning capabilities through large-scale reinforcement learning, even without supervised fine-tuning [13]. - The model's ability to self-validate and reflect on its answers enhances its performance on complex programming and scientific problems [13]. - The research indicates that the R1 model has become the most popular open-source reasoning model globally, with over 10.9 million downloads on Hugging Face [10]. Group 3: Industry Impact and Peer Review - The publication of the R1 model in "Nature" sets a precedent for transparency in AI research, addressing concerns about the reliability of benchmark tests and the potential for manipulation [11]. - The research emphasizes the importance of independent peer review in validating the capabilities of AI systems, which is crucial in an industry facing scrutiny over performance claims [11].
DeepSeek,打破历史!中国AI的“Nature时刻”
证券时报· 2025-09-18 04:51
Core Viewpoint - The article highlights the significant achievement of the DeepSeek-R1 inference model, which has become the first Chinese large model research to be published in the prestigious journal Nature, marking a milestone for China's AI technology on the global stage [1][2]. Group 1: Publication and Recognition - DeepSeek-R1's research paper was published in Nature after a rigorous peer review process involving eight external experts, breaking the trend where major models like those from OpenAI and Google were released without independent validation [2][3]. - Nature's editorial praised DeepSeek for filling the gap in the independent peer review of mainstream large models, emphasizing the importance of transparency and reproducibility in AI research [3]. Group 2: Model Training and Cost - The training of the R1 model utilized 512 H800 GPUs for 198 hours and 80 hours respectively, with a total training cost of $294,000 (approximately 2.09 million RMB), which is significantly lower compared to other models that can cost tens of millions [3][4]. - The paper disclosed detailed training costs and methodologies, addressing previous criticisms regarding data sourcing and the "distillation" process, asserting that all data was sourced from the internet without intentional use of proprietary models [4]. Group 3: Future Developments and Innovations - There is ongoing speculation about the release of the R2 model, with delays attributed to computational limitations, while the recent release of DeepSeek-V3.1 has sparked interest in the advancements leading to R2 [5][6]. - DeepSeek-V3.1 introduces a mixed inference architecture and improved efficiency, indicating a shift towards the "Agent" era in AI, and highlights the use of UE8M0 FP8 Scale parameter precision, which is designed for upcoming domestic chips [6][7]. - The adoption of FP8 parameter precision is seen as a strategic move to enhance the performance of domestic AI chips, potentially revolutionizing the landscape of AI model training and inference in China [7].
“这一空白终于被打破”,梁文锋论文登上《自然》封面
Guan Cha Zhe Wang· 2025-09-18 03:27
《科技日报》则在报道中介绍称,梁文锋参与的研究表明,大语言模型的推理能力可通过纯强化学习来 提升,从而减少增强性能所需的人类输入工作量。训练出的模型在数学和STEM领域研究生水平问题等 任务上,比传统训练的大语言模型表现更好。 DeepSeek-R1包含一个在人类监督下的深入训练阶段,以优化推理过程。梁文锋团队报告称,该模型使 用了强化学习而非人类示例来开发推理步骤,减少了训练成本和复杂性。DeepSeek-R1在被展示优质的 问题解决案例后,会获得一个模板来产生推理过程,即这一模型通过解决问题获得奖励,从而强化学习 效果。在评估AI表现的各项测试中,DeepSeek-R1-Zero和DeepSeek-R1的表现都十分优异。 据智通财经9月18日消息,由DeepSeek团队共同完成、梁文锋担任通讯作者的DeepSeek-R1推理模型研 究论文,登上了国际权威期刊《自然(Nature)》的封面。 与今年1月发布的DeepSeek-R1的初版论文相比,本次论文披露了更多模型训练的细节,并正面回应了 模型发布之初的蒸馏质疑。DeepSeek-R1也是全球首个经过同行评审的主流大语言模型。Nature评价 道:目前几 ...
梁文锋执笔的R1论文登上Nature封面!首次回应外界三大质疑
AI前线· 2025-09-18 02:28
Core Viewpoint - The article highlights the significant breakthrough of DeepSeek's AI model, DeepSeek-R1, which has successfully passed peer review and is recognized as the first large language model to achieve this milestone, marking a notable advancement for domestic AI research on the global stage [3][8]. Summary by Sections Model Development and Features - DeepSeek-R1 utilizes reinforcement learning (RL) to develop reasoning capabilities without relying on extensive human-annotated data, showcasing a novel approach in AI model training [3][12]. - The model was built on DeepSeek-V3 Base, with a focus on rewarding correct predictions to enhance the generation of longer and more logical responses [3][12]. - The training cost for DeepSeek-R1 was approximately $294,000, significantly lower than competitors that often spend tens of millions [6][12]. Peer Review Process - The peer review process for DeepSeek-R1 involved eight external experts over five months, resulting in a comprehensive review document that was three times the length of the original paper [9][12]. - The review addressed various aspects, including originality, methodology, and robustness, leading to improvements in the final published version [9][12]. Data and Safety Measures - The pre-training data for DeepSeek-V3 Base was sourced entirely from the internet, with a significant effort made to clean the data to avoid contamination, removing around 6 million potentially polluted samples [6][12]. - DeepSeek-R1 has implemented external risk control mechanisms and real-time audits, demonstrating superior safety performance compared to other mainstream models like Claude-3.7-Sonnet and GPT-4o [6][12]. Impact and Future Directions - The innovative use of pure reinforcement learning in DeepSeek-R1 is expected to influence future research in large language models, with many researchers looking to apply similar methods to enhance reasoning capabilities across various domains [12][14]. - Despite some concerns regarding the transparency of training data composition, the model has shown competitive performance in balancing accuracy and cost in scientific task challenges [14][12].
梁文锋发表Nature封面论文:揭开DeepSeek-R1背后的科学原理——强化学习激励大模型推理能力
生物世界· 2025-09-18 01:44
Core Viewpoint - The article discusses the development and capabilities of DeepSeek-R1, a reasoning model that significantly reduces computational costs while enhancing reasoning abilities in large language models (LLMs) through pure reinforcement learning [1][2]. Group 1: Model Development and Training - DeepSeek-R1 was launched by a startup in Hangzhou, China, on January 20, 2025, and has gained global attention for its strong reasoning capabilities and low computational requirements [1]. - The training cost for DeepSeek-R1 was only $294,000, which is significantly lower than similar models that often cost tens of millions [2]. - The model employs a pure reinforcement learning approach, minimizing reliance on human-annotated reasoning paths, which allows for more autonomous exploration of reasoning capabilities [6][10]. Group 2: Performance and Capabilities - DeepSeek-R1-Zero, a precursor to DeepSeek-R1, demonstrated remarkable performance improvements in reasoning tasks, achieving an average pass@1 score of 77.9% in the American Mathematics Invitational Exam (AIME) 2024, up from 15.6% [17]. - The model also excelled in programming competitions and graduate-level problems in biology, physics, and chemistry, showcasing its versatility [19]. - The research indicates that advanced reasoning behaviors, such as self-validation and reflection, emerged organically during the reinforcement learning process [29]. Group 3: Challenges and Limitations - Despite its strengths, DeepSeek-R1-Zero faces challenges such as poor readability and language mixing issues, particularly when responding in both English and Chinese [21]. - The model's performance in broader domains like writing and open-domain Q&A remains limited due to its focus on reasoning tasks during training [22]. - The article highlights potential ethical risks associated with enhanced reasoning capabilities, including vulnerability to jailbreak attacks and the generation of dangerous content [27][28].
梁文锋论文登上《自然》封面
财联社· 2025-09-18 00:49
Core Viewpoint - The DeepSeek-R1 inference model research paper, led by Liang Wenfeng, has been published in the prestigious journal Nature, marking a significant milestone in the field of large language models [1][4]. Group 1 - The latest paper provides more detailed insights into the model training process compared to the initial version released in January [4]. - DeepSeek-R1 is recognized as the first mainstream large language model to undergo peer review, addressing previous concerns regarding its distillation [4]. - Nature highlighted that most mainstream large models have not yet been independently peer-reviewed, and DeepSeek has filled this gap [4].
梁文锋论文登上《自然》封面
Mei Ri Jing Ji Xin Wen· 2025-09-18 00:42
(文章来源:每日经济新闻) 与今年1月发布的DeepSeek-R1的初版论文相比,本次论文披露了更多模型训练的细节,并正面回应了 模型发布之初的蒸馏质疑。DeepSeek-R1也是全球首个经过同行评审的主流大语言模型。Nature评价 道:目前几乎所有主流的大模型都还没有经过独立同行评审,这一空白"终于被DeepSeek打破"。 由DeepSeek团队共同完成、梁文锋担任通讯作者的DeepSeek-R1推理模型研究论文,登上了国际权威期 刊《自然(Nature)》第645期的封面。 ...
8点1氪:西贝回应“公筷喂狗”事件;美联储宣布降息25个基点;DeepSeek梁文锋论文登上《自然》封面
36氪· 2025-09-18 00:19
Group 1 - The incident at Xibei restaurant involved customers using restaurant utensils to feed a pet dog, raising concerns about dining safety [4] - The restaurant confirmed that all utensils used by the customers were discarded and a thorough disinfection of the premises was conducted [4] - Local authorities stated there are currently no legal grounds to penalize the restaurant for allowing pets, as the customer's actions were deemed personal behavior [4] Group 2 - The Federal Reserve announced a 25 basis point cut in the federal funds rate, marking its first rate decrease since December 2024 [4] Group 3 - NIO Group successfully completed a financing round of $1.16 billion, aimed at enhancing its technological capabilities and expanding charging infrastructure [20] - AI chip startup Groq raised $750 million in a new funding round, achieving a post-money valuation of $6.9 billion [20] - "Qingyun New Materials" announced the completion of a multi-hundred million C round financing to support the development of advanced materials [20] Group 4 - The month of September saw a significant increase in lemon prices, doubling from 7.83 yuan per kilogram to 15 yuan per kilogram over the past year, leading to supply shortages at some stores [15] - The mooncake industry in China is transitioning from seasonal demand to year-round consumption, with over 20,000 related enterprises currently registered [24]
早报|美联储宣布降息25个基点;清华学霸晒1.67亿元年薪引调查;多家餐饮店抹掉无预制菜字样;携程被约谈
虎嗅APP· 2025-09-18 00:17
Group 1 - The Federal Reserve announced a 25 basis point interest rate cut, bringing the target range to 4.00%-4.25%, aligning with market expectations [2][3] - This marks the first rate cut since December 2024, occurring after a 9-month interval [3] Group 2 - China Ping An clarified that recent rumors about relocating from Shanghai are unfounded, stating that the adjustments are regulatory compliance measures rather than a withdrawal from the city [4][5][6] - The company emphasized that its subsidiaries based in Shanghai will remain unchanged, and the adjustments pertain to employees returning to the Shenzhen headquarters [5][6] Group 3 - CATL announced that its sodium-ion batteries will have a range exceeding 500 kilometers and will begin mass production next year, targeting over 40% of the domestic passenger vehicle market [7][8] - The sodium-ion battery has a density of 175 Wh/kg and offers advantages in low-temperature performance and safety compared to lithium-ion batteries [7] Group 4 - Peak Group's chairman denied reports of widespread salary cuts, stating that the overall reduction is less than 10%, with adjustments primarily affecting high-salary positions and loss-making departments [16] - The company reported a loss of 130 million yuan in its direct sales business for the first seven months of the year, prompting the salary adjustments [16] Group 5 - The Tianjin Medical Insurance Consumables Directory will come into effect, including 3,062 types of medical consumables, with 1,896 classified as Class A, setting payment standards to reduce high prices [30]