Seek .(SKLTY)
Search documents
DeepSeek刷屏论文背后:除了梁文锋,还有一个18岁中国高中生,曾写出神级提示词
3 6 Ke· 2025-09-19 03:32
Core Insights - DeepSeek has published a paper in Nature, showcasing advancements in reasoning within large language models (LLMs) through reinforcement learning, which includes richer implementation details and experimental analysis compared to earlier versions [2][4][38] - The paper highlights the contributions of notable researchers, including Liang Wenfeng, Tu Jinhao, and Luo Fuli, indicating a strong presence of Chinese AI talent in global academic circles [4][38] Group 1 - The Nature publication represents a significant achievement for DeepSeek, marking a historical moment for Chinese AI development on a global stage [38] - The paper emphasizes the importance of the reasoning process in AI models, suggesting that a comprehensive thinking approach is crucial for improving the quality of AI responses [30][38] - The research team includes young talents, such as Tu Jinhao, who has gained recognition for innovative approaches in AI competitions and model enhancements [6][30] Group 2 - Luo Fuli, another key contributor, has a strong academic background and has been involved in significant projects, including leading the development of multilingual pre-trained models at Alibaba [34][36] - The publication reflects a broader trend of increasing representation of Chinese AI researchers in top-tier academic publications, enhancing the visibility of China's contributions to the AI field [38] - The collaborative nature of the research team underscores the importance of teamwork in achieving significant milestones in AI research [38]
AI医学的“DeepSeek时刻”快来了?
Di Yi Cai Jing· 2025-09-19 00:32
Core Insights - The article highlights the emergence of AI technologies in the pharmaceutical and medical fields, particularly focusing on the advancements made by Chinese AI company DeepSeek and its large model R1, which has gained recognition in the scientific community [2] - The integration of AI in drug discovery and clinical applications is accelerating, with significant investments from major pharmaceutical companies aiming to revolutionize the drug development process [4][5] Group 1: AI in Drug Discovery - Major pharmaceutical companies, including Bristol-Myers Squibb and Sanofi, are investing billions in AI drug discovery, hoping to achieve breakthroughs that will transform the drug development process [4] - Medidata's data indicates that the proportion of clinical trials initiated by Chinese companies has surged from approximately 3% to 30% by 2024, positioning China as the second-largest clinical trial market globally [4] - AI is expected to drive a new wave of drug development, becoming a crucial force in the transformation of new drug research [4] Group 2: AI in Medical Applications - The "Meta-Medical" laboratory, launched by Zhongshan Hospital affiliated with Fudan University, aims to develop AI agents and apply large model technologies to enhance medical knowledge digitization and productization of diagnostic capabilities [6] - AI is changing the paradigm of diagnosis and treatment, with significant advancements in areas such as heart disease risk prediction and real-time monitoring through wearable devices [6] - The successful application of AI in specific medical fields has reached clinical levels, exemplified by the monitoring of intermittent atrial fibrillation using wearable technology [6] Group 3: Challenges and Ethical Considerations - Despite the potential of AI in drug discovery, challenges remain, including a 90% failure rate in clinical trials and the need to address complex biological issues and regulatory hurdles [5] - Ethical considerations are paramount, with the responsibility for medical decisions still resting with physicians, who must ensure that AI technologies are used safely and effectively in clinical settings [7]
DeepSeek 创始人梁文锋在《自然》杂志回应质疑,R1 训练真 29.4 万美金
Xin Lang Cai Jing· 2025-09-19 00:03
Core Insights - DeepSeek-R1 has made a significant impact in the AI field by being featured on the cover of Nature, highlighting its innovative approach to enhancing reasoning capabilities in large language models (LLMs) through reinforcement learning (RL) [1][3][5]. Group 1: Achievements and Recognition - The paper "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning" was published in January and has now been recognized on the cover of a leading journal, Nature [3]. - DeepSeek-R1 has become the most popular model on Hugging Face after its open-source release, achieving over 10.9 million downloads [5]. - The training cost for DeepSeek-R1 was remarkably low at $294,000, which is significantly less than the costs incurred by competitors like OpenAI and Google [6][7]. Group 2: Training Methodology - DeepSeek-R1 utilizes a novel RL framework that focuses solely on the task format and reward signals based on the correctness of the final answer, allowing for a more organic development of reasoning capabilities [10]. - The model's reasoning accuracy improved dramatically from 15.6% to 77.9% during training, with a peak accuracy of 86.7% when combined with "self-consistent decoding" techniques [10]. Group 3: Self-Evolution and Advanced Strategies - The model exhibited self-evolution behaviors, such as increasing the length of generated text and employing advanced reasoning strategies like self-reflection and systematic exploration of alternative solutions [12][14]. - A notable "Aha Moment" was observed when the model began using the word "wait" more frequently, indicating a shift in its reasoning approach [15][17]. Group 4: Future Development Plans - To address the limitations of DeepSeek-R1, a multi-stage refinement plan has been initiated, which includes cold starting with high-quality conversational data, followed by multiple rounds of RL and supervised fine-tuning [18][19]. - The model's performance has improved by 17%-25% on various benchmarks after undergoing this multi-stage training process [21]. Group 5: Algorithm and Reward System - DeepSeek employs the GRPO (Group Relative Policy Optimization) algorithm, which optimizes model performance by evaluating a group of answers rather than a single best answer, thus reducing resource consumption while maintaining stability [23][24]. - A dual reward system has been established, incorporating both rule-based rewards for reasoning tasks and model-based rewards for general tasks, ensuring the model aligns with human preferences while maintaining its reasoning capabilities [25][26]. Group 6: Challenges and Limitations - Despite its advancements, DeepSeek-R1 faces challenges in structured outputs and tool usage, and it is sensitive to prompts, which limits its effectiveness in complex scenarios [35][37]. - The potential for reward hacking exists, particularly in subjective tasks, which could undermine the model's performance if the reward signals are not robust [37].
DeepSeek团队发表重磅论文,《自然》配发社论狂赞呼吁同行效仿
Yang Zi Wan Bao Wang· 2025-09-18 13:19
Group 1 - The DeepSeek-R1 inference model research paper has been published on the cover of the prestigious journal Nature, marking it as the first mainstream large language model (LLM) to undergo peer review, which is significant for AI model development [2][4] - The paper reveals more details about the model's training compared to its initial version released in January, indicating that the reasoning capabilities of LLMs can be enhanced through pure reinforcement learning, reducing the human input required for performance improvement [2][9] - Since its release in January, DeepSeek-R1 has become the most downloaded product for solving complex problems on the platform, and it has undergone evaluation by eight experts on originality, methodology, and robustness [9] Group 2 - Nature's editorial emphasizes the importance of peer review for AI models, noting that almost all mainstream large models have not undergone independent peer review until DeepSeek broke this gap [4][6] - Peer review helps clarify the workings of LLMs and assess whether they truly achieve their claimed functionalities, which is particularly crucial given the significant implications and potential risks associated with LLMs [6][10] - The editorial calls for other AI companies to follow DeepSeek's example, suggesting that if this practice becomes a trend, it could greatly promote the healthy development of the AI industry [10]
9.18犀牛财经晚报:生猪产能调控超预期 DeepSeek首次回应蒸馏OpenAI质疑
Xi Niu Cai Jing· 2025-09-18 10:30
Group 1: Banking and Financial Products - Several private banks, including Suzhou Bank and Huari Bank, have launched large-denomination certificates of deposit with interest rates exceeding 2%, contrasting with the declining rates of state-owned and joint-stock banks [1] - Huari Bank introduced two products with interest rates of 2.15% for 18-month deposits and 2.35% for 2-year deposits, both requiring a minimum subscription of 200,000 yuan [1] - The high-interest products are primarily aimed at customer acquisition and enhancing retail market competitiveness, rather than being a sustainable long-term strategy [1] Group 2: Agriculture and Livestock - The Ministry of Agriculture and Rural Affairs has mandated leading pig farming companies to reduce production capacity, including cutting the number of breeding sows and controlling the weight of pigs at around 120 kg [2] - This marks the first time the National Development and Reform Commission has explicitly required a reduction in the number of breeding sows, indicating a shift in regulatory focus [2] - Financial measures are being implemented alongside production controls, such as restricting credit for expanding pig farming capacity and reducing subsidies that encourage production growth [2] Group 3: Technology and Robotics - The Ministry of Science and Technology is promoting the accelerated application of humanoid robots in sectors like automotive manufacturing, logistics, and power inspection, laying a foundation for a trillion-dollar industry [1] - Significant advancements have been made in key technologies such as multi-modal perception and brain-machine interfaces, which have already benefited patients with paralysis and blindness [1] Group 4: Pharmaceuticals - Chongqing-based Runsheng Pharmaceutical has received approval for its inhalation powder product, fluticasone propionate, marking a significant breakthrough in the high-end inhalation powder market [2] Group 5: IPO and Investment - Zijin Mining International is set to attract major investors like GIC and Millennium Management in its upcoming Hong Kong IPO, which is expected to raise over $3 billion [5] - The IPO is anticipated to be the largest globally since May, with cornerstone investors likely to subscribe to about half of the shares [5] Group 6: Real Estate and Construction - Tian Di Yuan has successfully acquired a state-owned construction land use right in Xi'an for 2.015 billion yuan [10] - Palm Holdings has won a bid for a high-standard farmland construction project in Lankao County, valued at 433 million yuan, which represents 14.12% of its projected annual revenue [9] Group 7: Corporate Financial Issues - Sunac Real Estate has been ordered to execute payments totaling over 920 million yuan due to various legal disputes, contributing to its extensive financial liabilities exceeding 45.4 billion yuan [6] - He Shun Technology has received a warning from the Zhejiang Securities Regulatory Bureau for failing to disclose government subsidies and shareholder contributions in a timely manner [7]
DeepSeek 首登《自然》封面:中国大模型创造新历史,做了 OpenAI 不敢做的事
3 6 Ke· 2025-09-18 09:56
Core Insights - DeepSeek's AI model, R1, has gained significant recognition by being featured on the cover of Nature, a prestigious scientific journal, highlighting its impact in the AI industry [2][10][12] - The training cost for R1 was notably low at $294,000, which contrasts sharply with the multi-million dollar investments typical for models from companies like OpenAI [7][48] - The model's development process involved rigorous peer review, setting a new standard for transparency and scientific validation in AI [11][15][16] Group 1: Model Development and Training - DeepSeek R1's training process was detailed in a paper published on arXiv, which was later expanded upon in the Nature article, showcasing a comprehensive methodology [6][7] - The model was trained using a pure reinforcement learning framework, allowing it to develop reasoning capabilities without relying on human-annotated data [19][41] - R1 achieved an impressive accuracy of 77.9% in the AIME 2024 math competition, surpassing human average scores and even outperforming GPT-4 in certain tasks [23][31] Group 2: Peer Review and Industry Impact - The peer review process for R1 involved independent experts scrutinizing the model, which is a departure from the typical practices of major AI companies that often do not submit their models for academic evaluation [10][11][15] - Nature's editorial team has called for other companies to submit their models for peer review, emphasizing the importance of transparency and accountability in AI development [15][16] - The recognition from Nature not only validates R1's scientific contributions but also positions DeepSeek as a leader in the push for more rigorous standards in AI research [12][50] Group 3: Technical Innovations - R1's architecture is based on a mixture of experts (MoE) model with 671 billion parameters, which was pre-trained on a vast dataset of web pages and e-books [25] - The model's training involved a unique approach where it was rewarded solely based on the correctness of its answers, fostering an environment for self-reflection and dynamic adjustment during problem-solving [29][38] - The final version of R1 was developed through a multi-stage training process that combined reinforcement learning with supervised fine-tuning, enhancing both reasoning and general capabilities [39][47]
DeepSeek紧急声明!
Zheng Quan Shi Bao· 2025-09-18 09:26
Core Viewpoint - DeepSeek has issued a statement regarding fraudulent activities where individuals impersonate the company or its employees to collect fees under the guise of "computing power leasing" and "equity financing," which severely harms user rights and the company's reputation [1][2]. Group 1 - DeepSeek has never requested users to make payments to personal or unofficial accounts, and any such requests are considered scams [2]. - Any activities conducted under the company's name for "computing power leasing" or "financing" are illegal, and the company will pursue legal action against such actions [2]. - Users are advised to obtain information through official channels, as the company's official website and app products are currently free [2]. Group 2 - Since the release of the R1 model earlier this year, DeepSeek has established itself as a benchmark for open-source models globally [2]. - On September 17, DeepSeek's research paper on the R1 inference model was featured on the cover of the prestigious journal Nature, marking a significant achievement for Chinese AI research [2]. - This paper, co-authored by the DeepSeek team with Liang Wenfeng as the corresponding author, presents important findings on enhancing large model inference capabilities solely through reinforcement learning [2]. - Nature's editorial praised DeepSeek for breaking the gap in independent peer review for mainstream large models, highlighting the significance of this achievement in the international scientific community [2].
登上《自然》!DeepSeek-R1训练方法发布
Ke Ji Ri Bao· 2025-09-18 08:39
Core Insights - The DeepSeek-AI team has published a new open-source AI model, DeepSeek-R1, which utilizes a large-scale reasoning model training method to enhance the reasoning capabilities of large language models (LLMs) through pure reinforcement learning, thereby reducing the human input required for performance enhancement [1] Group 1: Model Performance - DeepSeek-R1 outperforms traditionally trained LLMs in tasks related to mathematics, programming competitions, and graduate-level STEM problems [1] - The model achieved scores of 77.9% and 79.8% in mathematical benchmark tests for DeepSeek-R1-Zero and DeepSeek-R1, respectively, demonstrating superior performance in programming competitions and graduate-level biology, physics, and chemistry problems [1] Group 2: Training Methodology - The model incorporates a deep training phase under human supervision to optimize the reasoning process, utilizing reinforcement learning instead of human examples to develop reasoning steps, which reduces training costs and complexity [1] - The team emphasizes that the model receives a template to generate reasoning processes after being shown high-quality problem-solving cases, reinforcing learning through problem-solving rewards [1] Group 3: Future Research Directions - Future research may focus on optimizing the reward process to ensure more reliable reasoning and task outcomes [1]
DeepSeek,严正声明!
Zhong Guo Ji Jin Bao· 2025-09-18 08:37
Core Viewpoint - DeepSeek has issued a statement regarding fraudulent activities where criminals impersonate the company or its employees to scam users, severely harming user rights and the company's reputation [1][2]. Group 1: Fraudulent Activities - Criminals have been using forged materials to solicit payments from users under the guise of "computing power leasing" and "equity financing" [1]. - DeepSeek emphasizes that it has never requested users to make payments to personal or unofficial accounts, and any such requests are fraudulent [2]. - The company warns users to verify information through its official website and certified accounts, as all official services are currently free [2]. Group 2: Company Background - DeepSeek was established in 2023 and is incubated by the well-known quantitative investment firm, Huansheng Quantitative [3]. - The founding team is led by quantitative expert Liang Wenfeng and includes top research talents from prestigious universities and experienced technical experts from international institutions [3]. - Recently, DeepSeek's research paper, DeepSeek-R1, was published on the cover of the prestigious journal Nature, marking it as the first major language model to undergo peer review [3].
训练成本29.4万美元,DeepSeek-R1登Nature封面,首个通过权威期刊同行评审的主流大模型获好评
3 6 Ke· 2025-09-18 07:55
Core Insights - DeepSeek-R1's research results have been published in Nature, marking it as the first mainstream large model to undergo peer review by a reputable journal, which has sparked significant discussion in the academic community [1][14][17] - The training cost of DeepSeek-R1 is reported to be only $294,000, significantly lower than the industry standard of tens of millions for leading models, despite an investment of approximately $6 million in the foundational LLM [1][2][17] Training Costs - The training costs for DeepSeek-R1 are broken down as follows: - DeepSeek-R1-Zero: $202,000 - SFT data creation: $10,000 - DeepSeek-R1: $82,000 - Total: $294,000 - The training utilized 648 H800 GPUs over approximately 198 hours for DeepSeek-R1-Zero and around 80 hours for DeepSeek-R1 [2] Reinforcement Learning and Reasoning Capabilities - The model employs Group Relative Policy Optimization (GRPO) to enhance reasoning capabilities without traditional supervised fine-tuning, allowing for more exploratory learning [3][4] - DeepSeek-R1-Zero demonstrates complex reasoning behaviors, generating longer responses that incorporate verification and exploration of different solutions [4][6] Performance Metrics - DeepSeek-R1-Zero achieved a pass@1 score of 77.9% in the AIME 2024 math competition, with further improvements to 86.7% using self-consistent decoding strategies, surpassing human average performance [6][8] - The model also excelled in programming competitions and graduate-level questions in biology, physics, and chemistry, validating the effectiveness of reinforcement learning in enhancing reasoning capabilities [6] Development Pipeline - The development of DeepSeek-R1 involved multiple stages, starting from data collection based on human-like dialogue to reinforcement learning and sampling, ultimately enhancing the model's utility and safety [9][11] - Experimental results indicate significant improvements in instruction execution across various development stages, with DeepSeek-R1 outperforming its predecessors in benchmark tests [11][13] Industry Impact - The peer review of DeepSeek-R1 is seen as a positive trend for AI research, promoting transparency and standardization in the field, which has been lacking for many mainstream AI models [14][16][17]