DeepSeek
Search documents
DeepSeek:薛定谔式“凉”了?
新财富· 2025-08-06 08:03
Core Viewpoint - The article discusses the recent decline in the usage and market share of DeepSeek, questioning the validity of the reported statistics and emphasizing the importance of considering third-party API usage in evaluating its performance [2][4][10]. Summary by Sections Market Share and Usage Statistics - Reports from Semianalysis indicate that DeepSeek's market share has dropped to below 5%, with a significant decline noted since January [4][10]. - The statistics cited by Semianalysis primarily focus on the official API usage, potentially overlooking significant third-party integrations and deployments [10][12]. Third-Party API Usage - DeepSeek's third-party API calls have reportedly increased nearly 20 times since the release of versions V3 and R1, indicating sustained interest from developers [11][12]. - The article argues that the decline in official API usage does not reflect the overall demand for DeepSeek, as many applications integrate it without being captured in the official statistics [10][12]. Comparative Performance - Data from OpenRouter shows that DeepSeek V3 has a tokens consumption of 378 billion, ranking it third behind Claude Sonnet 4 and ahead of Google’s Gemini [17][22]. - Despite a decline in market share, DeepSeek maintains over 50% of the domestic B-end demand, indicating its strong position in the market [33]. User Preference and Community Engagement - A survey by Artificial Analysis found that 53% of respondents still prefer DeepSeek, placing it fourth among AI product providers [39]. - DeepSeek-R1 continues to lead in popularity on platforms like Hugging Face, indicating strong community support despite market fluctuations [44]. Industry Context and Future Outlook - The rapid evolution of AI technology suggests that a decline in DeepSeek's market share may not indicate a loss of relevance but rather reflects the dynamic nature of the industry [49]. - The article highlights the importance of open-source contributions from DeepSeek in promoting AI equity, contrasting it with other companies that are moving away from open-source models [49][50].
Claude 小升级就赢了OpenAI 9年“开源神作”?高强度推理直接歇菜、幻觉率高达50%,写作还被Kimi 2吊锤?
AI前线· 2025-08-06 04:25
Core Viewpoint - OpenAI has released its first open-source language model series, gpt-oss, which includes gpt-oss-120b and gpt-oss-20b, both of which are fully customizable and support structured output [2][3]. Model Specifications - gpt-oss-120b requires 80GB of memory to run, while gpt-oss-20b only needs 16GB [2]. - The models utilize a mixture of experts (MoE) architecture, activating 5.1 billion parameters per token for gpt-oss-120b and 3.6 billion for gpt-oss-20b, with total parameters of 117 billion and 21 billion respectively [9]. - Both models support a context length of up to 128k and are designed for efficient deployment on consumer-grade hardware [10]. Training and Performance - The training process for gpt-oss models combines reinforcement learning and techniques from OpenAI's advanced internal models, focusing on reasoning capabilities and efficiency [8]. - gpt-oss models have shown strong performance in reasoning tasks, with gpt-oss-120b performing comparably to OpenAI's proprietary models in core inference benchmarks [10]. Comparison with Competitors - Claude Opus 4.1 has demonstrated superior programming performance with a score of 74.5% in SWE-bench Verified programming evaluations, outperforming previous versions [5]. - Independent benchmark tests indicate that gpt-oss-120b is less intelligent than DeepSeek R1 and Qwen3 235B, although it has advantages in efficiency due to its smaller parameter size [13]. User Feedback and Limitations - Users have reported mixed experiences with gpt-oss models, noting that gpt-oss-120b is particularly unstable for coding tasks, while gpt-oss-20b performs better [6][17]. - The models exhibit a higher hallucination rate, with gpt-oss-120b and gpt-oss-20b generating hallucinations at rates of 49% and 53% respectively, significantly higher than OpenAI's previous models [16]. Open Source and Accessibility - gpt-oss models are released under the flexible Apache 2.0 license, making them accessible for various applications, including agent workflows and tool usage [11][10]. - The models are available for free download on Hugging Face, promoting wider adoption and experimentation within the developer community [2][3].
OpenAI时隔六年再开源
Cai Jing Wang· 2025-08-06 03:37
Core Insights - OpenAI has released two new open-source AI models, GPT-oss-120b and GPT-oss-20b, marking the first introduction of new open-source large language models since the release of GPT-2 in 2019 [1] - The release was initially planned for March but was delayed until August 6, following a global open-source movement sparked by DeepSeek earlier this year [1] - Both models are released under a permissive Apache 2.0 license, allowing businesses to use them commercially without prior payment or licensing [1] - OpenAI CEO Sam Altman described GPT-oss as a significant breakthrough, claiming it offers advanced open-weight inference capabilities comparable to o4-mini, and can be run locally on personal computers or smaller devices [1]
AI竞技场,归根到底只是一门生意
3 6 Ke· 2025-08-06 01:47
"XX发布最强开源大模型,多项基准测试全面超越XX等闭源模型!" "万亿参数开源模型XX强势登顶全球开源模型榜首!" "国产之光!XX模型在中文评测榜单拿下第一!" 随着AI时代的到来,各位的朋友圈、微博等社交平台是不是也常常被诸如此类的新闻刷屏了? 今天这个模型拿到了冠军,明天那个模型变成了王者。评论区里有的人热血沸腾,有的人一头雾水。 一个又一个的现实问题摆在眼前: 这些模型所谓的"登顶"比的是什么?谁给它们评分,而评分的依据又是什么?为什么每个平台的榜单座次都不一样, 到底谁更权威? 如果各位也产生了类似的困惑,说明各位已经开始从"看热闹"转向"看门道"。 本文之中,我们便来拆解一下不同类型"AI竞技场"——也就是大语言模型排行榜——的"游戏规则"。 01 类型一:客观基准测试(Benchmark),给AI准备的"高考" 人类社会中,高考分数是决定学生大学档次的最主要评判标准。 同样地,在AI领域,也有很多高度标准化的测试题,用来尽可能客观地衡量AI模型在特定能力上的表现。 因此,在这个大模型产品频繁推陈出新的时代,各家厂商推出新模型后,第一件事就是拿到"高考"考场上跑个分,是 骡子是马,拉出来遛遛。 ...
解决方案走在市民诉求前,“12345”热线探索超大城市智慧高效治理
Chang Jiang Ri Bao· 2025-08-06 00:21
Core Viewpoint - The article highlights the proactive approach of Wuhan's citizen service hotline "12345" in addressing urban issues before residents even voice their concerns, showcasing a shift from reactive to proactive governance through data analysis and technology [1][3][4]. Group 1: Proactive Governance - The "12345" hotline has implemented a system that analyzes citizen complaints to predict and resolve issues before they are reported, leading to over 20,000 preemptive actions in the first seven months of the year [1][3]. - The service has successfully reduced the number of complaints by 28% during the period when proactive alerts were issued, indicating the effectiveness of the "未诉先办" (preemptive action) strategy [3]. Group 2: Data-Driven Decision Making - The "灵醒" platform utilizes extensive historical and real-time data to monitor urban issues, allowing for a shift from reactive "firefighting" to proactive "fire prevention" in city management [4][5]. - The platform's new modules, such as "民生十二时辰" and "民情月历," help track and analyze the frequency and trends of citizen complaints, enhancing the ability to anticipate and address issues [4]. Group 3: Technological Integration - The integration of AI technology, specifically the "灵醒AI助手," has improved the efficiency of the hotline, allowing for rapid and accurate responses to citizen inquiries, with a 92.7% accuracy rate in dispatching information [6]. - The AI assistant has transformed the hotline's operations, enabling quicker responses and reducing the time taken to fill out service requests, thereby enhancing citizen satisfaction [6].
六年来首次!OpenAI新模型开放权重,Altman称为"全球最佳开放模型"
Hua Er Jie Jian Wen· 2025-08-05 20:05
Core Insights - OpenAI has released two open-weight language models, gpt-oss-120b and gpt-oss-20b, marking its first open-weight model launch since 2019 and responding to competition from Meta, Mistral AI, and DeepSeek [1][2][12] Model Specifications - gpt-oss-120b and gpt-oss-20b are designed for low-cost options, with gpt-oss-20b able to run on a laptop with 16GB RAM and gpt-oss-120b requiring approximately 80GB RAM [2][5] - gpt-oss-120b has a total of 117 billion parameters, activating 5.1 billion parameters per token, while gpt-oss-20b has 21 billion parameters, activating 3.6 billion parameters per token [5][6] Performance Evaluation - gpt-oss-120b performs comparably to OpenAI's o4-mini in core inference benchmarks, while gpt-oss-20b matches or exceeds the performance of o3-mini [7][8] - Both models utilize advanced pre-training and post-training techniques, focusing on efficiency and practical deployment across environments [5][11] Security Measures - OpenAI has implemented extensive security measures to prevent malicious use of the models, filtering harmful data during pre-training and conducting specialized fine-tuning for security assessments [11] - The company collaborates with independent expert groups to evaluate potential security risks associated with the models [11] Market Impact - The release of these models is seen as a strategic shift for OpenAI, which had previously focused on proprietary API services, now responding to competitive pressures in the open-weight model space [12][15] - OpenAI has partnered with major cloud service providers like Amazon to offer these models, enhancing accessibility for developers and researchers [3][11]
OpenAI releases two new open-weight AI models
CNBC Television· 2025-08-05 19:15
Open AAI just announcing two new openw weight AI models. Those are models where some of the parameters around how they're trained are public and accessible. McKenzie Sagalos is here.She's got more on how this could change the AI race in today's tech check. And Mac, we're talking about not completely open source, but a little bit. And that's a really important distinction here, Becky.So, OpenAI is shifting strategy today, making its tech more accessible than it's been in six years. Because until now you coul ...
X @Bloomberg
Bloomberg· 2025-08-05 17:02
OpenAI is releasing a pair of open and freely available AI models months after China's DeepSeek found success with a similar approach https://t.co/6AzFhFInXK ...
OpenAI releases lower-cost models to rival Meta, Mistral and DeepSeek
CNBC· 2025-08-05 17:00
Core Insights - OpenAI has released two open-weight language models, gpt-oss-120b and gpt-oss-20b, marking the first release since GPT-2 in 2019, aimed at providing lower-cost options for developers and researchers [1] - Open-weight models have publicly available parameters, offering transparency and control, but differ from open-source models which provide full source code [2] - OpenAI collaborated with major tech companies like Nvidia and AMD to ensure compatibility of the models across various chips [3] Industry Context - The release of open-weight models by OpenAI contributes to a growing ecosystem, with other companies like Meta and Mistral AI also launching similar models [2][3] - Nvidia's CEO highlighted OpenAI's role in advancing innovation in open-source software, indicating a significant impact on the AI landscape [4]
大模型大逃杀:一山不容「六小虎」|深氪
36氪· 2025-08-05 10:38
Core Viewpoint - The article discusses the challenges and transformations faced by the "Six Little Tigers" in the AI industry, highlighting their struggles with competition, internal turmoil, and the impact of external pressures from investors and market dynamics [6][9][60]. Group 1: Industry Overview - The "Six Little Tigers" were initially valued at over 20 billion RMB, but have faced significant setbacks in the rapidly evolving AI landscape, leading to a loss of confidence among employees and high executive turnover [6][13][60]. - The emergence of DeepSeek as a dominant player has shifted the competitive landscape, forcing the Six Little Tigers to reevaluate their strategies and operations [41][60]. Group 2: Internal Challenges - The article details the internal restructuring and layoffs within the Six Little Tigers, with many employees leaving due to a loss of faith in the companies' futures [10][12][13]. - A significant percentage of employees (41.07%) have reported being in a job-seeking status as of July 2025, indicating widespread dissatisfaction and uncertainty [13]. Group 3: Strategic Missteps - The pursuit of a "Super App" strategy led to ineffective competition and internal chaos, as companies focused on rapid growth metrics rather than sustainable product development [17][24][40]. - The aggressive marketing and product strategies, driven by FOMO (fear of missing out), resulted in a misalignment between product capabilities and market needs, ultimately harming the companies' long-term viability [18][21][40]. Group 4: Market Dynamics - The competitive pressure from DeepSeek has forced the Six Little Tigers to adopt open-source strategies, which they previously avoided, in an attempt to regain market relevance [48][49]. - The article emphasizes that the market is increasingly favoring a few top players, suggesting that only the strongest models will survive in the long run [60][64]. Group 5: Future Outlook - Despite the current turmoil, there remains potential for recovery and innovation within the Six Little Tigers, as they still possess significant resources and talent to pivot towards more sustainable business models [70][75]. - The article concludes that the journey towards achieving AGI (Artificial General Intelligence) is ongoing, with the possibility of resurgence for the companies if they can adapt and learn from past mistakes [76][75].