开源模型

Search documents
为了不被挤下牌桌,OpenAI又开源了
Sou Hu Cai Jing· 2025-08-07 04:59
Core Insights - OpenAI has shifted its strategy by re-entering the open-source domain with the release of two models, gpt-oss-120b and gpt-oss-20b, marking a significant change from its previous closed-source approach [2][5][17] - The open-source models are designed to cater to different use cases, with gpt-oss-120b focusing on high inference needs and gpt-oss-20b aimed at localized applications [8][15] - OpenAI's decision to open-source these models is seen as a response to increasing competition in the AI space, particularly from companies like Anthropic and Google, which are gaining market share in the enterprise sector [3][22] OpenAI's Market Position - As of August, ChatGPT boasts 700 million weekly active users, a fourfold increase year-on-year, with daily message volume exceeding 3 billion [3] - OpenAI's paid user base has grown from 3 million to 5 million, with Pro and enterprise users contributing over 60% of revenue [3] - Despite its consumer market dominance, OpenAI faces challenges in the enterprise market, where competitors are encroaching on its share [3][22] Open-Source Strategy - OpenAI's initial open-source philosophy has evolved, with a notable shift to a closed-source model in 2020, which drew criticism for deviating from its mission to benefit humanity [5][16] - The newly released models follow a permissive Apache 2.0 license, allowing for extensive commercial and research use, which contrasts with the previous API-dependent model [14][15] - The open-source models are expected to enhance OpenAI's market influence, as they can now be deployed on major cloud platforms like Amazon AWS, allowing for broader accessibility [17][19] Competitive Landscape - The rise of open-source models has led to a more competitive environment, with companies like DeepSeek and Alibaba's Qwen series gaining traction in the market [18][22] - OpenAI's re-entry into open-source is anticipated to reshape the competitive dynamics, as more companies adopt a hybrid approach of open and closed models [17][22] - The trend indicates that open-source models are becoming increasingly viable, with the performance gap between open-source and closed-source models narrowing [17][18] Financial Implications - OpenAI is projected to achieve an annual recurring revenue (ARR) of $12 billion by the end of July, significantly outpacing its closest competitor, Anthropic, which is expected to reach $5 billion [19][22] - The financial model of open-source remains challenging, as companies may hesitate to adopt open-source strategies due to the lack of direct revenue generation from model usage [19][22]
全网开测GPT-oss!技术架构也扒明白了
量子位· 2025-08-07 00:56
Core Insights - The article highlights the impressive performance of GPT-oss, which surpasses many existing open-source models and is poised to lead in the SaaS fast-fashion era [1][3][4]. Performance Testing - GPT-oss has successfully passed multiple performance tests, achieving top rankings in various benchmarks, including GPQA Diamond, AIME 2024, AIME 2025, and Codeforces, outperforming models like DeepSeek R1, Qwen3, and Llama 4 [5][6]. - In the MMLU benchmark, GPT-oss achieved scores of 85.9 for the low 120B model and 88 for the medium model, while Qwen3-235B performed slightly better in MMLU [6][7]. Model Architecture - The architecture of GPT-oss is noted for its wider structure, more attention heads, and higher hidden dimensions compared to similar models, incorporating advanced techniques such as attention bias units [22][24][26]. - The model retains the core MoE Transformer architecture while optimizing performance and reducing complexity, making it suitable for open-source applications [26][28]. Cost and Training - The estimated cost for training the GPT-oss-120B model is between $4.2 million and $23.1 million, while the 20B model costs between $420,000 and $2.3 million [30]. - There are indications that the model may have limitations in non-English text performance, with a significant portion of responses containing grammatical or spelling errors [30]. User Applications - Users have begun exploring various applications for GPT-oss, including its integration into platforms for academic paper understanding and data transformation [17][19][20]. - The model can be easily accessed and utilized through platforms like LM Studio and AWS, facilitating rapid development of AI applications [33][34]. Community Engagement - The article encourages users to test GPT-oss and share their experiences, indicating a growing community interest in the model's capabilities [39].
中国“霸榜”全球开源大模型:光环下的隐忧与挑战丨人工智能AI瞭望台
证券时报· 2025-08-07 00:12
Core Viewpoint - China's open-source large models are rising in a "cluster-style" manner, reshaping the global AI landscape, while also presenting challenges such as frequent iterations leading to compatibility issues and a tendency towards homogenization [2][5][10]. Group 1: Open-source Model Surge - In recent weeks, major Chinese companies have released multiple open-source models, marking a resurgence in the domestic large model scene, reminiscent of the "hundred model battle" of 2023 [2][4]. - As of July 31, 2023, nine out of the top ten open-source large models listed by Hugging Face are from China, with notable models like Zhipu's GLM-4.5 and Alibaba's Tongyi Qianwen series dominating the rankings [4][5]. Group 2: Shift from Closed to Open-source - The success of DeepSeek has been pivotal in shifting the industry towards open-source models, prompting more companies to follow suit and focus on model optimization and iteration [4][5]. - The open-source approach is seen as a way for latecomers in the AI field, particularly in China, to break the dominance of established closed-source models [7][8]. Group 3: Economic and Technical Implications - The rise of open-source models in China is driven by the availability of vast amounts of quality Chinese language data and the maturation of domestic computing power, creating a strong feedback loop [5][8]. - Open-source models lower the barriers to entry for smaller companies, enabling them to leverage advanced models at reduced costs, thus accelerating AI integration into various sectors [8][10]. Group 4: Challenges and Concerns - The rapid iteration of open-source models has led to a phenomenon described as "tuning internal competition," where the lack of disruptive innovation results in similar capabilities across models [10][11]. - Developers face challenges such as high compatibility costs and frequent changes in model interfaces, which complicate integration efforts [10][11]. - Experts suggest that to avoid stagnation, there is a need for unified API standards and a focus on foundational algorithm innovation [11].
DeepSeek终于把OpenAI逼急了
阿尔法工场研究院· 2025-08-07 00:08
Core Viewpoint - The release of OpenAI's first open-source language model, GPT-OSS, marks a significant shift in the AI landscape, challenging the previously held belief that the strongest models would remain closed-source and proprietary [5][12][13]. Group 1: OpenAI's Strategic Shift - OpenAI has transitioned from a closed-source, paid model to an open-source collaborative ecosystem, potentially signaling a competitive stance against domestic Chinese models [14][16]. - The newly released models, GPT-OSS-120B and GPT-OSS-20B, are designed to be efficient and accessible, with the former featuring 117 billion parameters and the latter 21 billion parameters, both capable of running on consumer-grade hardware [9][10][11]. Group 2: Impact of Chinese Open-Source Models - The rapid development of Chinese open-source models, such as DeepSeek and Tongyi Qwen, has prompted OpenAI to reconsider its strategy, as these models have gained significant traction and market presence [18][20]. - The Chinese open-source model ecosystem is expected to flourish by 2025, with multiple influential teams emerging in various AI domains, including programming and multi-modal applications [20][22]. Group 3: Competitive Landscape - The competitive dynamics in the AI sector are intensifying, with Meta also reconsidering its open-source strategy, indicating a broader trend among major players to reassess their approaches in light of emerging competition [22].
资金动向 | 北水买日港股超90亿港元,加仓腾讯、阿里
Ge Long Hui· 2025-08-06 19:07
Group 1 - Tencent Holdings saw a net buy of HKD 15.18 billion, marking a total of HKD 59.44 billion in net buys over the last 10 days [1] - Alibaba-W experienced a net buy of HKD 8.76 billion, with a total of HKD 21.16 billion in net buys over the last 3 days [1] - SMIC recorded a net buy of HKD 6.11 billion, totaling HKD 9.51 billion in net buys over the last 3 days [1] Group 2 - Alibaba-W launched a new membership system integrating various Alibaba resources, aimed at enhancing consumer experience and user engagement [5] - SMIC is set to release its earnings report on August 7, with expected revenue of USD 2.185 billion for Q2 2025, a year-on-year increase of 14.91% [6] - Crystal Technology announced a significant order with DoveTree, totaling approximately USD 6 billion, setting a record for AI pharmaceutical overseas orders [6] Group 3 - Ideal Auto-W, along with China Automotive Research and Dongfeng Liuzhou Motor, issued a joint statement advocating for self-discipline and ethical competition in the automotive industry [6] - Bubble Mart received a report from Morgan Stanley stating that its intrinsic value exceeds its current IP holdings, maintaining an "overweight" rating with a target price of HKD 365 [6]
时隔六年,OpenAI 为什么再次开源?
Founder Park· 2025-08-06 14:00
Core Viewpoint - OpenAI's release of the open-source model gpt-oss marks a significant strategic shift, indicating a clearer understanding of its value proposition beyond just the model itself, focusing on its user base and application ecosystem [2][4][13]. Group 1: OpenAI's Open-Source Model Release - OpenAI has launched its first open-source model, gpt-oss, since GPT-2, with performance comparable to its proprietary o4 mini model while reducing costs by at least 10 times [2][10]. - The gpt-oss-120b model achieved a score of 90.0 on the MMLU benchmark, while the gpt-oss-20b scored 85.3, indicating competitive performance in the open-source landscape [3][8]. - The models are designed to run efficiently on various hardware, from consumer-grade GPUs to cloud servers, and are licensed under Apache 2.0, allowing for commercial deployment without downstream usage restrictions [7][8]. Group 2: Strategic Implications - OpenAI's move to open-source is not merely a technical sharing but aims to build an application ecosystem, targeting enterprises looking to deploy open-source AI models [5][12]. - The release reflects OpenAI's recognition that its core competitive advantage lies in its large user base and application ecosystem rather than just the models themselves [4][13]. - OpenAI's decision to avoid releasing training data, code, or technical reports suggests a strategy to attract businesses while potentially impacting academic research and the true open-source AI community [19][22]. Group 3: Competitive Landscape - The introduction of gpt-oss is expected to challenge existing API products, with OpenAI positioning itself aggressively in the market by offering a model that significantly undercuts the cost of its proprietary offerings [10][11]. - The architecture of gpt-oss aligns with industry trends towards sparse MoE models, indicating a shift in design preferences within the AI community [14]. - The competitive landscape is evolving, with OpenAI's release potentially reversing the previous lag in open-source model applications compared to Chinese counterparts [21][22]. Group 4: Future Considerations - The open-source model's ecosystem remains chaotic, with high-scoring models not necessarily being user-friendly, which could slow adoption rates [17][18]. - OpenAI's approach to model safety and fine-tuning raises questions about the balance between usability and security, which will need community validation [15][16]. - The ongoing competition between U.S. and Chinese open-source models highlights the need for strategic actions to maintain relevance and leadership in the AI space [20][22].
OpenAI、谷歌等深夜更新多款模型 展示开源、智能体、世界模型进展
Di Yi Cai Jing· 2025-08-06 04:59
Core Insights - Major AI companies released new products, showcasing shifts in product strategies, particularly OpenAI's transition to open-source models and Anthropic's focus on incremental updates [1][3] OpenAI - OpenAI launched two open-source models: gpt-oss-120b with 117 billion parameters and gpt-oss-20b with 21 billion parameters, both utilizing MoE architecture [2] - The gpt-oss-120b model can run on an 80GB GPU, while gpt-oss-20b can operate on consumer devices with 16GB memory, allowing local deployment on laptops and mobile phones [2] - These models achieved top-tier performance in benchmark tests, with gpt-oss-120b scoring close to or exceeding the closed-source o4-mini model [2] Anthropic - Anthropic introduced Claude Opus 4.1, marking a shift towards more frequent, incremental updates rather than focusing solely on major version releases [3] - The new model demonstrated improved capabilities in complex multi-step problem-solving and coding tasks, with a SWE-bench Verify score of 74.5%, surpassing the previous version [4] Google - Google launched Genie 3, its first world model allowing real-time interaction, building on previous models Genie 1 and Genie 2 [5] - Genie 3 can simulate diverse environments and natural phenomena, maintaining visual consistency for up to several minutes at 720p resolution [6] - Despite advancements, Genie 3 has limitations, such as restricted action space and challenges in simulating multiple agents in shared environments [9]
OpenAI、谷歌等深夜更新多款模型,展示开源、智能体、世界模型进展
Di Yi Cai Jing· 2025-08-06 04:49
Core Insights - The recent product launches by OpenAI, Anthropic, and Google indicate a shift in product strategies among major AI model developers, with a focus on open-source models and incremental updates [1][3][5] OpenAI - OpenAI has released two open-source models, gpt-oss-120b with 117 billion parameters and gpt-oss-20b with 21 billion parameters, both utilizing the MoE architecture [2] - The gpt-oss-120b model can run on a single 80GB GPU, while gpt-oss-20b can operate on consumer devices with 16GB memory, allowing for local deployment on laptops and mobile devices [2] - OpenAI's CEO, Sam Altman, emphasized the importance of releasing powerful open-source models, which are the result of billions of dollars in research [1][2] Anthropic - Anthropic has shifted its strategy to focus on more frequent incremental updates rather than solely major version releases, exemplified by the launch of Claude Opus 4.1 [3] - Claude Opus 4.1 shows improvements in coding capabilities, scoring 74.5% on the SWE-bench Verify benchmark, surpassing its predecessor [4] - The new model is designed to handle complex multi-step problems more effectively, positioning it as a more capable AI agent [3][4] Google - Google introduced Genie 3, its first world model that supports real-time interaction, building on previous models like Genie 1 and Genie 2 [5] - Genie 3 can simulate diverse interactive environments and model physical properties, allowing for realistic navigation and interaction within generated worlds [5][6] - Despite its advancements, Google acknowledges limitations in Genie 3, such as restricted action spaces and challenges in simulating multiple agents in shared environments [9]
谁在拆 OpenAI 的围墙?
3 6 Ke· 2025-08-06 01:41
你可能也刷到了。 昨晚,OpenAI 突然搞了个大动作:宣布开源两款新模型,叫gpt-oss-120b和gpt-oss-20b。 可以说,这是GPT-2以来,OpenAI重新向开源社区开放模型权重,关于模型参数、推理性能、训练细 节,网上已经铺天盖地了,我就不啰嗦了。 但我想说:你有没有想过,这次开源到底意味着什么? 01 智远认为这是一次战略转折点。 要知道,过去几年,OpenAI 一直是"闭源派"的头号代表。它靠 GPT-3、GPT-4 的技术优势,用 API 收 费、订阅制赚钱,建起了高墙,几乎垄断了大模型时代的入口和定价权,说白了,它就是定规则的人。 后来,风向变了。 DeepSeek火了后,局面开始松动。一批开源模型不仅性能逼近 GPT-4,成本还只有人家的 1/20。更关键 的是,它们用极度宽松的开源协议,允许你随便用、随便改、还能商用,几乎零门槛。 面对这种冲击,连 Sam Altman 都在今年 2 月 1 日公开承认了一句扎心的话:我们可能站在了历史错误 的一边。 所以,半年后,OpenAI 终于行动了。但这波"开源",真不是低头认输,仔细一看,你会发现,这里面 门道不少。 智远认为,它在主 ...
OpenAI发布ChatGPT世代首个开源模型gpt-oss,4060Ti都能跑得动。
数字生命卡兹克· 2025-08-05 22:08
8月6号,真的今夕是何年了。 一晚上,三个我觉得都蛮大的货。 先是晚上10点,Google发了一个世界模型(但期货),Genie 3。 这个非常的强,我看的热血沸腾,我这两天也会单独写一篇文章,来聊聊这个玩意,真的,作为一个这么多年的游戏和VR玩家,看到Genie 3非常的激 动。 然后就是12点半,Anthropic突然就发布了Claude Opus 4.1,在编程能力上继续进化。 这节奏,感觉就是来狙击OpenAI的。 然后,重头戏来了。 凌晨1点,OpenAI在GPT-2后,在整个ChatGPT世代,官宣发布了他们的第一个开源模型,GPT-oss。 真的,不眠之夜。 直接来聊聊GPT-oss。 很强,非常强,OpenAI终于干了点人事。 也就是说,20B模型的大小就12.8GB ,最低只要16GB内存就能跑,我这个破壁5080的16G卡,也能本地跑的动了20B的gpt-oss了。 GPT-oss一共开源了两款模型,120B和20B,都是MoE,纯文本、非多模态的推理模型, Apache 2.0 许可,也就是最宽松的那种,你随便用 。 | Model | Layers | Total Params | A ...