开放权重模型
Search documents
饭都不香了,奥尔特曼焦虑:美国低估了中国…
Guan Cha Zhe Wang· 2025-08-19 04:07
Core Insights - OpenAI's new model GPT-5 faced significant backlash, leading to the reversion of ChatGPT's default model to a previous version, indicating a major setback for the company [1] - CEO Sam Altman expressed concerns about the U.S. underestimating the complexity and severity of China's AI development, emphasizing that the U.S.-China AI competition is multifaceted and cannot be simplified to a mere ranking of who is ahead [1][2] - OpenAI's decision to release open-weight models was influenced by competition from Chinese models, particularly open-source systems like DeepSeek, marking a strategic shift towards increasing technology accessibility [6][7] U.S.-China AI Competition - Altman warned that U.S. policies, such as semiconductor export controls, may not effectively address the rapid advancements in China's AI ecosystem, suggesting that relying solely on policy measures is unrealistic [2] - Despite the U.S. government's attempts to control AI development through export restrictions, China is building a comprehensive AI technology ecosystem that could circumvent these measures [2][3] - China's response to U.S. export controls reflects a commitment to developing a self-sufficient semiconductor supply chain, while still needing advanced AI processors [3][4] OpenAI's Strategic Shift - OpenAI's release of open-weight models, including gpt-oss-120b and gpt-oss-20b, represents a significant change in strategy, aimed at enhancing developer engagement and countering competition from Chinese open-source models [6][7] - The initial reception of these open-weight models has been mixed, with some developers noting a lack of core functionalities compared to OpenAI's commercial products [7] - Altman acknowledged that the open-weight models are optimized for specific applications, particularly for building local coding agents, with the potential for future adjustments based on market demands [7][8]
六年来首次!OpenAI新模型开放权重,Altman称为“全球最佳开放模型”
华尔街见闻· 2025-08-06 13:06
Core Viewpoint - OpenAI has made a significant move by releasing its first open-weight language models, gpt-oss-120b and gpt-oss-20b, in six years, responding to competition in the open-weight model space [1][2][16]. Model Details - The newly released models, gpt-oss-120b and gpt-oss-20b, are designed for low-cost options, allowing developers and researchers to run and customize them easily [2][5]. - gpt-oss-120b has a total parameter count of 117 billion, with 5.1 billion active parameters per token, while gpt-oss-20b has 21 billion total parameters and 3.6 billion active parameters per token [6][7]. - Both models support a context length of up to 128k tokens and utilize a mixture of experts (MoE) architecture to enhance efficiency [5][6]. Performance Evaluation - gpt-oss-120b's performance in core inference benchmarks is comparable to OpenAI's o4-mini, while gpt-oss-20b matches or exceeds the performance of o3-mini in similar evaluations [8][9]. - The models are optimized for local execution, with gpt-oss-20b requiring at least 16GB of memory and gpt-oss-120b needing around 80GB [2][8]. Strategic Shift - The release marks a strategic shift for OpenAI, which had previously delayed the launch of open-weight models due to extensive safety testing [3][16]. - OpenAI's decision to release these models is largely driven by competitive pressure, particularly from companies like DeepSeek [16]. Security Measures - OpenAI has implemented rigorous security measures to prevent malicious use of the models, including filtering harmful data during pre-training and conducting extensive internal and external testing [14][15]. - The company has collaborated with independent expert groups to assess potential security risks associated with the models [15]. Ecosystem and Market Impact - The models are available under the Apache 2.0 license and can be downloaded from platforms like Hugging Face and GitHub, with cloud service providers like Amazon and Microsoft also offering them [15][16]. - OpenAI is hosting a red team challenge with a $500,000 prize pool to encourage the identification of new security issues [15].
亚马逊云科技宣布接入OpenAI开放权重模型
Sou Hu Cai Jing· 2025-08-06 10:30
Group 1 - Amazon Web Services (AWS) announced that OpenAI's open-weight models can be accessed through Amazon Bedrock and Amazon SageMaker AI, enabling customers to quickly build generative AI applications [1][3] - OpenAI's latest open-weight foundational models, gpt-oss-120b and gpt-oss-20b, are designed as low-cost options that support local deployment, tool usage, and chain-of-thought processing, making them suitable for developers and researchers [3] - AWS aims to position itself as the best platform for running models, emphasizing the importance of open-weight models in the future development of generative AI [3] Group 2 - The collaboration with OpenAI is seen as a natural extension of AWS's commitment to bringing cutting-edge AI technologies to organizations worldwide [3] - AWS's extensive customer base is expected to transform the way OpenAI's advanced technologies are accessed [3]
OpenAI发布低成本模型 与Meta(META.US)和DeepSeek正面竞争
智通财经网· 2025-08-06 01:53
Core Insights - OpenAI has released its first open-weight language models, gpt-oss-120b and gpt-oss-20b, since the launch of GPT-2 in 2019, aimed at providing low-cost options for developers, researchers, and businesses [1][2] - The release follows multiple delays due to the need for additional safety testing and review of high-risk areas, with comprehensive safety training and testing implemented for the models [2] - The models are designed to run on various hardware environments, from consumer-grade devices to cloud services, showcasing advanced reasoning, tool invocation, and chain-of-thought processing capabilities [2] Company and Industry Developments - OpenAI collaborates with companies like NVIDIA, AMD, Cerebras, and Groq to ensure stable operation of the models across different chips [1] - The models can be downloaded via platforms like Hugging Face and GitHub under the Apache 2.0 license, with cloud services provided by Amazon, Baseten, and Microsoft [2] - OpenAI's president expressed excitement about the growth of the AI ecosystem and the company's role in pushing technological boundaries [1]
六年来首次!OpenAI新模型开放权重,Altman称为"全球最佳开放模型"
Hua Er Jie Jian Wen· 2025-08-05 20:05
Core Insights - OpenAI has released two open-weight language models, gpt-oss-120b and gpt-oss-20b, marking its first open-weight model launch since 2019 and responding to competition from Meta, Mistral AI, and DeepSeek [1][2][12] Model Specifications - gpt-oss-120b and gpt-oss-20b are designed for low-cost options, with gpt-oss-20b able to run on a laptop with 16GB RAM and gpt-oss-120b requiring approximately 80GB RAM [2][5] - gpt-oss-120b has a total of 117 billion parameters, activating 5.1 billion parameters per token, while gpt-oss-20b has 21 billion parameters, activating 3.6 billion parameters per token [5][6] Performance Evaluation - gpt-oss-120b performs comparably to OpenAI's o4-mini in core inference benchmarks, while gpt-oss-20b matches or exceeds the performance of o3-mini [7][8] - Both models utilize advanced pre-training and post-training techniques, focusing on efficiency and practical deployment across environments [5][11] Security Measures - OpenAI has implemented extensive security measures to prevent malicious use of the models, filtering harmful data during pre-training and conducting specialized fine-tuning for security assessments [11] - The company collaborates with independent expert groups to evaluate potential security risks associated with the models [11] Market Impact - The release of these models is seen as a strategic shift for OpenAI, which had previously focused on proprietary API services, now responding to competitive pressures in the open-weight model space [12][15] - OpenAI has partnered with major cloud service providers like Amazon to offer these models, enhancing accessibility for developers and researchers [3][11]
OpenAI重磅出手:六年来首发“开放权重”模型,微软独家协议迎挑战?
Hua Er Jie Jian Wen· 2025-07-10 05:58
Group 1 - OpenAI plans to release an "open weight" AI language model, marking its first such release since GPT-2 in 2019 and the first open model after signing an exclusive cloud service agreement with Microsoft [1] - The new model is expected to have reasoning capabilities similar to o3 mini and will be available on Azure, Hugging Face, and other major cloud platforms [1] - The open weight model allows enterprises and government agencies to run the model independently, differing from OpenAI's previous closed weight models [1][2] Group 2 - The open weight model represents a middle ground between open-source and closed-source methods, allowing users to view and modify weights without retraining the model [2] - This model could reduce usage costs for enterprises and enable customization, although its final "openness" will depend on the licensing terms and whether OpenAI provides full access to the model's code and training details [2] Group 3 - The timing of the model's release poses challenges for Microsoft's exclusive agreement with OpenAI, as it may weaken Microsoft's unique advantage in the AI business [3] - Competitors could host the model, potentially leading some Azure customers to switch to lower-cost alternatives or other cloud platforms [3] - Microsoft has invested over $13 billion in OpenAI, and the current revenue-sharing agreement allows Microsoft to receive 20% of the revenue from OpenAI's ChatGPT and API platforms [3][4] Group 4 - Ongoing negotiations between Microsoft and OpenAI involve key contract terms, including Microsoft's exclusive rights to sell OpenAI software through Azure and priority in providing computing infrastructure [4] - OpenAI must secure Microsoft's approval for its transition to a for-profit model by the end of the year to avoid losing significant investments from backers like SoftBank [4]
美媒:OpenAI开放语言模型即将问世
news flash· 2025-07-09 16:22
Core Insights - OpenAI is preparing to release an open-weight language model, marking a significant shift in its approach to AI model deployment [1] - The relationship between Microsoft and OpenAI is undergoing a dramatic change as they renegotiate contracts, allowing OpenAI to restructure as a for-profit entity [1] - The upcoming open-weight model will be available on OpenAI and Microsoft Azure servers, as well as other cloud service providers, enabling businesses and governments to deploy the model independently [1] Company Dynamics - The collaboration between Microsoft and OpenAI is becoming more complex, with potential implications for their future partnership [1] - OpenAI's decision to release an open-weight model contrasts with its previous closed-weight model strategy, indicating a shift towards greater accessibility [1] Industry Impact - The release of the open-weight model could intensify competition in the AI space, as it allows broader access to advanced AI capabilities [1] - This move may influence how other companies and governments approach AI deployment, potentially leading to increased innovation and development in the sector [1]
6月11日电,OpenAI首席执行官山姆·奥特曼(Sam Altman)表示,将在开放权重模型(open - weights model)上再多花些时间,也就是说,预计会在今年夏末推出,但不是6月。
news flash· 2025-06-10 23:26
Core Viewpoint - OpenAI's CEO Sam Altman announced that the company will spend more time on the open weights model, with an expected launch by late summer, rather than in June [1] Group 1 - OpenAI is focusing on the development of the open weights model [1] - The anticipated release timeline for the open weights model has been pushed to late summer [1] - The launch will not occur in June as previously speculated [1]
OpenAI首席执行官山姆·奥特曼:我们将在开放权重模型(open - weights model)上再多花些时间,也就是说,预计会在今年夏末推出,但不是6月。
news flash· 2025-06-10 23:14
Core Viewpoint - OpenAI's CEO Sam Altman announced that the company will spend more time on the open-weights model, with an expected launch by late summer, rather than in June [1] Group 1 - OpenAI is focusing on the development of an open-weights model [1] - The anticipated release timeline for the open-weights model has been pushed to late summer [1] - The initial expectation for a June launch has been revised [1]