Workflow
开放权重AI推理模型
icon
Search documents
六年来首次!OpenAI发布两款开放权重AI推理模型!奥尔特曼称其为“全球最佳开放模型”
Mei Ri Jing Ji Xin Wen· 2025-08-05 22:57
Core Insights - OpenAI has made a significant move towards open-source models by releasing the GPT-OSS, marking the first time in six years that the company has introduced open-weight models [1][5] Model Details - OpenAI released two open-weight AI inference models on August 5: the gpt-oss-120b with 117 billion parameters, which can be run on a single NVIDIA professional data center GPU, and the gpt-oss-20b with 21 billion parameters, which can operate on consumer-grade laptops with 16GB of memory [3][6] - Both models are released under a permissive Apache 2.0 license, allowing businesses to use them commercially without prior payment or licensing [5] Performance Evaluation - The gpt-oss-120b model performs comparably to OpenAI's o4-mini in core inference benchmarks, while the gpt-oss-20b model matches or exceeds the performance of o3-mini [7] - The gpt-oss-120b model activates 510 million parameters per token, while the gpt-oss-20b activates 3.6 billion parameters, supporting context lengths of up to 128k [6][7] Market Context - OpenAI's release of open-weight models is largely driven by competitive pressure in the market, with the company emphasizing the importance of safety and security in the deployment of these models [12] - Amazon has announced it will offer OpenAI's models on its Bedrock and SageMaker platforms, marking the first time Amazon provides OpenAI products [6] Technical Architecture - Both models utilize advanced pre-training and post-training techniques, focusing on inference efficiency and practicality across deployment environments, employing a mixture of experts (MoE) architecture [6][7] Limitations - The smaller models are noted to produce more "hallucinations" due to their limited world knowledge compared to larger models, with gpt-oss-120b and gpt-oss-20b generating hallucinations in 49% and 53% of questions, respectively [11]