Workflow
The Industry Reacts to gpt-oss!
Matthew Bermanยท2025-08-06 19:22

Model Release & Performance - OpenAI released a new open-source model (GPT-OSS) that performs comparably to smaller models like 04 mini and can run on consumer hardware such as laptops and phones [1] - The 20 billion parameter version of GPT-OSS is reported to outperform models two to three times its size in certain tests [7] - Industry experts highlight the model's efficient training, with the 20 billion parameter version costing less than $500,000 to pre-train, requiring 21 million H100 hours [27] Safety & Evaluation - OpenAI conducted safety evaluations on GPT-OSS, including fine-tuning to identify potential malicious uses, and shared the recommendations they adopted or didn't adopt [2][3] - Former OpenAI safety researchers acknowledge the rigor of OpenAI's OSS safety evaluation [2][19] - The model's inclination to "snitch" on corporate wrongdoing was tested, with the 20 billion parameter version showing a 0% snitch rate and the 120 billion parameter version around 20% [31] Industry Reactions & Implications - Industry experts suggest OpenAI's release of GPT-OSS could be a strategic move to commoditize the model market, potentially forcing competitors to lower prices [22][23] - Some believe the value in AI will increasingly accrue to the application layer rather than the model layer, as the price of AI tokens converges with the cost of infrastructure [25][26] - The open-source model has quickly become the number one trending model on Hugging Face, indicating significant community interest and adoption [17][18] Accessibility & Use - Together AI supports the new open-source models from OpenAI, offering fast speeds and low prices, such as 15 cents per million input tokens and 60 cents per million output tokens for the 120 billion parameter model [12] - The 120 billion parameter model requires approximately 65 GB of storage, making it possible to store on a USB stick and run locally on consumer laptops [15] - Projects like GPTOSS Pro mode chain together multiple instances of the new OpenAI GPT-OSS model to produce better answers than a single instance [10]