Open Models

Search documents
X @Demis Hassabis
Demis Hassabis· 2025-08-15 23:45
New hyper-efficient addition to our amazing Gemma open models: Gemma 3 270M packs a real punch for its tiny size! It’s super compact and power efficient, so you can easily run your own task-specific fine-tuned systems on edge devices. Enjoy building with it!Google AI Developers (@googleaidevs):Introducing Gemma 3 270M! 🚀 It sets a new standard for instruction-following in compact models, while being extremely efficient for specialized tasks. https://t.co/kC9OOPwzVi ...
The Rise of Open Models in the Enterprise — Amir Haghighat, Baseten
AI Engineer· 2025-07-24 15:30
AI Adoption in Enterprises - Enterprises' adoption of AI is crucial for realizing AI's full potential and impact [2] - Enterprises initially experiment with OpenAI and Anthropic models, often deploying them on Azure or AWS for security and privacy [7] - In 2023, enterprises were "toying around" with AI, but by 2024, 40-50% had production use cases built on closed models [9][10] Challenges with Closed Models - Vendor lock-in is not a primary concern for enterprises due to the increasing number of interoperable models [12][13] - Ballooning costs, especially with agentic use cases involving potentially 50 inference calls per user action, are becoming a significant concern [20] - Enterprises are seeking differentiation at the AI level, not just at the workflow or application level, leading them to consider in-house solutions [21] Reasons for Open Source Model Adoption - Frontier models may not be the right tool for specific use cases, such as medical document extraction, where enterprises can leverage their labeled data to build better models [16][17] - Generic API-based models may not suffice for tasks requiring low latency, such as AI voices or AI phone calls [18] - Enterprises aim to reduce costs and improve unit economics by running models themselves and controlling pricing [20][21] Inference Infrastructure Challenges - Optimizing models for latency requires both model-level and infrastructure-level optimizations, such as speculative decoding techniques like Eagle 3 [23][24][25][26] - Guaranteeing high availability (four nines) for mission-critical inference requires robust infrastructure to handle hardware failures and VLM crashes [27][28] - Scaling up quickly to handle traffic bursts is challenging, with some enterprises experiencing delays of up to eight minutes to bring up a new replica of a model [29]
The State of Agentic AI
DDN· 2025-05-15 19:50
AI has been around a line for a long time the modern AI movement started with Alex net in 2012 be he Jensen say it every GTC I'm sure he'll say it tomorrow he always does but in the end the the hockey stick for AI happened with chat GPT two and a half years ago that's the hockey stick that just took this to the roof but since then it's evolved so quickly every year it's changing and moving first how do you take over am's open models you know the Advent of llama the ad of mist all these open models that are ...