Open Source Models

Search documents
ARK AI Agents Research | 2025 Mid-Year Review
ARK Invest· 2025-08-14 15:30
AI Agent Transition & Productivity - The industry is transitioning from AI assistants to AI agents capable of performing longer-form tasks using multiple tools and personal/business context [1][2] - This transition is expected to drive significant productivity gains as AI agents handle more complex and valuable tasks [2] - Improvements in AI technology, cost declines, and product development are fueling the advancement of AI agents in both consumer and enterprise applications [3] Market Adoption & Consumer Trends - OpenAI launched an agent product integrated into ChatGPT, which has over 700 million weekly active users [4] - Meta reported that sales of Meta Ray-Ban glasses tripled year-over-year from the first half of 2024 to the first half of 2025, indicating growing consumer adoption [7] - Personal AI agents are expected to become the first point of contact for accessing products and services online, potentially disrupting traditional search and marketplaces [10] Enterprise Applications & Software Development - Customer service and software development are currently the highest-value use cases for AI in the enterprise [12] - AI-native development environments (IDEs) are experiencing rapid growth, with companies like Cursor and Replit seeing revenue increase by more than 10x from Q4 last year to halfway through 2025 [14] - Cursor's ARR grew from $50 million to over $500 million, with rumors suggesting it's approaching $1 billion [14] - Businesses are reallocating hiring plans towards revenue-driving roles, adjusting for the impact of AI on software development and customer support [13] Monetization & Investment - While net new ARR growth for public enterprise software companies has decelerated, AI companies in the private market are experiencing rapid growth [18] - There is a willingness to pay for high-priced monthly subscriptions (over $200) for access to advanced AI models like ChatGPT, Claude, and Grok [19] - Business spending on software is expected to accelerate throughout the decade, reaching investment levels not seen since the COVID-19 pandemic [17] Open Source Models & Geopolitical Competition - China has emerged as a leader in open-source AI models, surpassing US companies in model performance [20][21] - OpenAI released its first open-source model since GPT-2 in response to the growing competition from Chinese open-source models [22]
Top LLM Providers for Enterprises
Bloomberg Technology· 2025-08-04 20:17
In the enterprise domain in particular, you guys are saying that anthropic leads in terms of market share, then open air is just behind. Just for transparency. While our audience of course, Menlow Low is on the cap table of anthropic and you have a sort of partnership with them for compute access but break down the data and how in Prop X pulled ahead in that space.Yeah, I mean, there's really three things that stood out in the report. First is, you know, Anthropic really pulled ahead in terms of enterprise ...
The Rise of Open Models in the Enterprise — Amir Haghighat, Baseten
AI Engineer· 2025-07-24 15:30
AI Adoption in Enterprises - Enterprises' adoption of AI is crucial for realizing AI's full potential and impact [2] - Enterprises initially experiment with OpenAI and Anthropic models, often deploying them on Azure or AWS for security and privacy [7] - In 2023, enterprises were "toying around" with AI, but by 2024, 40-50% had production use cases built on closed models [9][10] Challenges with Closed Models - Vendor lock-in is not a primary concern for enterprises due to the increasing number of interoperable models [12][13] - Ballooning costs, especially with agentic use cases involving potentially 50 inference calls per user action, are becoming a significant concern [20] - Enterprises are seeking differentiation at the AI level, not just at the workflow or application level, leading them to consider in-house solutions [21] Reasons for Open Source Model Adoption - Frontier models may not be the right tool for specific use cases, such as medical document extraction, where enterprises can leverage their labeled data to build better models [16][17] - Generic API-based models may not suffice for tasks requiring low latency, such as AI voices or AI phone calls [18] - Enterprises aim to reduce costs and improve unit economics by running models themselves and controlling pricing [20][21] Inference Infrastructure Challenges - Optimizing models for latency requires both model-level and infrastructure-level optimizations, such as speculative decoding techniques like Eagle 3 [23][24][25][26] - Guaranteeing high availability (four nines) for mission-critical inference requires robust infrastructure to handle hardware failures and VLM crashes [27][28] - Scaling up quickly to handle traffic bursts is challenging, with some enterprises experiencing delays of up to eight minutes to bring up a new replica of a model [29]