Workflow
Tool Use
icon
Search documents
X @Avi Chawla
Avi Chawla· 2025-08-29 06:30
AI Agent Evolution - The industry has progressed from simple LLMs to sophisticated Agentic systems with reasoning, memory, and tool use [1] - Early transformer-based chatbots were limited by small context windows, exemplified by ChatGPT's initial 4k token limit [1] - The industry has seen upgrades to handle thousands of tokens, enabling parsing of larger documents and longer conversations [1] - Retrieval-Augmented Generation (RAG) provided access to fresh and external data, enhancing LLM outputs [1] - Multimodal LLMs can process multiple data types (text, images, audio), with memory introducing persistence across interactions [1] Key Components of Advanced AI Agents - Advanced AI Agents are equipped with short-term, long-term, and episodic memory [1] - Tool calling (search, APIs, actions) is a crucial feature of modern AI Agents [1] - Reasoning and ReAct-based decision-making are integral to the current AI Agent era [1]
OpenAI Dropped a FRONTIER Open-Weights Model
Matthew Berman· 2025-08-05 17:17
Model Release & Capabilities - Open AAI released GPTOSS, state-of-the-art open-weight language models in 120 billion and 20 billion parameter versions [1] - The models outperform similarly sized open-source models on reasoning tasks and demonstrate strong tool use capabilities [3] - The models are optimized for efficient deployment on consumer hardware, with the 120 billion parameter version running efficiently on a single 80 GB GPU and the 20 billion parameter version on edge devices with 16 GB of memory [4][5] - The models excel in tool use, few-shot learning, function calling, chain of thought reasoning, and health issue diagnosis [8] - The models support context lengths of up to 128,000 tokens [12] Training & Architecture - The models were trained using a mix of reinforcement learning and techniques informed by OpenAI's most advanced internal models [3] - The models utilize a transformer architecture with a mixture of experts, reducing the number of active parameters needed to process input [10][11] - The 120 billion parameter version activates only 5 billion parameters per token, while the 20 billion parameter version activates 36 billion parameters [11][12] - The models employ alternating dense and locally banded sparse attention patterns, group multi-query attention, and RoPE for positional encoding [12] Safety & Security - OpenAI did not put any direct supervision on the chain of thought for either OSS model [21] - The models were pre-trained and filtered to remove harmful data related to chemical, biological, radiological, and nuclear data [22] - Even with robust fine-tuning, maliciously fine-tuned models were unable to reach high capability levels according to OpenAI's preparedness framework [23] - OpenAI is hosting a challenge for red teamers with $500,000 in awards to identify safety issues with the models [24]
2025上半年,AI Agent领域有什么变化和机会?
Hu Xiu· 2025-07-11 00:11
Core Insights - The rapid development of AI Agents has ignited a trend of "everything can be an Agent," particularly evident in the competitive landscape of model development and application [1][2][10] - Major companies like OpenAI, Google, and Alibaba are heavily investing in the Agent space, with new products emerging that enhance user interaction and decision-making capabilities [2][7][8] - The evolution of AI applications is categorized into three phases: prompt-based interactions, workflow-based systems, and the current phase of AI Agents, which emphasize autonomous decision-making and tool usage [17][19] Group 1: Model Development - The AI sector has entered a "arms race" for model development, with significant advancements marked by the release of models like DeepSeek, o3 Pro, and Gemini 2.5 Pro [5][6][14] - The introduction of DeepSeek has demonstrated that there is no significant gap between domestic and international model technologies, prompting major players to accelerate their model strategies [6][10] - The focus has shifted from "pre-training" to "post-training" methods, utilizing reinforcement learning to enhance model performance even with limited labeled data [11][13] Group 2: Application Development - The launch of OpenAI's Operator and Deep Research has marked 2025 as the "Year of AI Agents," with a surge in applications that leverage these capabilities [7][8] - Companies are exploring various applications of AI Agents, with notable examples including Cursor and Windsurf, which have validated product-market fit in the programming domain [9][21] - The ability of Agents to use tools effectively has been a significant breakthrough, allowing for enhanced information retrieval and interaction with external systems [20][21] Group 3: Challenges and Opportunities - Despite advancements, AI Agents face challenges such as context management, memory mechanisms, and interaction with complex software systems [39][40] - The future of Agent applications may involve evolving business models, potentially shifting from subscription-based to usage-based or outcome-based payment structures [40][41] - The industry is witnessing a competitive landscape where vertical-specific Agents may offer more value due to their specialized knowledge and closer user relationships [42][46]
Deep Research类产品深度测评:下一个大模型产品跃迁点到来了吗?
Founder Park· 2025-04-23 12:37
以下文章来源于海外独角兽 ,作者拾象 Founder Park 正在搭建开发者社群,邀请积极尝试、测试新模型、新技术的开发者、创业者们加入,请扫码详细填写你的产品/项目信息,通过 审核后工作人员会拉你入群~ 海外独角兽 . 研究科技大航海时代的伟大公司。 Deep Research 产品可被理解为 一个以大模型能力为基础、集合了检索与报告生成的端到端系统,对信息进行迭代搜索和分析,并生成详细报告作为输 出。 参考 Han Lee 的 2x2 分析框架,目前 Deep Research 类产品在 输出深度、训练程度 两大维度呈现分异。 输出深度 即产品在先前研究成果的基础上进行了 多少次迭代循环以收集更多信息,可进一步被理解为 Agentic 能力的必要基础。 低训练程度 指代经过人工干预和调整的系统,比如使用人工调整的 prompt,高训练程度则是指利用机器学习对系统进行训练。 和传统 LLM Search 产品相比,Deep Research 是迈向 Agent 产品雏形的一次跃迁,可能也将成为具有阶段代表性的经典产品形态。 Deep Research 产品通过系列推理模型嵌入,已生长出了 Agent 产品 ...