Function Calling

Search documents
OpenAI Dropped a FRONTIER Open-Weights Model
Matthew Berman· 2025-08-05 17:17
Model Release & Capabilities - Open AAI released GPTOSS, state-of-the-art open-weight language models in 120 billion and 20 billion parameter versions [1] - The models outperform similarly sized open-source models on reasoning tasks and demonstrate strong tool use capabilities [3] - The models are optimized for efficient deployment on consumer hardware, with the 120 billion parameter version running efficiently on a single 80 GB GPU and the 20 billion parameter version on edge devices with 16 GB of memory [4][5] - The models excel in tool use, few-shot learning, function calling, chain of thought reasoning, and health issue diagnosis [8] - The models support context lengths of up to 128,000 tokens [12] Training & Architecture - The models were trained using a mix of reinforcement learning and techniques informed by OpenAI's most advanced internal models [3] - The models utilize a transformer architecture with a mixture of experts, reducing the number of active parameters needed to process input [10][11] - The 120 billion parameter version activates only 5 billion parameters per token, while the 20 billion parameter version activates 36 billion parameters [11][12] - The models employ alternating dense and locally banded sparse attention patterns, group multi-query attention, and RoPE for positional encoding [12] Safety & Security - OpenAI did not put any direct supervision on the chain of thought for either OSS model [21] - The models were pre-trained and filtered to remove harmful data related to chemical, biological, radiological, and nuclear data [22] - Even with robust fine-tuning, maliciously fine-tuned models were unable to reach high capability levels according to OpenAI's preparedness framework [23] - OpenAI is hosting a challenge for red teamers with $500,000 in awards to identify safety issues with the models [24]
AI“新晋顶流”出现了!大厂竞相布局
Zheng Quan Shi Bao· 2025-05-01 11:38
Core Insights - The emergence of the Model Context Protocol (MCP) is seen as a significant advancement in AI development, allowing for easier integration of external data sources and tools, thereby enhancing the efficiency of AI applications and agents [3][5][9] - Major tech companies, including Alibaba, Baidu, Tencent, and ByteDance, are actively adopting and promoting MCP, indicating a competitive landscape for AI agent development [9][10][11] Group 1: MCP Overview - MCP is likened to a "universal socket" for AI, enabling seamless connections between large models and external tools, which significantly reduces development costs and time [3][5][8] - The protocol was initially introduced by Anthropic in November 2022 but gained traction with the launch of the Manus AI agent in February 2023, showcasing the potential of MCP [7][13] - The adoption of MCP is expected to transform AI agents from simple information retrieval systems to more complex applications capable of executing tasks [8][12] Group 2: Industry Adoption - As of April 2025, various tech giants have integrated MCP into their services, with Baidu being the first to offer an enterprise-level MCP service [3][9] - Alibaba Cloud has launched a comprehensive MCP service that integrates over 200 leading models and nearly 100 mainstream MCP services, facilitating easier development of AI agents [10][12] - The introduction of payment MCP services by Alipay further enhances the capabilities of AI agents, allowing for streamlined transaction processes within applications [11][12] Group 3: Future Developments - The MCP ecosystem is still evolving, with ongoing improvements and adaptations expected as the technology matures [13][15] - The competition between MCP and other protocols, such as Google's Agent2Agent Protocol (A2A), highlights the dynamic nature of AI integration standards [14][15] - Industry experts believe that while MCP may face challenges, its foundational role in AI development will continue to be significant as it evolves [15][16]
李彦宏说的「MCP」,还有人不知道吗?
36氪· 2025-04-28 09:44
以下文章来源于智能涌现 ,作者邓咏仪 智能涌现 . 文 | 邓咏仪 编辑 | 苏建勋 来源| 智能涌现(ID: AIEmergence) 封面来源 | AI生成 大模型的风,如今又刮到了一个新名词上:MCP。 AI圈中不缺新鲜事,但这次不一样,互联网仿佛又回到了十多年前的春天。 "现在,基于MCP开发智能体,就像2010年开发移动APP。" 4月25日,百度 董事长李彦宏在百度Create大会上说到。 如果还没有听过MCP,但你肯定听过上一个热词:Agent(智能体)。2025年初,中国初创公司Manus的爆火,把这个名词瞬间推到了大众面前。 "真·能干活的AI",是Agent爆火的关键。在这之前,大模型可以答疑解惑,但它只是一个简单的对话窗口,依赖于模型接受过的训练,大模型内的数据往 往不是最新的,如果只有大模型本体,调用外部工具,要经历非常繁琐的过程。 MCP这个概念,就和Agent密不可分。 MCP是Agent愿景得以实现的的重要路径——大模型可以自由地调用支持MCP协议的外部工具,完成更具体的任 务。 现在,包括高德地图、微信读书在内的应用,就已经纷纷推出官方的MCP Server(服务器),这意味着 ...
李彦宏说的「MCP」,还有人不知道吗?
3 6 Ke· 2025-04-28 01:26
Core Viewpoint - The emergence of MCP (Model Context Protocol) is seen as a pivotal development in the AI industry, akin to the rise of mobile apps in 2010, enabling more efficient interactions between large models and external tools [1][2]. Group 1: Definition and Importance of MCP - MCP is an open standard that allows large models to interact with external data sources and tools, similar to a universal interface like USB [6][12]. - The adoption of MCP is expected to lead to a significant explosion in AI applications by 2025, as it simplifies the development process for AI applications [5][12]. Group 2: Current Trends and Adoption - Since February 2024, a global wave of MCP adoption has occurred, with major companies like OpenAI, Google, and others announcing support for the protocol [2][16]. - Over 4,000 MCP servers have been launched globally, indicating rapid growth in the ecosystem [12]. Group 3: Developer Experience and Challenges - Prior to MCP, developers faced high barriers in integrating external tools with large models, often requiring extensive coding and adaptation [10][11]. - With MCP, developers can focus on maintaining their applications rather than managing external tool performance, significantly reducing development workload [12][13]. Group 4: Competitive Landscape and Strategic Shifts - The shift towards MCP represents a strategic pivot for major AI companies, moving from isolated development to a more collaborative ecosystem [17][21]. - OpenAI's previous closed strategy has been contrasted with MCP's open approach, highlighting the advantages of a more inclusive development environment [18][21].