上下文工程
Search documents
Manus“删博、裁员、跑路新加坡”后,创始人首次复盘经验教训
Hu Xiu· 2025-07-19 06:44
Group 1 - Manus experienced rapid growth and controversy within four months, transitioning from a successful startup to facing significant public scrutiny [1][4][6] - The company raised $75 million in Series B funding led by Benchmark, achieving a valuation of $500 million, which generated high expectations from the market [5] - Controversies arose in late June, including unannounced layoffs, mass deletion of posts by the founding team, and the company's relocation to Singapore, leading to public outcry [6][7] Group 2 - Co-founder Ji Yichao addressed the controversies through a lengthy blog post, focusing on the product and technology rather than the company's issues [3][8] - Manus chose to focus on context engineering instead of developing an end-to-end model, learning from past experiences with large models like GPT-3 [8][12] - Key insights from the blog include the importance of KV cache hit rate, managing tool availability without dynamic changes, and treating the file system as an external memory [8][9][10][34] Group 3 - The company emphasizes the need to retain error information in the context to help the model learn from mistakes, which is crucial for improving agent behavior [11][50] - Manus aims to avoid being limited by few examples by introducing structured variations in actions and observations, which helps break patterns and adjust model attention [52][54] - The conclusion highlights that context engineering is vital for agent systems, influencing their speed, recovery ability, and scalability [56]
回应撤离中国市场原因,Manus首度披露技术侧经验教训
Di Yi Cai Jing· 2025-07-19 06:17
Core Insights - Manus has withdrawn from the Chinese market and is focusing on international expansion, citing operational efficiency adjustments and internationalization strategies as the main reasons for this shift [2] - The co-founder of Manus, Ji Yichao, emphasized the importance of context engineering in their technology strategy, aiming to enhance product iteration speed by leveraging memory and process construction [2][4] - The company has learned from past experiences, particularly from their previous venture, Peak Labs, and has decided to avoid investing in foundational model development, instead opting to utilize open-source models for training [5] Context Engineering - Context in large models refers to the information set that models reference when processing tasks or generating outputs, which enhances understanding and performance [3] - The concept of Lossless Long Context is crucial for AI-native products, as it allows for personalized interactions by effectively utilizing user interaction history [3] - The Key-Value Cache (KV-Cache) hit rate is vital for improving inference efficiency and optimizing resource utilization, thereby reducing computational costs [3] Lessons Learned - Ji Yichao reflected on the lessons learned from Peak Labs, where the decision to develop a model from scratch became irrelevant after the emergence of advanced models like OpenAI's GPT-3 [4] - The Manus team has undergone multiple adjustments to their Agent framework to achieve a locally optimal solution, recognizing the challenges of relying on external models for task execution [5] - Despite the focus on efficiency, Manus faces limitations compared to competitors like OpenAI, which utilize proprietary models for better handling of complex tasks [5] Market Challenges - As Manus shifts to the international market, it faces competition from larger platforms that attract developers and users, posing a threat to market share for startups [5] - The current landscape for Agent products is characterized by significant homogenization, unclear business models, and high costs, making it challenging for startups to differentiate themselves [5] - Continuous optimization of technical strategies and exploration of differentiated development paths are essential for Manus to navigate these market challenges [5]
Manus季逸超:构建Manus的经验教训 | Jinqiu Select
锦秋集· 2025-07-19 05:00
Core Viewpoint - The article discusses the choice between end-to-end training and context engineering in developing general AI agents, highlighting the latter as a more adaptable approach in a rapidly evolving landscape of large models [1][3]. Group 1: Context Engineering Insights - Manus AI's decision to adopt context engineering was influenced by past experiences where self-trained models quickly became obsolete after the release of GPT-3, emphasizing the need for flexibility in model development [4][5]. - The article outlines six core practices derived from Manus's experience, which significantly reduced product iteration cycles from weeks to hours, showcasing an effective technical path for startups [2][3]. Group 2: Key Practices for KV-Cache Optimization - The KV-cache hit rate is identified as the most critical metric for AI agents in production, directly affecting latency and cost, with a notable example showing a 10x cost difference between cached and uncached tokens [7][8]. - Strategies to enhance KV-cache hit rates include maintaining stable prompt prefixes, using only appended context, and employing file systems as external memory to overcome context limitations [8][19]. Group 3: Managing Tool Complexity - The article advises against dynamically adding or removing tools in the agent's action space, suggesting instead to manage tool availability through context-aware masking of token logits to maintain stability [12][13]. - This approach helps prevent confusion in the model when previous actions reference tools that are no longer defined, thereby reducing the risk of erroneous actions [12][17]. Group 4: Utilizing External Memory - Manus employs a file system as an externalized memory solution to address the limitations of context windows, allowing for persistent and unlimited storage that can be directly manipulated by the agent [18][22]. - This method mitigates the risks associated with irreversible context compression, ensuring that critical information is not lost [22]. Group 5: Attention Manipulation Techniques - The use of a todo.md file to continuously update task goals serves as a mechanism to keep the model focused on its objectives, preventing it from losing track during complex tasks [23][26]. - This technique helps maintain the model's attention on the task at hand, especially in lengthy interactions requiring multiple tool calls [26]. Group 6: Learning from Errors - Retaining failed attempts in the context is emphasized as a crucial learning mechanism, allowing the model to adapt and reduce the likelihood of repeating mistakes [30][31]. - The article argues that error recovery is a significant indicator of an agent's performance, yet it is often underrepresented in academic benchmarks [30]. Group 7: Avoiding Few-Shot Traps - The article warns against the pitfalls of few-shot learning in agent systems, where repetitive patterns in context can lead to suboptimal decision-making [32][34]. - Introducing structured variability in actions and observations can help break these patterns and enhance the model's adaptability [34]. Conclusion - Context engineering is presented as an essential and emerging science for agent systems, with the design of context playing a pivotal role in defining agent behavior, speed, recovery, and scalability [35].
Manus「删博跑路」后,创始人首次深度复盘:公开产品细节,总结教训
3 6 Ke· 2025-07-19 01:15
Core Insights - Manus AI has abruptly withdrawn from the Chinese market, clearing all social media content and seemingly pausing the development of its Chinese version, following the relocation of its global headquarters to Singapore [1] - The co-founder of Manus AI, Ji Yichao, published a technical blog to refocus attention on the product's technology amidst the controversy, sharing valuable lessons learned during the development of Manus [3][9] Group 1: Company Developments - Manus AI has moved its global headquarters to Singapore and has offices in Tokyo and California, indicating a strategic shift in its operational focus [1] - The company has faced scrutiny and speculation regarding potential layoffs and whether it is abandoning the Chinese market [1] Group 2: Technical Insights from the Blog - The blog emphasizes the importance of context engineering over traditional model training, allowing for quicker product updates [6][10] - Key practices for improving KV-cache hit rates are outlined, including maintaining stable prompts, appending context only, and marking cache breakpoints [12][16][17] - The use of a file system for persistent context is recommended to manage the limitations of context windows in modern AI models [25][30] - The blog discusses the significance of maintaining attention through continuous updates to a todo list, which helps keep the model focused on its goals [31][34] - It highlights the importance of retaining error logs to improve model behavior and reduce the likelihood of repeating mistakes [35][38] - The introduction of structured variations in actions and observations is suggested to prevent the model from falling into repetitive patterns [39][41] Group 3: Future Implications - The article concludes that context engineering is essential for the future of agent systems, as it defines the behavior, speed, recovery, and scalability of AI agents [42]
来自 Manus 的一手分享:如何构建 AI Agent 的上下文工程?
Founder Park· 2025-07-18 18:51
Manus 官网昨天更新了一篇文章,分享了他们为 Manus 搭建合适的上下文工程的经验教训。 作者季逸超 (Peak),Manus 公司联合创始人、首席科学家。 文章基于 Kimi K2 翻译,我们进行了一些调整。 在 Manus 项目伊始,我和团队就面临一个关键抉择:是利用开源基础模型训练一个端到端的智能体,还是依托前沿模型的上下文学习能力,在其之上 构建智能体? 在我投身 NLP 的第一个十年里,我们并没有这种奢侈的选择。遥想当年 BERT 问世(没错,那已是七年前),模型必须先经过微调——还要评估—— 才能迁移到新任务。每次迭代往往耗时数周,尽管那时的模型体积与今日的 LLMs 相比微不足道。对于快速迭代的应用,尤其是 PMF 之前的阶段,如 此缓慢的反馈循环几乎是致命的。这是我上一家初创公司留下的惨痛教训:当时我从零开始训练模型,用于开放信息抽取和语义搜索。随后 GPT-3 与 Flan-T5 横空出世,我那些自研模型一夜之间便失去了意义。颇具讽刺意味的是,正是这些新模型开启了上下文学习的大门——也为我们指明了一条全 新的道路。 这个来之不易的教训让选择变得清晰:Manus 将押注于上下文工程。这让 ...
当 LLM 编程陷入“幻觉陷阱”,字节工程师如何用 ABCoder 精准控场
AI科技大本营· 2025-07-16 06:19
Core Insights - The article discusses the limitations of large language models (LLMs) in handling complex enterprise-level programming tasks, highlighting the "hallucination" problem where AI generates inaccurate or irrelevant code outputs [1] - A study by METR revealed that using AI programming assistants did not improve efficiency but instead increased development time by an average of 19%, due to high costs associated with reviewing and debugging AI-generated content [1] - ByteDance has introduced ABCoder, a tool designed to address these challenges by providing a clear and unambiguous code "worldview" through deep parsing of abstract syntax trees (AST), enhancing the model's contextual understanding [2] Group 1 - The hallucination problem in LLMs leads to inaccurate code generation, particularly in complex systems [1] - The METR study involved 16 experienced engineers completing 246 programming tasks, showing a 19% increase in development time when using AI tools [1] - ABCoder aims to improve the reliability of AI programming by enriching the model's context acquisition capabilities, thus reducing hallucinations and enabling more accurate code generation [2] Group 2 - ABCoder's implementation will be explained in a live session, showcasing its real-world applications in backend development [3] - The live session will feature a case study on the CloudWeGo project, demonstrating how ABCoder enhances code development efficiency and optimizes the programming experience [3] - ABCoder functions as a powerful toolbox for developers, offering tools for code understanding and conversion to tackle complex programming challenges [3]
DeepSeek流量暴跌,全球AI霸主地位遇滑铁卢;90后开发者6个月狂赚8000万;人形机器人A轮5亿融资|混沌AI一周焦点
混沌学园· 2025-07-11 07:55
Core Trends - The "Chaos AI Business Practical National Tour" has successfully commenced, aiming to ignite practical applications of AI across 20 innovative cities in China, with events already held in Changsha and Nanchang [1][2] - The AI application landscape is evolving with lower entry barriers due to open-source models and contextual engineering, enabling disruptive innovations that empower ordinary individuals [2] - AI penetration in vertical industries is increasing, particularly in pharmaceuticals, digital healthcare, and live service sectors, indicating potential transformative changes [2] AI Applications - Feishu has launched a comprehensive upgrade of its AI product matrix, including knowledge Q&A and AI meetings, along with the industry's first AI application maturity standard to facilitate enterprise AI adoption [3][4] - Google DeepMind's spinoff, Isomorphic Labs, is set to begin human trials for its AI-assisted cancer drug, marking a significant milestone in the pharmaceutical industry [12][13] Investment and Financing - Star Sea Map has raised over $100 million in its A4/A5 financing rounds, with a total pre-A and A round financing of nearly 1.5 billion yuan, reflecting strong capital interest in the embodied intelligence sector [6][7] - TARS, founded by former Huawei employees, completed a record $122 million angel round financing, showcasing investor confidence in embodied intelligence technologies [13] - Cloud Deep has secured nearly 500 million yuan in financing, positioning itself as a leader in the quadruped robot field with over 600 industry projects [14] - Star Motion Era has raised nearly 500 million yuan in its A round financing, emphasizing breakthroughs in humanoid robot technology and significant global demand [16] Business Cases - Wix's acquisition of AI startup Base44 for $80 million highlights the trend of AI enabling entrepreneurship, with Base44 allowing users to generate full-stack application code through natural language [7][8] - The AI personal finance assistant, Kapi Accounting, has gained over one million users in six months, indicating a shift in personal finance management through AI [21][22] Market Insights - The digital human market in China is projected to reach 30 billion yuan by 2025, with significant cost reductions in enterprise live streaming [19][20] - The rise of "contextual engineering" in Silicon Valley is reshaping AI model development, enhancing efficiency and application quality [18][20] Technology Developments - Baidu has open-sourced ten major models, significantly lowering the barriers for AI development and enhancing multi-modal capabilities [21] - The introduction of the Star Stream Agent, designed for Chinese designers, aims to revolutionize the design industry with automated processes and multi-modal content creation [24]
7月19日,相聚北京!一起聊聊ACL 2025爆点研究
机器之心· 2025-07-10 08:35
Core Insights - The AI field continues to be an exciting area in 2025, with numerous research releases from major tech companies and institutions [1] - The rapid pace of technological advancements in AI is overwhelming, with new models and paradigms emerging almost weekly [3][4] - Developers and researchers are increasingly engaging in conferences and academic sharing to stay updated on cutting-edge research [5] Event Overview - The ACL conference, a significant event in the NLP field, received over 8,000 submissions this year, marking a historical high [6] - The ACL 2025 conference will take place from July 27 to August 1 in Vienna, Austria, featuring various activities such as keynote speeches, paper presentations, roundtable discussions, and poster sessions [6][7] - The event aims to provide a platform for domestic AI talent, with a full schedule of presentations and discussions announced [6] Keynote Speakers and Topics - The keynote address on "Trends and Outlook for ACL 2025" will be delivered by Che Wanxiang, a prominent professor from Harbin Institute of Technology [9][17] - Liu Pengfei from Shanghai Jiao Tong University will present on "Reinforcement Learning and Complex Reasoning in Large Models" [11][19] Paper Presentations - Various papers will be presented, covering topics such as the intrinsic self-correction of large language models and the acceleration of inference in large language models [9][12] - The event will also feature poster sessions and opportunities for industry engagement [21]
苹果开发者自曝用Claude完成95%开发,开发应用已上架
量子位· 2025-07-07 09:35
闻乐 发自 凹非寺 量子位 | 公众号 QbitAI 苹果开发者自曝用AI开发应用程序, Claude含量95% ! 事情是这样的,一位苹果开发者最新发布了一款用于调试MCP服务器的原生macOS应用 Context —— 一款几乎完全由 Claude Code 构建的应用程序。 作者 indragiek 从2008年就开始为Mac开发软件。 这次,他的目标是使用Apple的SwiftUI框架,打造一款在macOS平台上使用起来很顺手且实用的开发者工具。 与以往不同的是,Claude Code承担了Context项目95%的工作量,indragiek声称: 在这个 20000行 代码的项目中,我亲手编写的代码估计 不到1000行 。 "工程师"Claude也是好起来了,能给苹果打工(doge)。 调侃归调侃,下面让我们来"学习"一下这位开发者是怎么用Claude的。 苹果开发者教你"驯服"Claude 作为一名经验丰富的工程师,Indragie像许多同行一样,拥有一个"烂尾项目"list。 尽管能够构建项目原型,但最后20%的交付工作往往耗费巨大时间和精力,导致项目搁置。 所以,他已经6年未能成功发布任何一个 ...
Karpathy最新脑洞「细菌编程」:优秀的代码应该具备细菌的三大特质
量子位· 2025-07-07 04:02
Core Viewpoint - The article discusses Andrej Karpathy's new concept of "Bacterial Code," which emphasizes small, modular, self-contained code blocks that are easy to copy and paste, inspired by the evolutionary strategies of bacteria [1][5][6]. Group 1: Concept of Bacterial Code - Bacterial Code has three main characteristics: small code blocks, modularity, and self-containment, allowing for easy replication [1][6][12]. - The idea is that open-source communities can thrive through "horizontal gene transfer," similar to how bacteria share genetic material [2][12]. - Karpathy's insights are derived from the survival strategies of bacteria, which have evolved to colonize diverse environments through efficient genetic coding [7][8]. Group 2: Principles of Bacterial Code - The first principle is "smallness," where each line of code consumes energy, leading to a natural self-optimization mechanism [8][11]. - The second principle is "modularity," where code should be organized into interchangeable modules, akin to bacterial operons, promoting high cohesion and low coupling [11][12]. - The third principle is "self-containment," meaning code snippets should be independent and not reliant on complex configurations or external libraries [13][14]. Group 3: Limitations and Future Directions - While Bacterial Code is effective for rapid prototyping, it is not suitable for building complex systems, which require more intricate structures like eukaryotic genomes [15][16]. - Karpathy suggests a hybrid approach, utilizing the strengths of both bacterial and eukaryotic coding strategies [16]. Group 4: Evolution of Software Development - Karpathy has previously introduced concepts like Software 3.0, which represents a shift towards programming with natural language models [18][25]. - He notes that software has undergone significant transformations in recent years, moving from traditional coding to model training and now to natural language programming [19][23][31]. - The future of software development will involve a collaboration between humans and large models, leading to semi-autonomous applications [28][30]. Group 5: Context Engineering - Context Engineering is highlighted as a crucial skill for effectively utilizing large language models (LLMs), requiring a balance of information to optimize performance [36][39]. - This discipline involves understanding the behavior of LLMs and integrating various elements like task descriptions and multimodal data [40][41].