Workflow
Miles
icon
Search documents
AI 已能写 80% 代码,但 Agent 也有致命短板!OpenAI Codex 技术总监:问错了,比不会写更麻烦
AI前线· 2026-03-28 05:33
Core Insights - The article discusses the evolution of software engineering tools and the importance of problem-solving in the age of AI programming, emphasizing that the key differentiator is not the speed of coding but the choice of problems to solve and the definition of a "better system" [2][3]. Group 1: Key Perspectives on Engineering - Many engineering breakthroughs stem from dissatisfaction with the status quo and rapid hands-on validation [3]. - The influence of engineers ultimately depends on whether they address the "real concerns of the company" [3]. - In the AI programming era, 80%-90% of code can be generated by models, but critical parts still require human oversight [3]. - The ability to ask the right questions is becoming more important than writing code [3]. Group 2: Transition from Engineer to Tool Creator - The journey from being an engineer to a tool creator is driven by problem-solving [5]. - The initial project, Chickenfoot, was a Firefox extension that aimed to facilitate web programming through user commands [6]. - The project shared similarities with current AI programming assistants, highlighting the evolution of natural language processing [7]. Group 3: Experiences at Google and Facebook - Google was attractive due to its innovative culture and focus on user experience, but there were internal divisions between product and infrastructure teams [9]. - The experience at Google was mixed, with a realization that personal interests often did not align with the company's core revenue-generating projects [10][11]. - At Facebook, the focus shifted to mobile development, with a significant emphasis on building tools for Android [12][13]. Group 4: Building Efficient Systems - The need for a new build system at Facebook arose from the inefficiencies of existing tools, leading to the creation of a more modular and efficient system [14][20]. - The new system improved performance by caching results and allowing for more granular incremental builds [19][20]. - The approach represented not just a technical optimization but a shift in mindset towards reducing repetitive work [22]. Group 5: Development of IDE and Virtual File Systems - After addressing Android build issues, the focus shifted to improving IDEs, particularly due to the limitations of Xcode for large-scale applications [25][28]. - The development of a virtual file system aimed to alleviate the challenges posed by monorepo structures, enhancing efficiency by delaying file loading [39][40]. - The new system allowed for dynamic file generation and improved search capabilities, significantly enhancing performance [42][43]. Group 6: Transition to OpenAI and Codex - The move to OpenAI was motivated by the desire to work with top talent and to engage in consumer-facing projects, contrasting with the internal focus at Meta [51][53]. - The Codex project faced initial challenges but gained traction through community engagement and open-source contributions [57].
Z Tech | LMSYS 团队发布大规模  MoE 强化学习框架 Miles,不积跬步无以至千里
Z Potentials· 2025-11-20 04:12
Core Insights - The article introduces Miles, a new reinforcement learning framework designed for enterprise-level large-scale MoE training and production workloads, developed by the LMSYS team as a fork of the lightweight framework slime [1][4]. Group 1: Framework Features - Miles inherits the lightweight and modular design principles of slime, making it a preferred tool for model scientists exploring algorithms [3]. - It implements Infrastructure-level True On-Policy to eliminate discrepancies between training and inference, achieving bit-wise consistency [5]. - The framework introduces speculative training through MTP Online Training, resulting in over 25% rollout acceleration [3][9]. Group 2: Memory Optimization - Miles incorporates advanced memory management techniques to maximize GPU performance without triggering out-of-memory (OOM) errors [8]. - It features online SFT for Draft Models, which enhances performance by preventing a decline in acceptance length during training [9]. - The framework includes mechanisms to avoid benign OOM errors and implements memory margin strategies to address NCCL-related OOM issues [10]. Group 3: Technical Upgrades - Miles supports full-stack optimization for SGLang and Megatron, ensuring compatibility with rapid iterations in training and inference frameworks [6]. - The modular design allows researchers to easily modify components like algorithms, data, sampling, and evaluation with minimal code changes [6]. - It provides a user-friendly interface for model scientists, allowing them to adjust important sampling or loss dynamics without delving into lower-level code [6]. Group 4: Future Development - The LMSYS team plans to enhance the FSDP backend for improved stability in large-scale distributed training [14]. - Future developments include independent rollout deployment, additional debugging tools, and formal mathematical verification for SFT/RL scripts [14]. - The roadmap also aims to support next-generation hardware like GB300 and expand capabilities for multi-modal training [18].
速递|红杉、a16z竞逐AI语音战场:初创公司Sesame获2亿美元融资
Z Potentials· 2025-03-31 06:34
Core Viewpoint - The article discusses the interest of major venture capital firms, including Sequoia Capital and A16Z, in investing in the voice AI startup Sesame, which focuses on developing AI voice assistants and wearable devices [1][2]. Group 1: Investment and Valuation - Sesame is reportedly in discussions to raise at least $200 million, with potential valuation discussions reaching into the billions [2]. - A16Z has led Sesame's Series A funding round, although specific terms and timelines have not been disclosed [4]. Group 2: Technology and Product Development - Sesame has launched its voice assistants, Maya and Miles, which are accessible via smartphones and laptops, and aims to integrate voice assistants into glasses for hands-free communication [3]. - The voice AI technology is based on Meta's Llama large language model and has been enhanced through training on approximately 1 million hours of audio primarily in English [3]. Group 3: Market Context and Competition - The growing interest in AI that can communicate like humans is highlighted, with industry leaders like OpenAI and Meta also developing voice capabilities for their text-based AI products [2][6]. - Sesame may become an acquisition target for companies like Meta, OpenAI, Anthropic, or xAI, which are already working on integrating voice features into their AI systems [5].