Workflow
代码智能体
icon
Search documents
华为云码道代码智能体公测版正式发布 打造全流程企业级研发智能体
Huan Qiu Wang· 2026-02-26 09:42
华为云码道的推出旨在系统性破解企业开发效率低、交付质量不稳等核心研发痛点,助力企业实现研发效能的长期可控。 据悉,华为云码道深度融合代码大模型、IDE与自主开发模式,内置华为在规模化开发中沉淀的需求管理、系统设计、软件开发等高频场景"专家技能"。产 品支持整仓代码索引、多模型适配及合规管控,可实现代码生成、调试、优化等全流程落地能力。 该产品特别针对金融、制造等对合规要求较高的行业场景,提供全流程研发工具链、代码安全管控和合规保障;同时面向注重代码质量和工程可落地性的团 队,支持项目级代码生成与整仓代码管理。值得关注的是,华为云码道原生支持鸿蒙系统官方推荐的ArkTS开发语言,依托Codebase代码库索引优势,为鸿 蒙开发者提供专属适配能力。 【环球网科技报道 记者 张阳】2月26日,华为云码道(CodeArts)代码智能体公测版正式发布。该产品以"AI编码实干派"为定位,依托华为30余年研发实践 积累与千亿级代码库沉淀,打造覆盖全流程的企业级研发智能体。 ...
前 Codex 大神倒戈实锤,吹爆 Claude Code:编程提速 5 倍,点破 OpenAl 死穴在上下文
3 6 Ke· 2026-02-09 11:17
Calvin French-Owen 是 Segment 联合创始人、前 OpenAI 工程师、Codex 项目的早期研发者。他最近在一档播客中,对当前最火的代码智能体 Codex、 Claude Code 和 Cursor 进行了锐评。 结论出人意料,他最常用、也最偏爱的,是 Claude Code,他表示搭配 Opus 模型更"香"。 Calvin 用了一个极具画面感的比喻,来形容用 Claude Code 的体验: 就像残疾人换上了一副仿生膝盖,写代码的速度直接提升了 5 倍。 在他看来,Claude Code 真正的杀手锏,是极其有效的 上下文拆分能力。 面对复杂任务,Claude Code 会自动生成多个 探索型子智能体,独立扫描代码仓库、检索上下文,再将关键信息汇总反馈。这种设计,显著降低了上下文 噪音,也解释了它为何能稳定输出高质量结果。 不过,他也肯定了自家产品,认为 Codex 很有"个性",像 AlphaGo。在调试复杂问题时的表现上,Codex 堪称超人类,很多 Opus 模型解决不了的问题, Codex 都能搞定。 "上下文管理",是 Calvin French-Owen 在整期播客中 ...
前 Codex 大神倒戈实锤!吹爆 Claude Code:编程提速 5 倍,点破 OpenAl 死穴在上下文
AI前线· 2026-02-09 09:12
Core Insights - The article discusses the preferences of Calvin French-Owen, co-founder of Segment and early developer of OpenAI's Codex, who favors Claude Code for its superior coding experience and context management capabilities [4][6][8]. Group 1: Product Comparison - Claude Code is preferred for its effective context-splitting ability, which allows it to generate multiple exploratory sub-agents that independently scan code repositories and summarize key information, significantly reducing context noise [6][17]. - Codex is acknowledged for its unique personality and exceptional performance in debugging complex issues, often outperforming other models in problem-solving [6][8][31]. Group 2: Context Management - Context management is emphasized as a critical factor in the performance of coding agents, with Calvin suggesting that when context token usage exceeds 50%, it is essential to clear the context to maintain efficiency [7][20][26]. - A practical method shared involves embedding verifiable but irrelevant information in the context to detect when the model begins to forget, indicating context pollution [7][28]. Group 3: Future Trends - The distribution model for products is becoming increasingly important, with a shift towards bottom-up distribution where engineers adopt tools without waiting for approvals [9][10][33]. - The future may see smaller companies with more individual smart agents, allowing engineers to manage tasks more effectively and focus on higher-level decision-making [12][36]. Group 4: Development and Integration - The integration and orchestration capabilities of coding agents are seen as new constraints, particularly in code review processes and ensuring the validity of code modifications [50]. - Testing is highlighted as crucial for enhancing coding efficiency, with a strong emphasis on achieving high test coverage to ensure stability and reliability in code execution [50][51]. Group 5: Industry Implications - The article suggests that the rise of coding agents like Claude Code and Codex will lead to a transformation in how software development is approached, with a focus on automation and efficiency [36][48]. - The potential for a future where every worker has their own cloud-based intelligent team is discussed, indicating a shift in workplace dynamics and productivity [38][39].
开源框架让代码AI偷师GitHub,bug修复率飙升至69.8%,性能创纪录
3 6 Ke· 2026-01-16 09:54
Core Insights - The article discusses the limitations of current AI in bug fixing and introduces the MemGovern framework, which aims to enhance AI's ability to learn from human experiences in software engineering [2][3][28]. Group 1: AI Limitations and Challenges - Current AI systems struggle to effectively utilize the vast amount of historical data available on platforms like GitHub due to their "closed world" cognitive limitations [3][4]. - Human engineers often rely on community knowledge to solve complex issues, but AI's direct access to unstructured data from GitHub is hindered by noise and ambiguity [4][8]. Group 2: MemGovern Framework - MemGovern introduces an "Experience Refinement Mechanism" that transforms raw GitHub data into structured, AI-friendly "experience cards" [9][10]. - The framework employs a hierarchical selection process to filter high-quality repositories and cleanses data to retain only complete repair records [9][10]. Group 3: Experience Cards - Experience cards consist of two layers: an Index Layer for efficient symptom-based retrieval and a Resolution Layer that encapsulates root cause analysis, fix strategies, patch summaries, and verification methods [10][12]. - The structured design of experience cards enhances the usability of knowledge and allows for better retrieval of repair logic [10][12]. Group 4: Search Mechanism - MemGovern utilizes a "Search-then-Browse" approach, allowing AI to first search for relevant cases based on symptoms and then browse detailed solutions, mimicking human search behavior [12][13]. - This method enables AI to understand repair logic more deeply and filter out irrelevant information [12][13]. Group 5: Experimental Results - MemGovern has shown significant improvements in bug resolution rates across various models, with Claude-4-Sonnet+MemGovern achieving a 69.8% resolution rate, a 3.2% increase over the baseline [15][16]. - GPT-4o+MemGovern's resolution rate increased from 23.2% to 32.6%, marking a 9.4% improvement [16]. Group 6: Broader Implications - The MemGovern framework not only enhances performance metrics but also provides a clear pathway for AI agents to effectively utilize vast amounts of unstructured human debugging experience [28]. - The methodology has potential applications beyond coding, such as in legal consulting and medical diagnosis, where historical case knowledge is crucial [28].
第一名方案公开,代码智能体安全竞赛,普渡大学拿下90%攻击成功率
机器之心· 2025-08-23 10:51
Core Insights - The article highlights the vulnerabilities of AI programming assistants, indicating that even well-aligned large language models can inadvertently generate code with security flaws, which can be exploited by malicious users to accelerate malware development [2][4][29] - The Amazon Nova AI Challenge showcased the effectiveness of red team strategies in identifying security vulnerabilities in AI code models, with the PurCL team achieving over 90% success in attacks [7][29] Group 1: AI Model Security Challenges - Recent studies reveal that the security of AI models is compromised by subtle flaws in the reasoning chain, not just by explicit input-output issues [2][4] - The PurCL team developed a comprehensive red team system based on AI cognitive modeling, which was shared with the research community [3][21] - The challenge of aligning code models lies in extending alignment techniques to complex real-world problems and enhancing the security relevance of model reasoning [4][32] Group 2: Amazon Nova AI Challenge - The competition involved 12 teams over eight months, with a total investment of one million dollars, focusing on identifying vulnerabilities in AI code models [3][7] - The competition's structure included red teams attempting to find vulnerabilities and blue teams applying security alignment practices to defend against these attacks [7][29] - The PurCL team emerged as the winner of the red team category, demonstrating the inadequacy of current AI safety research in addressing real-world model security issues [7][29] Group 3: AI Cognitive Modeling - The PurCL team proposed a cognitive modeling approach that divides human cognition into "problems," "inference," and "solutions," which can be applied to AI code generation [12][14] - Their research identified that existing security classifiers struggle with domain-specific knowledge, leading to a significant drop in effectiveness in complex fields like cybersecurity [19][20] - The team developed a knowledge modeling system to identify potential security risks in complex domains, revealing significant gaps in current alignment solutions [23][29] Group 4: ASTRA Reasoning Path Analysis - The ASTRA method was created to analyze the reasoning paths of AI models, identifying weaknesses in their inference processes [25][29] - This method allows for the generation of targeted input modifications to bypass model defenses, significantly enhancing red team testing depth [25][29] - The PurCL team found that many state-of-the-art models, including GPT-5, could assist in generating malicious code under certain conditions [29][30]