Evo 2
Search documents
注意力机制大变革?Bengio团队找到了一种超越Transformer的硬件对齐方案
机器之心· 2026-01-07 05:16
Core Insights - The article discusses the evolution of large language models (LLMs) and highlights the limitations of existing linear recurrence and state space models in terms of computational efficiency and performance [1][3]. - A new approach proposed by Radical Numerics and the Montreal University team focuses on redefining linear recurrences as hardware-aligned matrix operations, aiming to enhance GPU memory utilization and computational efficiency [1][2]. Group 1: Challenges and Limitations - The primary challenge identified is breaking through the "memory wall" associated with linear recurrences, which limits performance due to high communication costs in modern hardware [3][7]. - Traditional parallel scan algorithms, while theoretically efficient, struggle with data access patterns that lead to frequent global memory synchronization, thus failing to leverage data locality effectively [4][5][6]. Group 2: Proposed Solutions - The paper introduces the Sliding Window Recurrences (SWR) as a method to achieve high throughput by strategically truncating the computational horizon, utilizing a jagged window structure that aligns with hardware workloads [10][11]. - The Block Two-Pass (B2P) algorithm is developed to implement this theory, dividing the computation into two phases to optimize memory access and minimize data movement [14][15]. Group 3: Phalanx Layer and Performance - A new computing layer called Phalanx is designed based on the B2P algorithm, serving as a seamless replacement for sliding window attention or linear recurrence layers, ensuring numerical stability during long sequence processing [19][20]. - In systematic tests on a model with 1.3 billion parameters, the Phalanx hybrid model demonstrated significant performance advantages, achieving 10% to 40% end-to-end speedup in training throughput across varying context lengths [23][24]. Group 4: Industry Implications - The findings from the paper indicate that true efficiency in LLMs arises not just from reduced algorithmic complexity but from a deep understanding and alignment with the physical characteristics of underlying computational hardware [31][32]. - As LLMs evolve towards larger context sizes and real-time embodied intelligence post-2025, hardware-aware operator design will be crucial for developing more efficient and powerful AI systems [33].
图数室丨回看2025,AI那些“封神”瞬间
Xin Lang Cai Jing· 2025-12-26 09:28
Core Insights - The year 2025 is marked as a transformative period for artificial intelligence (AI), transitioning from experimental concepts to practical applications in everyday life, signifying the "Year of AI for All" [2] Group 1: Major Developments in AI - DeepSeek launched its R1 model on January 20, 2025, creating a significant impact in the global AI community [4] - On February 17, xAI released its latest AI model, Grok 3, followed by the introduction of the Evo 2 AI biology model by a collaboration of institutions on February 19 [6] - Google unveiled its strongest reasoning model, Gemini 2.5 Pro, on March 25, the same day OpenAI introduced a native image generation feature based on the GPT-4o model [6] Group 2: Notable Events and Milestones - The first humanoid robot marathon was held in Beijing on April 13, with the humanoid robot "Tiangong" winning the race [6] - By July 2025, China had registered 433 large models that were operational [8] - The EU's AI Act came into effect in August, establishing the first comprehensive regulatory framework for AI globally [8] Group 3: Product Launches and Innovations - On May 22, Anthropic launched the Claude 4 series of large models, and on July 18, OpenAI released a new intelligent product called "ChatGPT Agent" [8] - In October, ByteDance's Volcano Engine released an updated version of the Doubao model, which became the first in China to support "tiered adjustment of thinking length" [10] - Google's "Solar Catcher" project was publicly announced on November 5, aiming to create space-based machine learning data centers [12]
世界首个AI设计的病毒诞生,能够感染并杀死耐药细菌,人类达成AI生成生命的里程碑!
生物世界· 2025-09-20 03:05
Core Viewpoint - The article discusses a significant advancement in artificial intelligence (AI) where the first AI-designed virus (bacteriophage) has been created, capable of infecting and killing antibiotic-resistant bacteria, showcasing the potential of AI in developing biotechnological tools and therapies for bacterial infections [3][11]. Group 1: AI and Virus Design - The research published on September 17, 2025, demonstrates the ability of AI to design a complete viral genome, marking a major step towards AI-generated life [3][6]. - The AI models, Evo 1 and Evo 2, were trained on extensive genomic data, enabling them to predict and generate DNA sequences at a genome scale [7][11]. - The team used the bacteriophage ΦX174 as a template for designing the viral genome, which consists of 5,386 nucleotides and encodes 11 genes, showcasing the complexity of the design process [8][9]. Group 2: Research Findings - The research team evaluated thousands of AI-generated sequences, narrowing them down to 302 viable candidates, with 16 capable of self-replication and specifically infecting E. coli strains [11][12]. - Among the functional genomes, each contained 67-392 new mutations compared to their closest natural counterparts, indicating the potential for creating new species through AI [12]. - The AI-designed phages demonstrated the ability to infect and kill three different E. coli strains, outperforming the wild-type ΦX174 phage [11][12]. Group 3: Safety and Future Applications - Concerns about the potential misuse of AI models for harmful virus design are addressed, emphasizing that the training data excluded viruses affecting eukaryotes, including humans [14]. - The research team aims to utilize this method for safely generating AI-designed viruses to tackle various diseases and public health issues, particularly antibiotic resistance [15].