Transformer架构
Search documents
画到哪,动到哪!字节跳动发布视频生成「神笔马良」ATI,已开源!
机器之心· 2025-07-02 10:40
Core Viewpoint - The article discusses the development of ATI, a new controllable video generation framework by ByteDance, which allows users to create dynamic videos by drawing trajectories on static images, transforming user input into explicit control signals for object and camera movements [2][4]. Group 1: Introduction to ATI - Angtian Wang, a researcher at ByteDance, focuses on video generation and 3D vision, highlighting the advancements in video generation tasks due to diffusion models and transformer architectures [1]. - The current mainstream methods face a significant bottleneck in providing effective and intuitive motion control for users, limiting creative expression and practical application [2]. Group 2: Methodology of ATI - ATI accepts two basic inputs: a static image and a set of user-drawn trajectories, which can be any shape, including lines and curves [6]. - The Gaussian Motion Injector encodes these trajectories into motion vectors in latent space, guiding the video generation process frame by frame [6][14]. - The model uses Gaussian weights to ensure that it can "see" the drawn trajectories and understand their relation to the generated video [8][14]. Group 3: Features and Capabilities - Users can draw trajectories for key actions like running or jumping, with ATI accurately sampling and encoding joint movements to generate natural motion sequences [19]. - ATI can handle up to 8 independent trajectories simultaneously, ensuring that object identities remain distinct during complex interactions [21]. - The system allows for synchronized camera movements, enabling users to create dynamic videos with cinematic techniques like panning and tilting [23][25]. Group 4: Performance and Applications - ATI demonstrates strong cross-domain generalization, supporting various artistic styles such as realistic films, cartoons, and watercolor renderings [28]. - Users can create non-realistic motion effects, such as flying or stretching, providing creative possibilities for sci-fi or fantasy scenes [29]. - The high-precision model based on Wan2.1-I2V-14B can generate videos comparable to real footage, while a lightweight version is available for real-time interactions in resource-constrained environments [30]. Group 5: Open Source and Community - The Wan2.1-I2V-14B model version of ATI has been open-sourced on Hugging Face, facilitating high-quality, controllable video generation for researchers and developers [32]. - Community support is growing, with tools like ComfyUI-WanVideoWrapper available to optimize model performance on consumer-grade GPUs [32].
盘一盘,2017年Transformer之后,LLM领域的重要论文
机器之心· 2025-06-29 04:23
Core Insights - The article discusses Andrej Karpathy's concept of "Software 3.0," where natural language becomes the new programming interface, and AI models execute specific tasks [1][2]. - It emphasizes the transformative impact of this shift on developers, users, and software design paradigms, indicating a new computational framework is being constructed [2]. Development of LLMs - The evolution of Large Language Models (LLMs) has accelerated since the introduction of the Transformer architecture in 2017, leading to significant advancements in the GPT series and multimodal capabilities [3][5]. - Key foundational papers that established today's AI capabilities are reviewed, highlighting the transition from traditional programming to natural language interaction [5][6]. Foundational Theories - The paper "Attention Is All You Need" (2017) introduced the Transformer architecture, which relies solely on self-attention mechanisms, revolutionizing natural language processing and computer vision [10][11]. - "Language Models are Few-Shot Learners" (2020) demonstrated the capabilities of GPT-3, establishing the "large model + large data" scaling law as a pathway to more general artificial intelligence [13][18]. - "Deep Reinforcement Learning from Human Preferences" (2017) laid the groundwork for reinforcement learning from human feedback (RLHF), crucial for aligning AI outputs with human values [15][18]. Milestone Breakthroughs - The "GPT-4 Technical Report" (2023) details a large-scale, multimodal language model that exhibits human-level performance across various benchmarks, emphasizing the importance of AI safety and alignment [26][27]. - The release of LLaMA models (2023) demonstrated that smaller models trained on extensive datasets could outperform larger models, promoting a new approach to model efficiency [27][30]. Emerging Techniques - The "Chain-of-Thought Prompting" technique enhances reasoning in LLMs by guiding them to articulate their thought processes before arriving at conclusions [32][33]. - "Direct Preference Optimization" (2023) simplifies the alignment process of language models by directly utilizing human preference data, making it a widely adopted method in the industry [34][35]. Important Optimizations - The "PagedAttention" mechanism improves memory management for LLMs, significantly enhancing throughput and reducing memory usage during inference [51][52]. - The "Mistral 7B" model showcases how smaller models can achieve high performance through innovative architecture, influencing the development of efficient AI applications [55][56].
你的扫描全能王,作价217亿冲刺港股IPO
量子位· 2025-06-27 10:57
Core Viewpoint - The company, Shanghai Hehe Information Technology, is aiming to become the "first stock of intelligent text recognition" in Hong Kong, following its previous listing on the A-share Sci-Tech Innovation Board. The company has shown significant growth in revenue and user engagement, positioning itself as a leader in the AI sector with a focus on text intelligence technology [2][3][4]. Financial Performance - In 2024, the company reported a revenue of 1.438 billion RMB, a net profit of 400 million RMB, and a gross margin of 84.3% [4][25]. - The revenue growth from 2022 to 2024 was approximately 21% CAGR, with revenues of 989 million RMB, 1.187 billion RMB, and 1.438 billion RMB respectively [25]. - The C-end business accounted for a significant portion of total revenue, with contributions of 82.2%, 84.3%, and 83.8% from 2022 to 2024 [27]. User Engagement - The monthly active users (MAU) for C-end products reached 171 million in 2024, with a paid user ratio of 4.3% [21]. - The company ranks first in China and fifth globally among efficiency AI companies with MAU exceeding 100 million [21][22]. Product Portfolio - The company offers a range of products targeting both C-end and B-end markets, including "Scan All-in-One" and "Business Card All-in-One" for C-end, and "TextIn" and "Qixin Huayan" for B-end [8][12]. - The core technology is based on multi-modal text intelligence, which enhances efficiency in various applications [14][15]. Market Position - The company is positioned as a leading AI firm with a focus on text recognition and processing, competing with major players like OpenAI, Google, Adobe, and Microsoft [5][6][21]. - The global AI product market is projected to grow significantly, with estimates of 46.5 billion USD in 2024 and 228 billion USD by 2029, indicating a robust growth trajectory for the industry [66]. Research and Development - The company has been increasing its R&D investment, with expenditures of 280 million RMB, 323 million RMB, and 390 million RMB from 2022 to 2024, representing about 27% of total revenue [33]. - The workforce consists of 1,053 employees, with 60.6% in R&D roles, highlighting the company's commitment to innovation [35]. Future Plans - The funds raised from the Hong Kong listing will primarily be used for R&D, international expansion, and exploring investment and acquisition opportunities [50].
上海AI Lab主任周伯文:关于人工智能前沿的十个问题
机器人圈· 2025-06-26 10:46
Core Viewpoint - The Shanghai Artificial Intelligence Laboratory aims to become a world-class research institution in the field of artificial intelligence, focusing on strategic, original, and forward-looking scientific research and technological breakthroughs [1]. Group 1: Conference Overview - The inaugural Mingzhu Lake Conference, themed "Multidimensional Breakthroughs and Collaborative Innovation in Artificial Intelligence," will take place from June 12-16, 2025, in Shanghai, attracting nearly 60 young scholars and industry leaders [5][48]. - The conference emphasizes the importance of problem discovery alongside problem-solving, as highlighted by the laboratory's director Zhou Bowen [3][16]. Group 2: Key Questions in AI - Zhou Bowen presented ten critical questions regarding the frontiers of artificial intelligence, including the balance between overall intelligence and unit intelligence, resource allocation in deep reinforcement learning, and the relationship between agents and foundational models [4][19]. - The questions aim to address the challenges and opportunities in AI development over the next 3-5 years, focusing on the systematization, diversification, and advancement of intelligent capabilities [19][20]. Group 3: Importance of Scientific Community - The establishment of the Xinghe Community is intended to foster collaboration and innovation among researchers, emphasizing the need for a platform that encourages the discovery and articulation of significant scientific questions [7][17]. - Historical examples illustrate the impact of scientific communities on innovation, highlighting the necessity of collective efforts in addressing complex scientific challenges [10][12][46]. Group 4: Strategic Scientist Emergence - The emergence of strategic scientists is crucial for addressing major scientific challenges, as evidenced by historical examples where significant scientific advancements were achieved through collaborative efforts [46][47]. - The laboratory aims to cultivate strategic scientists by creating conditions that promote high-intensity input, concentrated task tackling, and dense talent development [47].
致敬钱学森,我国学者开发AI虚拟现实运动系统——灵境,解决青少年肥胖难题,揭示VR运动的减肥及促进大脑认知作用机制
生物世界· 2025-06-24 03:56
Core Viewpoint - Adolescent obesity is a global public health crisis with rising prevalence, leading to increased risks of cardiovascular and metabolic diseases, as well as cognitive impairments [2] Group 1: Research and Development - A research team from Shanghai Jiao Tong University and other institutions developed the world's first VR-based exercise intervention system, REVERIE, aimed at overweight adolescents [4][8] - The REVERIE system utilizes deep reinforcement learning and a Transformer-based virtual coach to provide safe, effective, and empathetic exercise guidance [4][8] Group 2: Study Design and Methodology - The study included a randomized controlled trial with 227 overweight adolescents, comparing outcomes between VR exercise, real-world exercise, and a control group [11] - Participants were assigned to different groups, including VR and real-world sports, with all groups receiving uniform dietary management over an eight-week intervention [11] Group 3: Results and Findings - After eight weeks, the VR exercise group lost an average of 4.28 kg of body fat, while the real-world exercise group lost 5.06 kg, showing comparable results [13] - Both VR and real-world exercise groups showed improvements in liver enzyme levels, LDL cholesterol, physical fitness, mental health, and exercise willingness [13] - VR exercise demonstrated superior cognitive function enhancement compared to real-world exercise, supported by fMRI findings indicating increased neural efficiency and plasticity [14] Group 4: Safety and Implications - The injury rate in the VR exercise group was 7.69%, lower than the 13.48% in the real-world exercise group, with no severe adverse events reported [15] - The REVERIE system is positioned as a promising solution for addressing adolescent obesity and promoting overall health improvements beyond weight loss [16][17]
Transformer 在具身智能“水土不服”,大模型强≠机器人强
3 6 Ke· 2025-06-18 11:55
Core Insights - The year 2025 is anticipated to be the "Year of Embodied Intelligence," driven by significant events and advancements in robotics and AI technologies [1] - There is a growing interest and investment in the field of general robotics, but concerns about sustainability and potential market bubbles persist [1] - Experts are exploring the challenges and advancements in embodied intelligence, focusing on the gap between technological ideals and engineering realities [1] Group 1: Industry Trends - A surge in robotics startups and investments indicates a strong belief in the potential of general robotics [1][2] - The transition from multi-modal large models to embodied intelligence is seen as a natural evolution, requiring substantial data and infrastructure improvements [3][4] - Current AI models face limitations in multi-task scenarios, highlighting the need for better adaptability and learning mechanisms [5][6] Group 2: Technical Challenges - The high energy consumption and training costs of large models pose significant challenges for their application in robotics [4][5] - There is a notable gap between the capabilities of large models and the multi-modal sensory systems of robots, complicating their integration [6][7] - The industry is exploring both modular and end-to-end architectures for embodied intelligence, with a shift towards more unified systems [9][10] Group 3: Research and Development - Research is focused on bridging the gap between human, AI, and robotic intelligence, aiming for better collaboration and understanding [16][18] - The current state of embodied intelligence is limited, with robots primarily executing pre-defined tasks rather than understanding human needs [18][19] - Future developments may involve creating systems that can interpret human intentions directly, bypassing traditional communication methods [20][21] Group 4: Future Outlook - Experts believe that achieving true embodied intelligence will require overcoming significant technical hurdles, particularly in understanding and interacting with the physical world [23][24] - The evolution of AI architectures, particularly beyond the current Transformer models, is essential for the long-term success of embodied intelligence [24][25] - The next five to ten years are expected to be critical for advancements in both hardware and software, potentially leading to widespread adoption of household robots [31][32]
一文了解DeepSeek和OpenAI:企业家为什么需要认知型创新?
混沌学园· 2025-06-10 11:07
Core Viewpoint - The article emphasizes the transformative impact of AI technology on business innovation and the necessity for companies to adapt their strategies to remain competitive in the evolving landscape of AI [1][2]. Group 1: OpenAI's Emergence - OpenAI was founded in 2015 by Elon Musk and Sam Altman with the mission to counteract the monopolistic power of major tech companies in AI, aiming for an open and safe AI for all [9][10][12]. - The introduction of the Transformer architecture by Google in 2017 revolutionized language processing, enabling models to understand context better and significantly improving training speed [13][15]. - OpenAI's belief in the Scaling Law led to unprecedented investments in AI, resulting in the development of groundbreaking language models that exhibit emergent capabilities [17][19]. Group 2: ChatGPT and Human-Machine Interaction - The launch of ChatGPT marked a significant shift in human-machine interaction, allowing users to communicate in natural language rather than through complex commands, thus lowering the barrier to AI usage [22][24]. - ChatGPT's success not only established a user base for future AI applications but also reshaped perceptions of human-AI collaboration, showcasing vast potential for future developments [25]. Group 3: DeepSeek's Strategic Approach - DeepSeek adopted a "Limited Scaling Law" strategy, focusing on maximizing efficiency and performance with limited resources, contrasting with the resource-heavy approaches of larger AI firms [32][34]. - The company achieved high performance at low costs through innovative model architecture and training methods, emphasizing quality data selection and algorithm efficiency [36][38]. - DeepSeek's R1 model, released in January 2025, demonstrated advanced reasoning capabilities without human feedback, marking a significant advancement in AI technology [45][48]. Group 4: Organizational Innovation in AI - DeepSeek's organizational model promotes an AI Lab paradigm that fosters emergent innovation, allowing for open collaboration and resource sharing among researchers [54][56]. - The dynamic team structure and self-organizing management style encourage creativity and rapid iteration, essential for success in the unpredictable field of AI [58][62]. - The company's approach challenges traditional hierarchical models, advocating for a culture that empowers individuals to explore and innovate freely [64][70]. Group 5: Breaking the "Thought Stamp" - DeepSeek's achievements highlight a shift in mindset among Chinese entrepreneurs, demonstrating that original foundational research in AI is possible within China [75][78]. - The article calls for a departure from the belief that Chinese companies should only focus on application and commercialization, urging a commitment to long-term foundational research and innovation [80][82].
大模型专题:大模型架构创新研究报告
Sou Hu Cai Jing· 2025-06-06 11:38
Core Insights - The report focuses on innovations in large model architectures, particularly addressing the limitations of the Transformer architecture and exploring industry pathways for improvement [1][2][7] - As model sizes increase, the secondary computational complexity of Transformers (O(n²)) leads to significant power consumption and efficiency bottlenecks in processing long sequences, prompting a demand for innovative solutions [1][2][15] - The industry is currently exploring two main paths for architectural breakthroughs: improvements to the Transformer architecture and exploration of non-Transformer architectures [1][2][7] Transformer Architecture Improvements - Improvements to the Transformer architecture focus on optimizing the Attention mechanism, Feed-Forward Network (FFN) layers, and normalization layers [1][2][18] - Techniques such as sparse attention and dynamic attention are being developed to enhance computational efficiency, while Mixture of Experts (MoE) aims to improve sparse connection efficiency in FFN layers [1][2][18] - LongRoPE and other technologies are enhancing positional encoding to better model long sequences [1][2][18] Non-Transformer Architecture Exploration - Non-Transformer architectures include new types of RNNs (e.g., RWKV, Mamba) and CNNs (e.g., Hyena Hierarchy), as well as other innovative architectures like RetNet and LFM [1][2][7] - RWKV optimizes state evolution through a generalized Delta Rule, while Mamba leverages state space models to enhance training efficiency [1][2][7] - RetNet combines state space and multi-head attention to achieve parallel computation [1][2][7] Industry Trends and Future Directions - The industry is witnessing a trend towards hybrid architectures that combine linear Transformers with non-Transformer architectures, balancing performance and efficiency [2][7] - The current phase is characterized by a peak in traditional Transformer paradigms and an impending wave of architectural innovations, with significant focus on new RNN/CNN theoretical breakthroughs and practical engineering optimizations [2][7] - Companies like ByteDance and Alibaba are accelerating their investments in hybrid architectures, driving the evolution of large models towards higher efficiency and lower energy consumption [2][7]
三位顶流AI技术人罕见同台,谈了谈AI行业最大的「罗生门」
3 6 Ke· 2025-05-28 11:59
Core Insights - The AI industry is currently experiencing a significant debate over the effectiveness of pre-training models versus first principles, with notable figures like Ilya from OpenAI suggesting that pre-training has reached its limits [1][2] - The shift from a consensus-driven approach to exploring non-consensus methods is evident, as companies and researchers seek innovative solutions in AI [6][7] Group 1: Industry Trends - The AI landscape is witnessing a transition from a focus on pre-training to exploring alternative methodologies, with companies like Sand.AI and NLP LAB leading the charge in applying multi-modal architectures to language and video models [3][4] - The emergence of new models, such as Dream 7B, demonstrates the potential of applying diffusion models to language tasks, outperforming larger models like DeepSeek V3 [3][4] - The consensus around pre-training is being challenged, with some experts arguing that it is not yet over, as there remains untapped data that could enhance model performance [38][39] Group 2: Company Perspectives - Ant Group's Qwen team, led by Lin Junyang, has faced criticism for being conservative, yet they emphasize that their extensive experimentation has led to valuable insights, ultimately reaffirming the effectiveness of the Transformer architecture [5][15] - The exploration of Mixture of Experts (MoE) models is ongoing, with the team recognizing the potential for scalability while also addressing the challenges of training stability [16][20] - The industry is increasingly focused on optimizing model efficiency and effectiveness, with a particular interest in achieving a balance between model size and performance [19][22] Group 3: Technical Innovations - The integration of different model architectures, such as using diffusion models for language generation, reflects a broader trend of innovation in AI [3][4] - The challenges of training models with long sequences and the need for effective optimization strategies are critical areas of focus for researchers [21][22] - The potential for future breakthroughs lies in leveraging increased computational power to revisit previously unviable techniques, suggesting a cycle of innovation driven by advancements in hardware [40][41]
自动驾驶未来技术趋势怎样?李想:现阶段VLA是能力最强的架构
news flash· 2025-05-07 13:27
Core Viewpoint - The CEO of Li Auto, Li Xiang, discussed the transition of the auxiliary driving system to the VLA architecture, questioning its efficiency compared to potential future architectures [1] Group 1 - VLA architecture is capable of addressing full autonomous driving, but its efficiency as the optimal solution is uncertain [1] - Li Xiang highlighted that VLA is still based on the transformer architecture, which raises questions about whether transformer is the most efficient architecture available [1] - Currently, VLA is considered the most powerful architecture in terms of capabilities [1]