具身智能之心
Search documents
今年9家盈利最高的人形机器人公司......
具身智能之心· 2025-12-03 03:47
Core Insights - The article provides an overview of the order amounts and shipment volumes of leading robotics companies projected for 2025, highlighting the top nine companies in terms of revenue and their core clients [1]. Group 1: Company Orders and Shipments - Zhongqing Robotics: Order amount of 200 million yuan over three years, with a shipment volume of 2,000 units, collaborating with major firms like Multi-Tech and Amazon [2]. - Songyan Power: Order amount exceeding 100 million yuan, with annual shipments surpassing 2,500 units, focusing on education, research, and commercial performances [2]. - Stardust Intelligence: Approximately 500 million yuan in orders, with plans to deploy over 1,000 AI robots in industrial manufacturing and logistics over the next two years [2]. - Zhifang Technology: Order amount of 500 million yuan, with over 1,000 units to be delivered in three years, primarily for industrial applications [2]. - Leju Robotics: Order amount around 500 million yuan, with nearly 2,000 units shipped annually [2]. - Zhiyuan Robotics: Order amount approximately 700 million yuan, with thousands of units shipped, serving various industrial applications [2]. Group 2: Major Players and Their Orders - UBTECH Technology: Order amount exceeding 800 million yuan, with around 2,700 units shipped, primarily serving automotive manufacturers and data collection needs [3]. - Yuejiang Robotics: Order amount around 1.1 billion yuan, with annual shipments of about 20,000 units, projecting 80,000 units in 2024 and 100,000 units in 2025 [3]. - Yushu Technology: Order amount close to 1.2 billion yuan, with over 10,000 units shipped, collaborating with various educational institutions and companies for research and development [3].
五年,终于等来Transformers v5
具身智能之心· 2025-12-03 03:47
编辑丨 机器之心 点击下方 卡片 ,关注" 具身智能之心 "公众号 >> 点击进入→ 具身 智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区: 具身智能之心知识星球(戳我) ,这里包含所有你想要的! 刚刚,Transformers v5 发布首个 RC(候选) 版本 v5.0.0rc0。 GitHub:https://github.com/huggingface/transformers/releases/tag/v5.0.0rc0 这次更新标志着这一全球最流行的 AI 基础设施库,正式跨越了从 v4 到 v5 长达 五年 的技术周期。 作为 Hugging Face 最核心的开源项目,自 2020 年 11 月 v4 版本发布以来,Transformers 的日下载量已从当时的 2 万次激增至如今的超过 300 万次 ,总安装量突破 12 亿次 。 它定义了业界如何使用模型,支持的架构也从最初的 40 个扩展至超过 400 个 ,涵盖了文本、视觉、音频及多模态领域,社区贡献的模型权重更是超过 75 万个 , 涵盖了文本、视觉、音频及多模态领域。 官方表示,在人工智能领域,「重塑」是保持长 ...
智源发布具身数据创新基座,携手行业共筑物理AGI基础设施
具身智能之心· 2025-12-03 03:47
Core Insights - The article discusses the launch of the RoboXstudio platform and the RoboCOIN dataset by the Beijing Zhiyuan Artificial Intelligence Research Institute, aimed at addressing challenges in embodied data production and enhancing research efficiency in embodied intelligence [6][19]. Group 1: Challenges in Embodied Data - Embodied data faces three main challenges: data silos, lack of quality control, and high costs [7][8]. - Data silos arise from non-standardized formats and isolated data tools, complicating data processing [7]. - Quality control issues include frame loss, stuttering, and timestamp misalignment, leading to unreliable data records [8]. - The cost of generating embodied data remains high due to reliance on manual operations and the absence of mature platforms for scalability [8]. Group 2: CoRobot Software Framework - The CoRobot framework was developed to standardize operations, improve quality, and enhance efficiency in embodied data management [10]. - It consists of five components: data collection tools, format conversion tools, data processing tools, data management tools, and model training tools [10]. Group 3: RoboCOIN Dataset - The RoboCOIN dataset is a collaboration involving multiple companies and universities, designed to be the global benchmark for dual-arm robot data [14][16]. - It features the largest number of dual-arm entities, with 180,000 data entries covering over ten scenarios, including industrial and retail applications [16]. - The dataset is noted for its fine-grained labeling and ease of use, facilitated by the CoRobot framework [16]. Group 4: RoboXstudio Platform - The RoboXstudio platform aims to streamline the entire process of data collection, annotation, management, model training, evaluation, and deployment [19][22]. - It supports diverse robot types and tasks, ensuring comprehensive data collection without gaps [22]. - The platform integrates open-source frameworks and multimodal models to reduce operational costs and enhance user accessibility [22]. Group 5: Open Source and Collaboration - The Zhiyuan Institute emphasizes the importance of collaborative innovation in advancing artificial intelligence, with a significant number of downloads of their open-source models [23]. - The RoboCOIN dataset and CoRobot framework are made available to the public to foster industry-wide collaboration and innovation [23][25].
免训练!使用贝叶斯去微调VLM,机器人操作任务取得SOTA!
具身智能之心· 2025-12-03 03:47
Core Insights - The article discusses the advancements in Visual Language Models (VLM) and introduces T²-VLM, a novel framework that generates temporally consistent rewards for robotic tasks without requiring training [2][5]. Group 1: VLM and T²-VLM Overview - VLMs have significantly improved performance in embodied tasks such as goal decomposition and visual understanding, but providing precise rewards for robotic operations remains challenging due to the lack of domain-specific knowledge in pre-training datasets and high computational costs [2]. - T²-VLM is designed to track the state changes of sub-goals derived from VLMs to generate accurate rewards, enhancing long-term decision-making capabilities and improving fault recovery performance through reinforcement learning [2]. Group 2: Methodology and Results - The T²-VLM method queries the VLM before each interaction to establish spatially aware sub-goals and initial completion estimates, utilizing a Bayesian tracking algorithm to dynamically update the target completion state [2]. - Extensive experiments demonstrate that T²-VLM achieves state-of-the-art performance in two robotic operation benchmarks while reducing computational costs and exhibiting superior reward accuracy [2]. Group 3: Live Session Details - A live session is scheduled for December 3rd, from 19:30 to 20:30, focusing on the background of real-machine reinforcement learning, the current state of reward generation research based on VLMs, and reflections on the T²-VLM method [5][6].
VLM也能「自我进化」!RL自我进化框架VisPlay突破视觉推理难题
具身智能之心· 2025-12-02 09:30
Core Insights - The article discusses the introduction of VisPlay, a self-evolving reinforcement learning framework for Vision-Language Models (VLM), which allows for self-improvement using vast amounts of unlabeled image data [2][3][18] Group 1: Challenges in VLM - VLMs have made significant progress in perception tasks but struggle with complex visual reasoning due to reliance on high-quality labeled data [5] - Traditional methods like supervised fine-tuning and reinforcement learning face bottlenecks as the cost and speed of manual labeling cannot keep up with the evolving model demands [5][4] Group 2: VisPlay Framework - VisPlay is designed to address the challenges of VLMs by implementing a self-evolution mechanism that allows models to learn autonomously from unlabeled images [7][8] - The framework divides the VLM into two roles: the "Questioner," which generates challenging visual questions, and the "Reasoner," which provides answers based on the images and questions [10][12] Group 3: Reward Mechanism - VisPlay employs a sophisticated reward mechanism that includes Difficulty Reward and Diversity Reward to enhance the quality of generated questions and answers [10][11] - This approach effectively mitigates common issues in self-evolving models, such as low answer quality and high question redundancy, leading to significant improvements in capability [11] Group 4: Experimental Results - VisPlay has been tested on mainstream VLM models like Qwen2.5-VL and MiMo-VL across eight benchmark datasets, showing consistent and significant accuracy gains [15][17] - The framework demonstrates strong generalization capabilities, particularly in unseen complex reasoning combinations, and effectively reduces the occurrence of "hallucinations" in VLMs [17][18]
上交&ai lab团队联合提出MM-ACT:一个统一的VLA模型实现感知-规划-执行的高效协同
具身智能之心· 2025-12-02 09:30
点击下方 卡片 ,关注" 具身智能 之心 "公众号 编辑丨 具身智能之心 本文只做学术分享,如有侵权,联系删文 >> 点击进入→ 具身智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有你想要的。 在机器人操作领域,"通用性" 与 "高效性" 的平衡始终是核心挑战——现有方案要么缺乏动态建模能力,难以应对复杂环境交互;要么推理速度慢,无法满足实时 控制需求。 上海 AI 实验室、上海交通大学等团队联合提出的 MM-ACT ,以 "统一多模态表征 + 并行解码架构" 为核心,创新引入 "上下文共享多模态学习" 范式,实现了文 本、图像、动作的协同生成,既具备精准的语义理解与环境预测能力,又能高效输出执行动作,在模拟与真实场景中均展现出超越现有方案的综合性能。 为什么需要重构视觉 - 语言 - 动作(VLA)模型架构? 当前 VLA 模型陷入 "三重矛盾":语义理解与动态建模难以兼顾、多模态生成效率低下、训练目标存在错位,核心问题可归结为 "无法在统一框架内实现'感知 - 规 划 - 执行'的高效协同": | 方案类型 | 代表思路 | | 核 ...
清华成立具身智能与机器人研究院
具身智能之心· 2025-12-02 09:30
编辑丨 量子位 点击下方 卡片 ,关注" 具身智能之心 "公众号 >> 点击进入→ 具身 智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区: 具身智能之心知识星球(戳我) ,这里包含所有你想要的! 具身智能,火得有些过分, 就在昨天,清华大学宣布成立 具身智能与机器人 研究院。 这是今年3月底,清华开设 具身智能系统北京市重点实验室 以来,围绕具身智能的又一次大动作。 这并非孤例——清华的布局,恰好踩中了国内高校整体加快具身智能布局的节点上。 这一年,国内高校几乎清一色地围着 "具身智能" 出牌: 从研究院、实验室、研究中心到本科专业;从密集的demo演示,到一场场揭牌仪式、行业对谈、学术论坛与专题会议…… 毫不夸张地说,具身智能硬是用一年时间,走完了大模型三年的发展路程。 清华成立具身智能与机器人研究院 11月30日,清华大学 具身智能与机器人研究院 正式揭牌成立。 清华大学校长 李路明 院士等校领导出席揭牌仪式,新晋中国科学院院士、清华大学 刘云浩 教授也在现场发表致辞。 据悉,新研究院由清华大学自动化系主任、信息科学技术学院副院长 张涛 教授出任院长。 智能技术与系统国家重点实验室副 ...
担心买得起机械臂,玩不转代码?小白友好,你的第一台科研机械臂
具身智能之心· 2025-12-02 09:30
Core Viewpoint - The article emphasizes the advantages of the Imeta-Y1 robotic arm, highlighting its user-friendly features and cost-effectiveness for beginners and researchers in the field of embodied intelligence [5][6][9]. Group 1: Workflow Improvement - Before using Imeta-Y1, users spend 70% of their time on hardware communication and sensor calibration, facing challenges in code adaptation between simulation and real machines [1]. - After adopting Imeta-Y1, the workflow transforms to quick simulation verification in Gazebo, allowing for seamless deployment and fine-tuning of algorithms [2]. Group 2: Product Features - Imeta-Y1 is designed as a lightweight, high-cost performance robotic arm, suitable for students, educators, and novice developers [5][6]. - It offers a complete open-source toolchain and code examples, supporting both Python and C++ interfaces, ensuring quick onboarding for users [7][21]. - The arm is compatible with ROS1 and ROS2, providing URDF models for seamless switching between simulation and real machines [7][20]. Group 3: Technical Specifications - The robotic arm has a weight of 4.2 kg, a rated load of 3 kg, and 6 degrees of freedom, with a working radius of 612.5 mm and a repeat positioning accuracy of ±0.1 mm [11][22]. - It operates on a 24V power supply and utilizes CAN communication, with a compact design suitable for embedded AI and robotic learning platforms [10][22]. Group 4: Development Support - The product includes a comprehensive open-source SDK, facilitating rapid application development with support for major frameworks like TensorFlow and PyTorch [33][39]. - Users can perform end-to-end algorithm deployment, from data collection to model training and inference, significantly reducing development risks and debugging costs [20][39]. Group 5: Customer Support - The company provides 24-hour rapid response for customer support, ensuring users do not face delays in their learning and development processes [7][22]. - There are bulk purchase discounts available, and the company supports project development and educational training based on the product [22].
竟速机器人“母港”! 2026具身智能首展,3月杭州集结!
具身智能之心· 2025-12-02 03:03
Core Insights - The core viewpoint of the article emphasizes the rapid growth and potential of the embodied intelligence market in China, predicting a market size of nearly 5.3 billion yuan by 2025 and over 20 billion yuan by 2026, with a global market projection of 870 billion yuan by 2030 [3][5]. Market Potential - The embodied intelligence market in China is expected to approach 5.3 billion yuan by 2025, accounting for over 25% of the global market share [3]. - By 2026, the market is projected to exceed 20 billion yuan, indicating a significant growth trajectory [3]. Industry Ecosystem - Hangzhou is identified as a "global innovation mother port," housing over 700 key enterprises across the entire supply chain from research and development to manufacturing [4][6]. - The city benefits from a complete industrial ecosystem, innovative resource aggregation, and proactive policy legislation, positioning it as a leader in the embodied intelligence sector [6]. Upcoming Events - The 2026 Third China Embodied Intelligent Robot Industry Conference and Exhibition will take place from March 11-13, 2026, at the Hangzhou International Expo Center, featuring over 500 exhibitors and 30,000 professional attendees [8][9]. - The event aims to create a comprehensive industrial ecosystem that integrates conferences, exhibitions, technology, and trends [8][14]. Exhibition Focus - The exhibition will cover a wide range of products, including complete embodied intelligent robots, power systems, industrial robots, control and computing systems, and various application solutions [16][20]. - It aims to connect all segments of the supply chain, facilitating high-quality supply and application of innovative technologies [14]. Awards and Recognition - The 2026 China Embodied Intelligence Industry Annual Awards Ceremony will recognize significant achievements and benchmark forces in the sector, with awards for categories such as "Top Ten Outstanding Complete Machine Brands" and "Top Ten Innovative Enterprises" [38][46]. - The awards will serve as a catalyst for industry development and a gathering point for key stakeholders [46]. Industry Collaboration - The conference will feature discussions on technological breakthroughs and practical applications across various industries, fostering collaboration between academia, industry leaders, and investors [29][52]. - It aims to provide a comprehensive understanding of the future landscape of the embodied intelligence sector, enabling participants to seize opportunities in a changing environment [29].
IPO辅导收官!A股首个人形机器人正式开启冲刺
具身智能之心· 2025-12-02 03:03
点击下方 卡片 ,关注" 具身智能 之心 "公众号 自2020年起,宇树科技已步入盈利轨道,未出现持续亏损或业绩剧烈波动等影响财务健康的突出问题。其研 发支出的资本化处理及收入确认等关键会计操作均严格遵循相关会计准则,财务数据真实准确,无需进行额 外的追溯调整。 商业模式明确聚焦: 公司核心业务围绕四足机器人展开,已形成清晰的客户群体划分——涵盖B端工业应用与C端消费市场,并建 立了稳定且可持续的收入流。同时,人形机器人业务正处于积极的研发推进及小批量试产阶段。整体而言, 宇树的业务布局集中,边界清晰,不存在盲目跨界经营或业务多元化导致的资源分散问题。 写在最后 编辑丨具身智能之心 本文只做学术分享,如有侵权,联系删文 >> 点击进入→ 具身智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有你想要 的。 2025年11月29日,宇树向浙江证监局提交了更新后的IPO辅导进展报告,其辅导状态正式转为"辅导工作完 成"。这标志着宇树科技已成功通过中国证监会关于A股上市的前期合规审查,即将迈出提交招股说明书的关 键一步,有望成为"A股人形机器人第 ...