Workflow
具身智能之心
icon
Search documents
合伙人招募!和我们一起运营这个具身社区吧~
具身智能之心· 2025-09-21 10:00
Core Viewpoint - The article emphasizes the importance of collaboration in the field of embodied intelligence, aiming to create a platform that adds real value to the industry rather than just serving as a media outlet [1]. Group 1: Course Development - The company invites collaboration to develop courses that benefit beginners and promote industry advancement, targeting both consumer and enterprise training as well as academic curriculum development [2][3]. Group 2: Hardware Development - The goal is to create an affordable and user-friendly research platform for embodied intelligence, ensuring accessibility for developers and ease of use for beginners [4]. Group 3: Open Source Projects - The company seeks to build globally influential open source projects in collaboration with others in the field [5][6]. Group 4: Consulting Services - There is an invitation to partner in providing consulting services for both B2B and B2C sectors, focusing on embodied data, ontology, algorithms, and deployment to facilitate industry upgrades and talent development [7][8]. Group 5: Job Opportunities - The company is looking for individuals with engineering experience in the field or those holding a PhD or higher, offering competitive compensation and access to industry resources for both full-time and part-time positions [9][10].
灵御智能遥操TeleAvatar机器人开始交付啦!
具身智能之心· 2025-09-21 04:01
Core Insights - Lingyu Intelligent has achieved a significant milestone in the field of artificial intelligence and robotics with the delivery of its first TeleAvatar robot, marking a breakthrough in product commercialization and market expansion [2][3]. Group 1: Product Delivery - The first TeleAvatar robot (model 001) was officially delivered to the Xiganghu Robotics Institute, with multiple strategic clients set to receive their first batch of robots in the coming weeks [2][3]. - The delivery event was attended by key representatives, including Dr. Xu Qinhua, the technical director of the Xiganghu Robotics Institute, highlighting the importance of this collaboration [3]. Group 2: Technological Innovation - TeleAvatar is an embodiment of advanced technology, featuring high-precision motion control, multi-modal perception integration, and low-latency remote operation capabilities [5][6]. - The robot boasts an operating precision at the sub-millimeter level, with an end-to-end operation delay of less than 30 milliseconds, ensuring real-time responsiveness [6]. Group 3: Market Potential - The application areas for TeleAvatar are extensive, including scientific data collection, smart manufacturing, medical services, research exploration, and emergency response, providing robust technical support for industry transformation [7]. - The competitive pricing of TeleAvatar, starting at 79,900 yuan, positions it as a cost-effective solution in the market [6]. Group 4: Company Background - Lingyu Intelligent was founded by a top team from Tsinghua University's Department of Automation, focusing on robot planning control and human-machine interaction [10]. - The company's mission is to create practical benchmarks for embodied intelligence, liberating humans from dangerous, heavy, and tedious tasks through a comprehensive self-research solution that spans hardware, software, and data platforms [10].
具身领域的大模型基础部分,都在这里了......
具身智能之心· 2025-09-20 16:03
Core Viewpoint - The article emphasizes the importance of a comprehensive community for learning and sharing knowledge about large models, particularly in the fields of embodied AI and autonomous driving, highlighting the establishment of the "Large Model Heart Tech Knowledge Planet" as a platform for collaboration and technical exchange [1][3]. Group 1: Community and Learning Resources - The "Large Model Heart Tech" community aims to provide a platform for technical exchange related to large models, inviting experts from renowned universities and leading companies in the field [3][67]. - The community offers a detailed learning roadmap for various aspects of large models, including RAG, AI Agents, and multimodal models, making it suitable for beginners and advanced learners [4][43]. - Members can access a wealth of resources, including academic progress, industrial applications, job recommendations, and networking opportunities with industry leaders [7][70]. Group 2: Technical Roadmaps - The community has outlined specific learning paths for RAG, AI Agents, and multimodal large models, detailing subfields and applications to facilitate systematic learning [9][43]. - For RAG, the community provides resources on various subfields such as Graph RAG, Knowledge-Oriented RAG, and applications in AIGC [10][23]. - The AI Agent section includes comprehensive introductions, evaluations, and advancements in areas like multi-agent systems and self-evolving agents [25][39]. Group 3: Future Plans and Engagement - The community plans to host live sessions with industry experts, allowing members to engage with leading figures in academia and industry [66]. - There is a focus on job sharing and recruitment information to empower members in their career pursuits within the large model domain [70].
PhysicalAgent:迈向通用认知机器人的基础世界模型框架
具身智能之心· 2025-09-20 16:03
Core Viewpoint - The article discusses the development of a new robotic control framework called PhysicalAgent, which aims to overcome existing limitations in the field of robot manipulation by integrating iterative reasoning, diffusion video generation, and closed-loop execution [2][4]. Group 1: Key Challenges in Robotics - Current mainstream visual-language-action (VLM) models require task-specific fine-tuning, leading to a significant drop in robustness when switching robots or environments [2]. - World model-based methods depend on specially trained predictive models and carefully curated training data, limiting their generalizability [2]. Group 2: Framework Design and Principles - The PhysicalAgent framework separates perception and reasoning from specific robot forms, requiring only lightweight skeletal detection models for different robots, which minimizes computational costs and data requirements [4]. - The framework leverages pre-trained video generation models that understand physical processes and object interactions, allowing for quick integration without local training [4]. - It aligns human-like reasoning by generating visual representations of actions based on textual instructions, facilitating intuitive robot control [4]. Group 3: VLM's Grounding Reasoning Role - The VLM serves as the cognitive core of the framework, enabling grounding through multiple calls to achieve "instruction-environment-execution" rather than a single planning step [6]. - The framework innovatively reconstructs action generation as conditional video synthesis, moving away from traditional direct control strategy learning [6]. Group 4: Execution Process and Adaptation - The robot adaptation layer translates generated action videos into motor commands, which is the only part requiring robot-specific adaptation [6]. - The process includes task decomposition, contextual scene description, execution monitoring, and model independence, allowing for flexibility in model selection [6]. Group 5: Experimental Validation - Experiments validate the framework's cross-form and perception modality generalization, as well as the robustness of iterative execution [8]. - The first experiment demonstrated that the framework significantly outperformed task-specific baselines in success rates across different robotic platforms [12]. - The second experiment confirmed the robustness of the iterative "Perceive→Plan→Reason→Act" pipeline, achieving an 80% success rate across physical robots [13].
头部具身智能人形机器人公司最新估值/市值
具身智能之心· 2025-09-20 06:12
编辑丨具身智能之心 点击下方 卡片 ,关注" 具身智能之心 "公众号 >> 点击进入→ 具身 智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有 你想要的。 头部具身智能人形机器人公司最新估值或市值一览。除了已上市公司外,这里展示的都是已完成或 正在交割的真实估值,未经实际交割、未获交易确认的估值均未列入,单位为人民币。注意,各公 司成立时间和融资阶段差异大。估值高低与技术、商业化水平不能简单划等号。 以下数字仅做参考,如有不足或者遗漏,欢迎后台留言。 Figure AI 2736亿 优必选 555亿 Sklid AI 324亿 Physical Intelligence 170亿 宇树科技 160亿 智元机器人 150亿 Apptronik 144亿 Field AI 144亿 Agility Robotics 126亿 云深处机器人 80亿 傅利叶机器人 80亿 乐聚机器人 80亿 World labs 70亿 Sanctuary AI 70亿 Boston Dynamics 70亿 银河通用 70亿 星海图 70亿 自变量 60亿 ...
英伟达50亿美元入股英特尔,将发布CPU+GPU合体芯片,大结局来了?
具身智能之心· 2025-09-19 16:04
编辑丨 机器之心 点击下方 卡片 ,关注" 具身智能之心 "公众号 >> 点击进入→ 具身 智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有你想要的。 他们共同宣布,要把电脑上的 CPU 和 GPU 合成为超级 SoC。 周四晚间,英伟达收购 50 亿美元英特尔股份的新闻引爆了科技圈。 两家公司在 9 月 18 日同时发布公告,宣布达成长期战略合作。英伟达将投资 50 亿美元购买英特尔普通股,基于全新合作,两家公司将共同开发多代定制 数据中心和 PC 产品。 在具体内容上,两家公司将专注于利用 NVIDIA NVLink 无缝连接 NVIDIA 和 Intel 架构 —— 将英伟达的 AI 和加速计算优势与英特尔领先的 CPU、x86 生态系统相结合,为客户提供顶尖解决方案。 对于数据中心,英特尔将构建英伟达定制版 x86 CPU,英伟达会将其集成到其 AI 基础设施平台中并提供给市场。 在个人计算领域,英特尔将打造并向市场推出集成 RTX GPU 芯片组的 x86 系统级芯片 (SoC)。这些全新的 x86 RTX SoC 将为各种需 ...
从数采方案来看,具身数据的需求有哪些?
具身智能之心· 2025-09-19 16:04
点击下方 卡片 ,关注" 具身智能 之心 "公众号 编辑丨 具身智能之心 本文只做学术分享,如有侵权,联系删文 >> 点击进入→ 具身智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有你想要的。 当前,具身智能已成为全球的新焦点,如何打造一个通用的本体和大脑是各个创业公司一直努力突破的,更是受到资本和产业界的高度关注。而数采作为基础模 块,是重中之重,好的数据更是很多算法取得效果的基础。 今天就为大家全面梳理下具备研发和产品力的数采领域相关公司,深入分析其技术特点、产品布局和应用场景,为公司提供行业全景图,助力战略决策和业务拓 展。 重点关注 :专注于数据采集设备与解决方案的企业,包括硬件采集设备、软件平台及整体解决方案。 国内公司 星海图 自研数采任务管理平台 :支持任务发布、上传、存储、审核等全流程 可视化管控,高效管理,无惧数据丢失! 一站式数据采集链路 :任务下发 → 采集 → 清洗 → 补采 → 压缩 → 上传 → 审核 → 标注 → 存储 兼容主流算法与格式 :输出格式:rosbag、ARIO、lerobot;适配模型:ACT ...
智源牵头举办的具身大模型挑战赛火热报名中!
具身智能之心· 2025-09-19 16:04
编辑丨 BAAI具身智能 点击下方 卡片 ,关注" 具身智能之心 "公众号 2025 第二届中关村具身智能机器人应用大赛 了解更多信息 欢迎大家踊跃报名参赛! 智源具身智能模型能力挑战赛火热报名中! 本届赛事以 「具身引智 · 应用未来」 为主题,打造一个 汇聚尖端技术与产业应用 的舞台。这里不仅是比拼模型实力的竞技场,更是展示创意与才华的舞台。让我们一起突破边界,提升模型能力,推动具身智能 走 出实验室,走进现实世界,创造真正的价值! 未来已来,等你出发! 指导教师荣誉:有机会获得"智源学者"身份,享受专项科研资金支持 10.23 - 10.24 11.02 - 11.16 11.17 - 11.18 决赛 初赛 真机调试与数据采集 资源支持 真机数据采集、标注一站式平台支持 充足的算力支持 机器人本体设备支持 智源专家全程技术指导 场地与环境保障 奖金与荣誉 单暴道奖金 优胜奖(第4-6名) 三等奖 2万 等奖 5万 二等奖 3万 学生选手福利:有机会获得直通智源研究院实习、入职机会 在智源,你将收获: 真机实战:人形机器人、高性能机械臂、移动操作平台等 顶级算力 & 自由科研:享用智源充足的算力与海量数 ...
NeurIPS 2025 | 人类认知对齐的CogVLA,突破VLA效率与性能瓶颈
具身智能之心· 2025-09-19 05:43
Core Insights - The article discusses the development of a new model called CogVLA, which addresses the efficiency challenges and semantic degradation in Vision-Language-Action (VLA) research, driven by the capabilities of pre-trained Vision-Language Models (VLM) [5][6][10]. Group 1: Background and Challenges - The transition from large models to embodied intelligence faces efficiency dilemmas and semantic degradation, with existing VLA methods often neglecting the semantic coupling between perception, language alignment, and action decoding [5]. - Key challenges include redundant perception, instruction-semantic disconnection, and action incoherence, which hinder the performance of traditional VLA models [6][10]. Group 2: Proposed Solution - CogVLA introduces a cognitive-aligned three-stage design that mimics human multimodal coordination mechanisms, consisting of EFA-Routing, LFP-Routing, and CAtten [12][14]. - EFA-Routing focuses on instruction-driven visual aggregation, LFP-Routing performs semantic pruning in language models, and CAtten ensures semantic consistency and action sequence coherence [16]. Group 3: Experimental Results - CogVLA outperforms advanced models like OpenVLA-OFT and π0, achieving a state-of-the-art (SOTA) success rate of 97.4% on LIBERO while maintaining an 8× visual compression ratio [18]. - The model significantly improves efficiency, with inference time reduced by 2.79 times, throughput increased by 22.54 times, and training costs lowered by 2.49 times compared to OpenVLA [20]. Group 4: Visualization and Performance - Visual analysis demonstrates CogVLA's ability to focus on task-relevant areas in input images, showcasing its human-aligned perception capabilities even in chaotic or unclear scenes [21].
智平方2026年大规模校园招聘来袭!具身算法/开发/仿真等
具身智能之心· 2025-09-19 00:03
Core Insights - The article highlights the advancements in AI and robotics, particularly focusing on the development of the world's first end-to-end VLA technology and the launch of the GOVLA model, which outperforms international benchmarks by 30% [2] - The company is recognized as the only domestic entity to open-source a robot model, showcasing its strong technical capabilities [2] - The introduction of the AlphaBot series demonstrates the company's commitment to creating versatile robots capable of seamless task switching across various scenarios [2] Technology and Innovation - The company has developed a unified technology platform that supports sustainable and compounding intelligent services across multiple real-world applications [2] - It has established partnerships with leading clients in sectors such as automotive manufacturing, semiconductors, biotechnology, and public services [2] Talent and Culture - The company promotes a flat organizational structure that encourages input from all levels, including fresh graduates [5] - It seeks individuals with a strong curiosity, learning ability, and practical skills, emphasizing the importance of collaboration and resilience in facing real-world uncertainties [12] Recruitment and Opportunities - The company is actively hiring for various positions across algorithms, engineering, product management, design, and manufacturing [8][9][10][11] - It offers a clear recruitment process, including online applications, resume screening, and interviews, aimed at attracting top talent from diverse educational backgrounds [13][14]