宇树开源UnifoLM-VLA-0
Xin Lang Cai Jing·2026-01-29 12:37
Core Viewpoint - Yuzhu announced the launch of UnifoLM-VLA-0, a visual-language-action (VLA) large model aimed at general humanoid robot operations, which seeks to overcome the limitations of traditional visual-language models (VLM) in physical interactions [1] Group 1 - UnifoLM-VLA-0 is part of the UnifoLM series and focuses on enhancing robot operation capabilities through continued pre-training on robot operation data [1] - The model represents an evolution from general "image-text understanding" to a "embodied brain" with physical common sense [1]