Workflow
世界模型
icon
Search documents
达晨财智领投 极佳视界完成2亿元A2轮融资
Xin Lang Cai Jing· 2025-12-08 15:14
Investment Overview - The company Jijiashijie has recently completed a new round of financing, raising 200 million yuan in Series A2 funding, led by Dacheng Caizhi, with participation from several notable institutions [1][3] - This round of financing follows three previous rounds (Pre-A, Pre-A+, A1) completed within three months, totaling 500 million yuan in Series A funding [1][3] Company Focus and Products - Jijiashijie specializes in general intelligence for the physical world, aiming for physical AGI (Artificial General Intelligence) and has plans to release a corresponding ontology by November 26, 2025 [1][3] - The company's product offerings include the GigaWorld platform (for driving and embodiment), GigaBrain (general embodied brain), and Maker (general embodied ontology), representing a full-stack approach to physical AI [1][3] Model Development - The company has introduced a native paradigm of "world model + action model + reinforcement learning," where each component is driven by the world model [1][3] - The current trend in model architecture is converging towards general action models, with a shift in data sources to real machine data and world model-generated data [2][4] Industry Trends - The company believes that physical AI is entering a new critical era, with the next 2-3 years being a key window for breakthroughs in physical AGI [5] - The advancements in world models and action models are accelerating the arrival of a "ChatGPT moment" in the physical world [5]
Roblox CEO感叹AI研究进展:曾博览群书的自己都快看不懂了
Sou Hu Cai Jing· 2025-12-08 11:28
巴祖基 2005 年创立 Roblox。创业初期,他几乎读遍了从物理模拟到图形渲染的各类研究,而且都能理解。然而 AI 时代的到来改变了一切。他称如今的研究浪潮"规模巨大、速度惊人",从 Transformer 到扩散模型,再到世界 模型,"内容多到让人难以完全掌握"。 IT之家 12 月 8 日消息,AI 研究更新速度飞快,新论文几乎每天出现,技术概念也越来越复杂,Roblox CEO 大 卫・巴祖基对此深有体会。 据《商业内幕》今日报道,巴祖基透露,自己休假时抽出大量时间系统阅读 AI 研究,却发现过程"发人深省"—— 想真正看懂所有论文"极其困难"。 尽管外界关注焦点集中在算力扩张,OpenAI 联合创始人伊利亚・苏茨克维却认为,真正决定 AI 走向的仍是"研 究本身":"我们重新回到研究时代,只不过现在用的是更大的计算机。" 而对于 Roblox 而言,巴祖基的结论是:AI 在"三维世界"里仍然处于非常初期的阶段。他指出,AI 依赖的是人类 制造出来的文本和图像,"我们在用自己创造的内容训练 AI,而不是用真实世界的三维原始数据"。 随着 AI 从学界扩展到国家战略高度,Meta、微软等公司纷纷建立自 ...
达晨、华控领投,极佳视界A2轮再融2亿,押注“世界模型+行动模型”原生架构
Tai Mei Ti A P P· 2025-12-08 07:17
Group 1 - The company, Jiga Vision, has completed a new round of financing, raising 200 million yuan in Series A2 funding, led by Dashen Caizhi, with participation from several notable investors, bringing the total funding raised in the last three months to 500 million yuan [2] - The founder and CEO, Dr. Huang Guan, has a strong background in AI and robotics, having previously worked at leading research institutions and has been instrumental in the evolution of physical AI from its inception to industrial application [2][3] - Jiga Vision has introduced a new paradigm for artificial general intelligence (AGI) that emphasizes a "world model + action model + reinforcement learning" framework, indicating a shift towards general action models in the industry [3] Group 2 - The company has officially launched two core models for physical AGI: GigaBrain-0, an end-to-end decision control model, and GigaWorld-0, a high-quality world model, along with the Maker H01 robot platform [4] - GigaBrain-0 enhances 3D spatial perception and structured reasoning capabilities, significantly improving navigation accuracy and task execution in complex environments, outperforming current state-of-the-art methods in various benchmarks [5] - GigaWorld-0 generates high-fidelity, controllable, and diverse interactive data, achieving nearly 300% performance improvement in key generalization dimensions, making it a cost-effective solution in the current market [6] Group 3 - Maker H01 is designed for open environments in home, commercial, and light industrial applications, featuring a dual-arm and omnidirectional mobile chassis, capable of performing precise operations and complex tasks [6][7] - The integration of GigaBrain-0, GigaWorld-0, and Maker H01 accelerates the transition of embodied intelligence from the laboratory to scalable applications, marking a significant step towards a reliable and generalizable physical AGI era [7]
哈萨比斯:DeepMind才是Scaling Law发现者,现在也没看到瓶颈
量子位· 2025-12-08 06:07
Core Insights - The article emphasizes the importance of Scaling Laws in achieving Artificial General Intelligence (AGI) and highlights Google's success with its Gemini 3 model as a validation of this approach [5][19][21]. Group 1: Scaling Laws and AGI - Scaling Laws were initially discovered by DeepMind, not OpenAI, and have been pivotal in guiding research directions in AI [12][14][18]. - Google DeepMind believes that Scaling Laws are essential for the development of AGI, suggesting that significant data and computational resources are necessary for achieving human-like intelligence [23][24]. - The potential for Scaling Laws to remain relevant for the next 500 years is debated, with some experts expressing skepticism about its long-term viability [10][11]. Group 2: Future AI Developments - In the next 12 months, AI is expected to advance significantly, particularly in areas such as complete multimodal integration, which allows seamless processing of various data types [27][28][30]. - Breakthroughs in visual intelligence are anticipated, exemplified by Google's Nano Banana Pro, which demonstrates advanced visual understanding [31][32]. - The proliferation of world models is a key focus, with notable projects like Genie 3 enabling interactive video generation [35][36]. - Improvements in the reliability of agent systems are expected, with agents becoming more capable of completing assigned tasks [38][39]. Group 3: Gemini 3 and Its Capabilities - Gemini 3 aims to be a universal assistant, showcasing personalized depth in responses and the ability to generate commercial-grade games quickly [41][44][45]. - The architecture of Gemini 3 allows it to understand high-level instructions and produce detailed outputs, indicating a significant leap in intelligence and practicality [46]. - The frequency of Gemini's use is projected to become as common as smartphone usage, integrating seamlessly into daily life [47].
死磕技术的自动驾驶黄埔军校,又更新了这些技术进展......
自动驾驶之心· 2025-12-07 02:05
点击下方 卡片 ,关注" 自动驾驶之心 "公众号 戳我-> 领取 自动驾驶近30个 方向 学习 路线 这一个月,自动驾驶之心星球又更新了很多技术内容,汇报给大家: 自动驾驶之心知识星球是我们一直在维护更新的星球! 如果您也想和自动驾驶学术界或工业界的大佬交流,欢迎加入自动驾驶之心知识星球。我们聊技术、聊趋 势、聊变化。未来柱哥还会持续邀请学术界和工业界的同行和大家交流。 欢迎加入自动驾驶之心知识星球,我们准备的大额的新人优惠...... 扛内卷,一个足够有料的社区 对于很多想入门的同学来说,试错成本有点高。没时间和缺乏完整的体系是最大问题,这也容易导致行业壁垒越来越高,如果想要卷赢那就更加困难了。 所以我们联合了诸多学术界和工业界的大佬,共同打造了我们维护三年之久的『自动驾驶之心知识星球』! 星球目前集视频 + 图文 + 学习路线 + 问答 + 求职交流 为一体,是一个综合类的自驾社区,已经超过4000人了。我们期望未来2年内做到近万人的规模。给大家打造一个交流+技术分享的聚集地,是许多初学者和进阶的 同学经常逛的地方。 如果你也想和我们一起推动自驾领域的进步,欢迎加入我们的社区团队,和我们一起推动! 我们准 ...
烧光700亿后,扎克伯格戳破元宇宙泡沫
Xin Lang Cai Jing· 2025-12-06 06:24
Core Viewpoint - Meta is significantly reducing its budget for the metaverse division by 30%, marking a shift away from its previous focus on the metaverse towards more profitable areas, particularly AI [1][4][20]. Group 1: Budget Cuts and Personnel Adjustments - Meta is considering a budget cut of up to 30% for its metaverse department, primarily affecting the Meta Horizon Worlds social platform and Quest VR hardware [4][19]. - Since early 2021, Reality Labs has incurred losses exceeding $70 billion, prompting this budget reduction as a direct response to Wall Street's pressure [5][20]. - The company may initiate layoffs affecting 10% to 30% of employees in the metaverse division as early as January 2026 [7][22]. Group 2: Talent Acquisition and Resource Allocation - Despite cutting VR budgets, Meta is investing in high-level talent by hiring former Apple design executive Alan Dye as Chief Design Officer for Reality Labs [23][24]. - Dye will lead a new creative studio focused on integrating design, fashion, and technology for next-generation AI products, emphasizing the strategic importance of design aesthetics [25][32]. - The funds saved from the budget cuts are expected to be redirected towards AI glasses and other wearable technology projects, as competitors slow their efforts in virtual reality [22][32]. Group 3: Divergence in Technical Direction - Meta's Chief AI Scientist Yann LeCun is leaving to establish a new AI company, AMI, which will focus on "world models" rather than the mainstream generative AI approach [27][29]. - LeCun criticizes the prevailing generative AI models, arguing they lack essential elements for achieving human-level intelligence [28][29]. - AMI will operate independently from Meta, although it will maintain a collaborative relationship without accepting Meta's investment to ensure research independence [30][31]. Group 4: Strategic Balance and Future Challenges - Meta's recent actions reflect a pragmatic strategy, reallocating resources from the metaverse to AI infrastructure and hardware that show market potential [32][33]. - The introduction of top design talent aims to ensure that future AI hardware excels in user interaction aesthetics [32]. - By allowing independent exploration in foundational AI theories, Meta is diversifying its technological bets, mitigating risks associated with focusing solely on generative models [32][33].
英伟达2025年技术图鉴,强的可怕......
自动驾驶之心· 2025-12-06 03:04
Core Viewpoint - NVIDIA has emerged as a leading player in the AI infrastructure space, achieving a market valuation of $5 trillion, which is an 11-fold increase over three years. The company has transitioned from a graphics chip manufacturer to a key player in AI, particularly in autonomous driving and embodied intelligence [2]. Group 1: NVIDIA's Technological Developments - The Cosmos series, initiated in January, focuses on world foundation models, leading to the development of Cosmos-Transfer1, Cosmos-Reason1, and Cosmos-Predict2.5, which lay the groundwork for autonomous driving and embodied intelligence [5]. - The Nemotron series aims to create a "digital brain" for the agent-based AI era, providing open, efficient, and precise models and tools for enterprises to build specialized AI systems [5]. - The embodied intelligence initiatives include GR00T N1 and Isaac Lab, which focus on simulation platforms and embodied VLA (Vision-Language-Action) models [5]. Group 2: Key Papers and Contributions - The paper "Isaac Lab" presents a GPU-accelerated simulation framework for multi-modal robot learning, addressing challenges in data scarcity and the simulation-to-reality gap [6]. - "Nemotron Nano V2 VL" introduces a 12 billion parameter visual language model that achieves state-of-the-art performance in document understanding and long video reasoning tasks [12]. - "Alpamayo-R1" proposes a visual-language-action model that integrates causal reasoning and trajectory planning to enhance safety and decision-making in autonomous driving [13]. Group 3: Innovations in AI Models - "Cosmos-Predict2.5" introduces a next-generation physical AI video world foundation model that integrates text, image, and video generation capabilities, significantly improving video quality and consistency [17]. - "Cosmos-Reason1" aims to endow multi-modal language models with physical common sense and embodied reasoning capabilities, enhancing their interaction with the physical world [32]. - "GR00T N1" is an open foundation model for generalist humanoid robots, utilizing a dual-system architecture for efficient visual language understanding and real-time action generation [35].
对话任少卿:2025 NeurIPS 时间检验奖背后,我的学术与产业观
雷峰网· 2025-12-05 10:24
Group 1 - NeurIPS is recognized as the "Oscar of AI" and serves as a global annual barometer for the artificial intelligence field [1] - The NeurIPS Time-Tested Award honors foundational works that have significantly influenced the discipline over a decade [1] - The award was given to the authors of "Faster R-CNN," which has been cited over 98,000 times, making it the most cited paper by a Chinese first author at this conference [2] Group 2 - "Faster R-CNN," developed in 2015, improved object detection efficiency by over 10 times and introduced an end-to-end real-time detection model [2] - The core ideas of this model have been deeply integrated into the foundational technologies of AI, impacting key sectors such as autonomous driving and medical imaging [2] - The collaboration between the authors, including Ren Shaoqing and He Kaiming, has led to significant advancements in deep learning frameworks [2] Group 3 - Ren Shaoqing joined NIO in August 2020, focusing on building a team and developing self-research chips for autonomous driving [13][14] - NIO's first generation of vehicles utilized the Mobileye solution, while the second generation was the first globally to mass-produce the NVIDIA Orin chip [14] - The challenges faced during the development included adapting to new architectures and ensuring the stability of the new chip [15] Group 4 - NIO emphasized the importance of data collection and analysis, focusing on corner cases to improve the performance of their models [19][20] - The company established a flexible system for cloud computing and data management, allowing for rapid iteration of models [21] - NIO's approach to active safety has enabled them to achieve a standard of 200,000 kilometers per false positive, significantly improving their testing efficiency [22] Group 5 - The concept of end-to-end solutions in autonomous driving has evolved, with discussions on integrating various technologies to enhance performance [24][25] - NIO is exploring the development of world models to improve long-term decision-making capabilities in autonomous systems [27][28] - The world model approach aims to address the limitations of traditional methods by incorporating both spatial and temporal understanding [30][31]
大晓机器人将于12月18日正式对外亮相,商汤王晓刚出任董事长
Xin Lang Cai Jing· 2025-12-05 00:42
Core Viewpoint - DaXiao Robotics is set to officially unveil its open-source "Kairos 3.0" world model on December 18, marking it as the first domestically open-source world model with commercial applications [1] Group 1: Product Launch - The Kairos 3.0 model is the first domestic open-source world model that has achieved commercial application [1] - Alongside the Kairos 3.0, DaXiao Robotics will release the A1 embodiment super brain module, which features a pure visual end-to-end VLA embodiment intelligence model with autonomous navigation capabilities [1] Group 2: Leadership Changes - Wang Xiaogang, co-founder and executive director of SenseTime, will serve as the chairman of DaXiao Robotics [1] - World-class AI scientist Tao Dacheng has been appointed as the chief scientist of DaXiao Robotics [1]
另辟蹊径赴欧洲创办新AI公司,杨立昆:硅谷不是AGI的土壤
3 6 Ke· 2025-12-05 00:04
Core Insights - Yann LeCun, the outgoing Chief AI Scientist at Meta, plans to establish a new startup in Europe that will pursue a different AI path compared to the generative models dominated by tech giants like OpenAI and Google [1][2] - The new company, named Advanced Machine Intelligence (AMI), aims to develop systems that understand the physical world rather than just generating text, with a focus on creating a significant revolution in AI capabilities [2][3] Group 1 - Yann LeCun announced his departure from Meta to focus on creating his own company, emphasizing the need for AI development outside of Silicon Valley [1][2] - The startup will be a "global entity" with multiple research bases worldwide, particularly in Europe, to harness local talent [2] - LeCun criticized current text-based language models for lacking essential capabilities that would allow AI to perform tasks comparable to a five-year-old child [2] Group 2 - The goal of AMI is to enable systems to understand the physical world, possess long-term memory, reason, and plan complex actions [2] - The new company will adopt a "non-generative" AI architecture to perceive environments and understand the physical world, opening up new application possibilities [2] - Meta will collaborate with AMI and provide access to its innovative technologies, but will not invest in the startup [2][3]