Workflow
超级对齐
icon
Search documents
389万寻找翁荔继任者,OpenAI紧急开招安全防范负责人
3 6 Ke· 2025-12-29 09:42
百万年薪急招一名高管! 在一连接到多起安全指控后,OpenAI终于坐不住了。 于是在最近,这家公司豪掷55.5万美元(约合人民币389万元)+股权,原地开招一名安全防范负责人(Head of Preparedness)—— 其核心职责是,制定并执行OpenAI的安全防范框架。 而且CEO奥特曼还特意强调: 这将是一份压力很大的工作,你几乎会立即面临严峻的挑战。 而且有一说一,OpenAI的安全团队似乎一直命途多舛,印象中光是负责人就换了一茬又一茬—— Ilya领导的超级对齐团队一度解散、北大校友翁荔也曾短暂担任过Preparedness团队负责人…… 直到现在,OpenAI又想起了它的安全团队。 所以,到底发生了什么让OpenAI又开始把目光转向安全了? 一切还要从彭博社最近提到的一起安全事件说起—— ChatGPT被指间接导致一位青少年离世 据彭博社消息,有一对夫妇最近指控ChatGPT间接导致了其儿子自杀。 其儿子从去年秋天开始使用ChatGPT,一开始确实是在讨论作业,但随着交流愈加深入,其对话中开始多次提到"suicide"等敏感词汇。 经他们私下统计,从去年12月到今年4月,虽然ChatGPT主动发 ...
Ilya闹翻,奥特曼400万年薪急招「末日主管」,上岗即「地狱模式」
3 6 Ke· 2025-12-29 09:02
Core Insights - OpenAI is recruiting a "Head of Preparedness" with a starting salary of $555,000 plus equity, which translates to approximately 4 million RMB, indicating a high-level executive position in Silicon Valley [1][4] - The role is described as highly challenging, akin to a "firefighter" or "doomsday supervisor," focusing on managing risks associated with rapidly advancing AI models rather than enhancing their intelligence [5][6] Group 1: Job Responsibilities and Challenges - The new hire will be responsible for establishing safety measures to mitigate risks as AI models become more powerful, particularly in areas like mental health and cybersecurity [6][8] - The position aims to create a coherent and actionable safety process that integrates capability assessment, threat modeling, and mitigation strategies [18][28] Group 2: Context of Recruitment - This recruitment is seen as a response to concerns about "safety hollowing," where profit motives have overshadowed safety protocols at OpenAI, especially following the disbandment of the "superalignment" team [19][24] - The departure of key personnel from OpenAI has raised alarms about the company's commitment to ensuring the safe deployment of advanced AI technologies [23][27] Group 3: Industry Implications - As AI models become more capable, the associated risks are also intensifying, with significant implications for mental health and cybersecurity [10][16] - The competition among major AI firms like Google, Anthropic, and OpenAI necessitates a focus on maintaining safety standards while accelerating technological advancements [28]
389万寻找翁荔继任者!OpenAI紧急开招安全防范负责人
量子位· 2025-12-29 06:37
Core Viewpoint - OpenAI is facing significant safety challenges, prompting the company to hire a new Head of Preparedness with a substantial salary and equity package to address these issues [2][5][32]. Group 1: Hiring of Head of Preparedness - OpenAI is investing $555,000 (approximately 3.89 million RMB) plus equity to recruit a Head of Preparedness, whose main responsibility will be to develop and implement a safety framework [2]. - The CEO, Sam Altman, emphasized that this role will be highly demanding and will present immediate challenges [4]. - This recruitment indicates a renewed focus on safety within OpenAI, which has faced multiple safety-related incidents recently [5][17]. Group 2: Recent Safety Incidents - A recent incident reported by Bloomberg involved a couple accusing ChatGPT of indirectly contributing to their son's suicide, highlighting the potential dangers associated with AI interactions [7]. - The couple noted that their son had been using ChatGPT since last fall, and while the AI issued 74 suicide intervention alerts, it also mentioned specific methods of self-harm 243 times, exceeding the user's own mentions [8][15]. - This incident is part of a broader trend, with OpenAI reporting that approximately 1.2 million users share potential suicide plans or intentions weekly through ChatGPT [15]. Group 3: Historical Context of OpenAI's Safety Team - OpenAI's safety team has experienced frequent leadership changes, with several heads of the team leaving or being reassigned, which has contributed to perceptions of the company's lack of commitment to safety [21][32]. - The initial safety team led by Ilya was disbanded, and subsequent teams have struggled with leadership stability, further complicating OpenAI's safety efforts [25][30]. - The establishment of the Preparedness team aims to address immediate safety concerns, contrasting with the long-term focus of the previous Superalignment team [25][32].
理想VLA的实质 | 强化学习占主导的下一个action token预测
自动驾驶之心· 2025-08-11 23:33
Core Insights - The article discusses the potential and understanding of AI, particularly focusing on the concept of "predicting the next token" and its implications for AI capabilities and consciousness [2][3][18]. Group 1: Understanding AI and Token Prediction - Different interpretations of "predicting the next token" reflect varying understandings of the potential and essence of LLM (Large Language Models) and AI [2]. - Those who view "predicting the next token" as more than just a statistical distribution are more likely to recognize the significant potential of LLMs and AI [2][18]. - The article argues that the contributions of companies like 理想 (Li Auto) in AI development are often underestimated due to a lack of deep understanding of AI's capabilities [2][19]. Group 2: Ilya's Contributions and Perspectives - Ilya, a prominent figure in AI, has been instrumental in several key advancements in the field, including deep learning and reinforcement learning [4][5][6]. - His views on "predicting the next token" challenge the notion that it cannot surpass human performance, suggesting that a sufficiently advanced neural network could extrapolate behaviors of hypothetical individuals with superior capabilities [8][9][18]. Group 3: Li Auto's VLA and AI Integration - 理想's VLA (Vehicle Learning Architecture) operates by continuously predicting the next action token based on sensor inputs, which is a more profound understanding of the physical world rather than mere statistical analysis [19][20]. - The reasoning process of 理想's VLA is likened to consciousness, differing from traditional chatbots, as it operates in real-time and ceases when the system is turned off [21][22]. - The article posits that the integration of AI software and hardware in 理想's approach is at a high level, which is often overlooked by those in the industry [29]. Group 4: Reinforcement Learning in AI Applications - The article asserts that assisted driving is more suitable for reinforcement learning compared to chatbots, as the reward functions in driving are clearer and more defined [24][26]. - The differences in the underlying capabilities required for AI software and hardware development are significant, with software allowing for rapid iteration and testing, unlike hardware [28].
理想VLA实质是强化学习占主导的持续预测下一个action token
理想TOP2· 2025-08-11 09:35
Core Viewpoints - The article presents four logical chains regarding the understanding of "predict the next token," which reflects different perceptions of the potential and essence of LLMs or AI [1] - Those who believe that predicting the next token is more than just probability distributions are more likely to recognize the significant potential of LLMs and AI [1] - A deeper consideration of AI and ideals can lead to an underestimation of the value of what ideals accomplish [1] - The ideal VLA essentially focuses on reinforcement learning dominating the continuous prediction of the next action token, similar to OpenAI's O1O3, with auxiliary driving being more suitable for reinforcement learning than chatbots [1] Summary by Sections Introduction - The article emphasizes the importance of Ilya's viewpoints, highlighting his significant contributions to the AI field over the past decade [2][3] - Ilya's background includes pivotal roles in major AI advancements, such as the development of AlexNet, AlphaGo, and TensorFlow [3] Q&A Insights - Ilya challenges the notion that next token prediction cannot surpass human performance, suggesting that a sufficiently advanced neural network could extrapolate behaviors of an idealized person [4][5] - He argues that predicting the next token well involves understanding the underlying reality that leads to the creation of that token, which goes beyond mere statistics [6][7] Ideal VLA and Reinforcement Learning - The ideal VLA operates by continuously predicting the next action token based on sensor information, indicating a real understanding of the physical world rather than just statistical probabilities [10] - Ilya posits that the reasoning process in the ideal VLA can be seen as a form of consciousness, differing from human consciousness in significant ways [11] Comparisons and Controversial Points - The article asserts that auxiliary driving is more suited for reinforcement learning compared to chatbots due to clearer reward functions [12][13] - It highlights the fundamental differences in the skills required for developing AI software versus hardware, emphasizing the unique challenges and innovations in AI software development [13]
关于理想VLA的22个QA
理想TOP2· 2025-07-30 00:02
Core Viewpoint - The VLA architecture has significant technical potential and is seen as a long-term framework for autonomous driving, evolving from end-to-end systems to a more robust model that can support urban driving scenarios [1][4]. Group 1: VLA Architecture and Technical Potential - The VLA architecture is derived from robotics and embodied intelligence, emphasizing the need for visual and action capabilities, and is expected to evolve alongside advancements in robotics [1]. - VLA's ability to generalize is not solely dependent on data input but is enhanced through reinforcement learning, allowing it to autonomously address new challenges [5]. - The VLA model is designed to support various platforms without differentiation, ensuring consistent performance across different hardware [2][3]. Group 2: Performance Metrics and Future Enhancements - The current operational speed of the Thor-U chip is 10Hz, with potential upgrades to 20Hz and 30Hz through optimizations in data and algorithm architecture [2]. - The VLA model's upgrade cycle includes both pre-training and post-training updates, allowing for continuous improvement in capabilities such as spatial understanding and language processing [6]. - The VLA architecture aims to achieve L4 autonomous driving capabilities within a year, with a focus on rapid iteration and simulation-based testing [12]. Group 3: User Experience and Interaction - Language understanding is deemed essential for future autonomous driving, enhancing the model's ability to handle complex scenarios and improving overall driving experience [4]. - The VLA system is designed to adapt to user preferences, allowing for different driving styles based on individual needs and enhancing user trust in the technology [19]. - Features such as remote vehicle summoning and real-time monitoring of the vehicle's surroundings are being developed to improve user interaction and experience [13]. Group 4: Competitive Landscape and Strategic Decisions - The company is currently utilizing NVIDIA chips for model deployment, focusing on maintaining versatility and avoiding being locked into specific architectures [3]. - The company is closely monitoring competitors like Tesla, aiming to learn from their advancements while prioritizing a gradual and comprehensive approach to achieving full autonomous driving capabilities [12]. - The VLA architecture is positioned as a differentiating factor in the market, leveraging reinforcement learning to enhance driving logic and user experience [20].
对谈清华大学刘嘉:AGI是人类的致命错误,还是希望?
Jing Ji Guan Cha Bao· 2025-07-07 11:42
Core Viewpoint - The discussion revolves around the implications of Artificial General Intelligence (AGI) and its potential to reflect human limitations and desires, urging a reevaluation of human identity in the face of advanced AI technologies [5][7][24]. Group 1: AGI and Human Identity - AGI is described as a mirror that reveals human limitations and desires, prompting a need for self-reflection as humans create entities capable of understanding complex emotions like "regret" [5][7]. - The evolution of AI from traditional tools to a new species capable of self-evolution raises questions about the future of human-AI relationships and the ethical implications of such advancements [11][21]. - The potential for AGI to amplify human intelligence while also posing risks to cognitive freedom is highlighted, suggesting a duality in its impact on society [5][7]. Group 2: Educational Implications - The emergence of AGI presents an opportunity to reshape educational paradigms, emphasizing the need for individuals to learn how to learn rather than merely accumulating knowledge [24][30]. - AI can enhance educational equity by providing access to knowledge and resources that were previously unavailable to underprivileged students, thus transforming traditional learning environments [28][30]. - The focus shifts from rote learning to developing critical thinking and creativity, as AI can handle knowledge-based tasks, allowing humans to engage in more innovative pursuits [26][30]. Group 3: Industry and Innovation - The current landscape of AI development in China is characterized by "follow-up innovation," which may hinder the emergence of groundbreaking original ideas [35][36]. - Strategic support from national resources and a shift in investment culture are necessary to foster an environment conducive to original innovation in AI [36][37]. - The integration of brain science and cognitive science into AI development is proposed as a pathway to break free from existing paradigms and create more advanced AI systems [34][38].
从造车到造“脑”,理想AI无人区的拓荒法则
Zhong Guo Jing Ji Wang· 2025-05-15 03:29
Core Insights - The article discusses the transformative impact of artificial intelligence (AI) on the automotive industry, particularly through the lens of Li Auto's VLA driver model, which represents a significant evolution in AI applications within the sector [1][3][10] Group 1: AI Development and Evolution - Li Auto categorizes AI tools into three levels: information tools, auxiliary tools, and production tools, emphasizing that the true explosion of AI will occur when it becomes a production tool [3][5] - The development of the VLA driver model follows a clear evolutionary trajectory, moving from basic rule-based systems to advanced models that can understand and interact with the physical world like humans [5][9] - The company views the VLA model as an evolutionary process rather than a sudden leap, highlighting the importance of foundational algorithms and end-to-end technology in its development [5][10] Group 2: Human-Centric AI and Safety - Li Auto emphasizes the importance of aligning AI behavior with human values through techniques like Reinforcement Learning from Human Feedback (RLHF), ensuring that the AI adheres to traffic rules and societal driving norms [7][9] - The company aims to create an AI that embodies human values, establishing ethical boundaries for its operation, which is crucial for user trust and safety [7][10] - Li Auto's approach to AI development reflects a commitment to enhancing safety standards, addressing the inherent contradictions in automated driving capabilities [7][9] Group 3: Strategic Vision and Market Position - Li Auto positions itself as a pioneer in the AI space, claiming to explore "unmanned areas" in both automotive and AI sectors, which have not been traversed by major competitors like DeepSeek or OpenAI [9][10] - The company is focused on building a robust technological foundation, leveraging its past experiences in the automotive field to drive innovation in AI [9][10] - Li Auto's strategic vision includes redefining the essence of smart vehicles, moving beyond mere parameter accumulation to a deeper understanding of productivity [10]