Transformer
Search documents
两年前猛裁1.2万人后,谷歌吃起了“回头草”:新招的AI工程师中,20%是「老面孔」
猿大侠· 2025-12-25 04:09
整理 | 郑丽媛 出品 | CSDN(ID:CSDNnews) 如果说过去两年,生成式 AI 的主战场属于 OpenAI、Anthropic 和 Meta,那么 2025 年,谷歌正在用一种更"工程化"、也更务实的方式,重新夺回话 语权。 在 OpenAI、Meta、Anthropic 等公司不断加码 AI 人才、疯狂互相"挖墙脚"的背景下, 谷歌并没有单纯加入竞价战,而是选择了一条更特殊的路径 : 把已经离开的人,再一次请回来 。 据 CNBC 最新 报道, 谷歌 2025 年新招募的 AI 软件工程师中, 约有 20% 是前员工。 这些人曾在谷歌工作过,因裁 员、战略分歧或个人选择离开, 如今又重新回到公司 —— 而 2 0 % 这一比例,已明显高于往年。 一场早有伏笔的"回流" 事实上 , 谷歌 "召回" 前员工 的 趋势并非偶然。 2023 年初, 谷歌 母公司 Alphabet 启动了公司历史上规模最大的一轮裁员: 约 1.2 万人被裁 , 占员工总数的 6%。 彼时, 全球科技行业正经历高通 胀、利率上行与需求放缓的多重冲击,大厂普遍选择"急刹车"。 但与许多公司不同的是,谷歌并未完全切断与离职员 ...
刚做了一份世界模型的学习路线图,面向初学者......
自动驾驶之心· 2025-12-25 03:24
最近和业内专家讨论了很多,分享一个最近被问到很多的问题: 世界模型是不是端到端? 第一个问题的答案是明确的:不是。 世界模型和端到端都不指某个具体的技术,而是一类具备某些特定能力的模型。可以理解为 世界模型只是一种实现端到端自 动驾驶的途径。 目前学术界和工业界把自动驾驶世界模型收敛到生成和重建两个领域,并且主流都在利用世界模型在做闭环仿真,所以我们看到了很多相关工作的推出。这也是业 内风格转换的一个趋势,Corner Case的成本过高,我们需要更有效的的其他手段...... 点击下方 卡片 ,关注" 自动驾驶之心 "公众号 戳我-> 领取 自动驾驶近30个 方向 学习 路线 讲师介绍 Jason:C9本科+QS50 PhD,已发表CCF-A论文2篇,CCF-B论文若干。现任国内TOP主机厂算法专家,目前从事端到端、大模型、世界模型等前沿算法的预研和量 产,并已主持和完成多项自动驾驶感知和端到端算法的产品量产交付,拥有丰富的端到端算法研发和实战经验。 课程大纲 这门课程讲如何展开 第一章:世界模型介绍 第一章主要针对自动驾驶世界模型概括性的内容讲解。 这一章老师会先复盘世界模型和端到端自动驾驶的联系,接着讲 ...
下周开课!我们设计了一份自动驾驶世界模型学习路线图....
自动驾驶之心· 2025-12-24 09:22
点击下方 卡片 ,关注" 自动驾驶之心 "公众号 戳我-> 领取 自动驾驶近30个 方向 学习 路线 最近和业内专家jason老师讨论了很多,分享一个最近被问到很多的问题: 世界模型是不是端到端?以及如何看待世界模型最近爆发式的工作发表。 第一个问题的答案是明确的:不是。 世界模型和端到端都不指某个具体的技术,而是一类具备某些特定能力的模型。可以理解为 世界模型只是一种实现端到端自 动驾驶的途径。 早鸟优惠!开课即止~ 目前学术界和工业界把自动驾驶世界模型收敛到生成和重建两个领域,并且主流都在利用世界模型在做闭环仿真,所以我们看到了很多相关工作的推出。这也是业 内风格转换的一个趋势,Corner Case的成本过高,我们需要更有效的的其他手段...... 先前平台和Jason老师共同打造的《端到端与VLA自动驾驶小班课》备受大家好评,因此我们进一步推出这门世界模型小班课, 课程聚焦于通用世界模型、视频生 成、OCC生成等世界模型算法,涵盖特斯拉世界模型、李飞飞团队Marble等。欢迎大家加入学习~ 讲师介绍 Jason:C9本科+QS50 PhD,已发表CCF-A论文2篇,CCF-B论文若干。现任国内TOP主 ...
谷歌创始人罕见反思:低估 Transformer,也低估了 AI 编程的风险,“代码错了,代价更高”
AI前线· 2025-12-21 05:32
Group 1 - The core viewpoint of the article emphasizes the rapid advancements in AI, particularly in code generation, while also highlighting the associated risks and challenges, as noted by Sergey Brin [2][3][20] - Brin pointed out that AI's ability to write code can lead to significant errors, making it more suitable for creative tasks where mistakes are less critical [2][38] - He reflected on Google's initial hesitations regarding generative AI and the underestimation of the importance of scaling computational power and algorithms [2][22][24] Group 2 - The discussion included a historical overview of Google's founding, emphasizing the creative and experimental environment at Stanford that fostered innovation [4][6][10] - Brin noted that the early days of Google were characterized by a lack of clear direction, with many ideas being tested without strict limitations [6][9] - The importance of a strong academic foundation in shaping Google's culture and approach to research and development was highlighted [12][13] Group 3 - Brin discussed the competitive landscape of AI, noting that significant investments in AI infrastructure have reached hundreds of billions, with companies racing to lead in this space [21][22] - He acknowledged that while Google has made substantial contributions to AI, there were missed opportunities in the past due to insufficient investment and fear of releasing products prematurely [22][23][24] - The conversation also touched on the evolving nature of AI, with Brin expressing uncertainty about its future capabilities and the potential for AI to surpass human abilities [27][29][30] Group 4 - Brin emphasized the need for a balance between computational power and algorithmic advancements, stating that algorithmic progress has outpaced scaling efforts in recent years [3][55] - He mentioned that deep technology and foundational research are crucial for maintaining a competitive edge in AI [24][25] - The discussion concluded with reflections on the role of universities in the future, considering the rapid changes in education and knowledge dissemination due to technology [41][42]
AGI为什么不会到来?这位研究员把AI的“物理极限”讲透了
3 6 Ke· 2025-12-17 11:43
Group 1 - The article discusses the skepticism surrounding the realization of Artificial General Intelligence (AGI), emphasizing that current optimism in the market may be misplaced due to physical constraints on computation [1][4]. - Tim Dettmers argues that computation is fundamentally bound by physical laws, meaning that advancements in intelligence are limited by energy, bandwidth, storage, manufacturing, and cost [3][4]. - Dettmers identifies several key judgments regarding AGI: the success of Transformer models is not coincidental but rather an optimal engineering choice under current physical constraints, and further improvements yield diminishing returns [4][6]. Group 2 - The article highlights that discussions about AGI often overlook the physical realities of computation, leading to misconceptions about the potential for unlimited scaling of intelligence [5][9]. - It is noted that as systems mature, linear improvements require exponentially increasing resource investments, which can lead to diminishing returns [10][16]. - The article points out that the performance gains from GPUs, which have historically driven AI advancements, are nearing their physical and engineering limits, suggesting a shift in focus is necessary [18][22]. Group 3 - Dettmers suggests that the current trajectory of AI development may be approaching a stagnation point, particularly with the introduction of Gemini 3, which could signal a limit to the effectiveness of scaling [33][36]. - The cost structure of scaling has changed, with past linear costs now becoming exponential, indicating that further scaling may not be sustainable without new breakthroughs [35][36]. - The article emphasizes that true AGI must encompass the ability to perform economically meaningful tasks in the real world, which is heavily constrained by physical limitations [49][50]. Group 4 - The discussion includes the notion that the concept of "superintelligence" may be flawed, as it assumes unlimited capacity for self-improvement, which is not feasible given the physical constraints of resources [56][58]. - The article argues that the future of AI will be shaped by economic viability and practical applications rather than the pursuit of an idealized AGI [59][60].
布林坦承谷歌低估Transformer,“还被OpenAI挖走了Ilya”
3 6 Ke· 2025-12-15 11:02
Core Insights - Google founder Sergey Brin reflected on the company's journey, acknowledging mistakes in the AI race and recognizing OpenAI's opportunity [1][4] - Brin emphasized the importance of not rushing to commercialize ideas without adequate preparation, using Google Glass as a cautionary example [25][27] Company History - Google was founded in 1998, evolving from a project called BackRub, which assessed webpage importance through links [11][12] - The name "Google" is derived from a mathematical term representing a 1 followed by 100 zeros, symbolizing the ambition to organize global information [14] AI Development - Google initially underestimated AI's potential after the release of the Transformer paper, leading to missed opportunities as OpenAI capitalized on the technology [20] - Despite setbacks, Google's long-term investment in AI research and development, including the creation of specialized TPU chips, has maintained its competitive edge [20] Future Technologies - Brin identified quantum computing and materials science as undervalued future technologies, suggesting a focus on their applications in AI [23] - He advised students to leverage AI in various aspects of life, while cautioning against pursuing fields where AI may excel, such as comparative literature [21][23] Entrepreneurial Advice - Brin warned young entrepreneurs against prematurely commercializing unrefined ideas, stressing the need for thorough preparation and cost management [25] - He shared insights from his return to Google, emphasizing the importance of staying engaged and continuously learning [27][29]
重磅!8 年后回到斯坦福,谷歌创始人谢尔盖·布林复盘:AI为什么落后,又如何实现绝地反击?(附视频)
美股IPO· 2025-12-15 00:24
Core Insights - The article discusses the evolution of Google and its approach to AI, highlighting both past mistakes and future opportunities in the context of education and technology [3][10][12]. Group 1: Google's AI Strategy - Google initially missed opportunities in AI commercialization due to hesitance in promoting chatbots, fearing they would produce nonsensical outputs [3][15]. - The company's competitive edge in AI stems from long-term investments in foundational technologies, such as AI-specific chips (TPUs) and large-scale data centers [4][16]. - Future breakthroughs in AI are expected to rely more on algorithmic advancements rather than merely scaling data and computational power [5][29]. Group 2: Education and Career Guidance - Students should view AI as a tool to enhance their capabilities rather than a reason to abandon traditional fields like computer science [7][18]. - The future of universities may shift away from geographical constraints, emphasizing remote learning and collaboration [20][21]. - The path from academia to industry may need reevaluation as the timeline for turning ideas into commercial products shortens [22][23]. Group 3: Research and Innovation - While industry leads many innovations, academic research remains crucial for long-term exploratory projects that require extensive timeframes [24][25]. - Emerging technologies, particularly in materials science, are seen as underappreciated areas with significant potential for impact [32][34].
AI医疗影像:在数据“围城”中如何突围
经济观察报· 2025-12-10 10:39
Core Viewpoint - The article emphasizes the importance of addressing data challenges in the medical imaging sector, which not only facilitates the revolutionary development of medical AI but also provides valuable experiences and models for AI applications across various industries [1]. Group 1: AI in Medical Imaging - The National Health Commission of China has set a timeline for the development of "AI + Healthcare," aiming for comprehensive coverage of intelligent diagnostic applications in primary care by 2030 [2]. - The AI medical imaging industry has matured, with major hospitals adopting AI products for diagnostic assistance [3]. - AI has significantly improved the efficiency of medical imaging diagnostics, reducing the time required for doctors to complete reports from approximately 30 minutes to 5-10 minutes, thus alleviating the workload of overburdened radiologists [5][6]. Group 2: Commercialization Challenges - Despite the substantial value created by AI in medical imaging, the industry faces a commercialization dilemma, with cumulative revenues projected to be less than 3 billion yuan from 2020 to 2024 [8]. - The low technical barriers and intense competition have led to a market where many companies offer similar products, often resorting to free trials to gain hospital access, which undermines profitability [9][10]. - Many hospitals, especially lower-tier ones, struggle with budget constraints, limiting their ability to invest in AI products, which further compresses the market potential [10]. Group 3: Future Potential of AI - To unlock greater potential, AI must enhance its value in medical imaging analysis, diagnosis, and treatment, which requires higher research and development barriers [12]. - Current AI models primarily based on Convolutional Neural Networks (CNN) have limitations in understanding complex medical images, while the introduction of Transformer models could significantly improve diagnostic capabilities [13][14]. - The integration of multi-modal data processing through Transformer models could lead to comprehensive clinical decision-making models, breaking down barriers between different types of medical data [14]. Group 4: Data Challenges - The transition from CNN to Transformer-based models presents significant data challenges, as training such models requires vast amounts of high-quality labeled data, which is difficult to obtain in the medical field due to privacy regulations [18][19]. - The complexity of multi-modal data integration further complicates the data landscape, necessitating extensive coordination and processing efforts [19]. - Addressing data issues is crucial for advancing AI in medical imaging, and companies that can establish robust capabilities in data collection, governance, and utilization will likely lead the next generation of medical AI [20].
北京大学:AI视频生成技术原理与行业应用 2025
Sou Hu Cai Jing· 2025-12-09 06:48
Group 1: AI Video Technology Overview - AI video technology is a subset of narrow AI focused on generative tasks such as video generation, editing, and understanding, with typical methods including text-to-video and image-to-video [1] - The evolution of technology spans from the exploration of GANs before 2016 to the commercialization of diffusion models from 2020 to 2024, culminating in the release of Sora in 2024, marking the "AI Video Year" [1] Group 2: Main Tools and Platforms - Key platforms include OpenAI Sora, Kuaishou Keling AI, ByteDance Jimeng AI, Runway, and Pika, each offering unique features in terms of duration, quality, and style [2] Group 3: Technical Principles and Architecture - The mainstream paradigm is the diffusion model, which is stable in training and offers strong generation diversity, with architectures categorized into U-Net and DiT [3] - Key components include the self-attention mechanism of Transformers for temporal consistency, VAE for compression, and CLIP for semantic alignment between text and visuals [3] Group 4: Data Value and Training - The scale, quality, and diversity of training data determine the model's upper limits, with prominent datasets including WebVid-10M and UCF-101 [4] Group 5: Technological Advancements and Breakthroughs - Mainstream models can generate videos at 1080p/4K resolution and up to 2 minutes in length, with some models supporting native audio-visual synchronization [5] - Existing challenges include temporal consistency, physical logic, and emotional detail expression, alongside computational cost constraints [5] - Evaluation frameworks like VBench and SuperCLUE have been established, focusing on "intrinsic authenticity" [5] Group 6: Industry Applications and Value - In the film and entertainment sector, AI is involved in the entire production process, leading to cost reductions and efficiency improvements [6] - The short video and marketing sectors utilize AI for rapid content generation, exemplified by Xiaomi's AI glasses advertisement [6] - In the cultural tourism industry, AI is used for city promotional videos and immersive experiences [7] - In education, AI facilitates the bulk generation of micro-course videos and personalized learning content [8] - In news media, AI virtual anchors enable 24-hour reporting, though ethical challenges regarding content authenticity persist [9] Group 7: Tool Selection Recommendations - Recommendations for tool selection include using Runway or Keling AI for professional film, Jimeng AI or Pika for short video operations, and Vidu for traditional Chinese content [10] - Domestic tools like Keling and Jimeng have low barriers to entry, while overseas tools require VPN and foreign currency payments [11] - A multi-tool collaborative workflow is advised, emphasizing a "director's mindset" rather than reliance on a single platform [12] Group 8: Future Outlook - The report concludes that AI video will evolve towards a "human-machine co-creation" model, becoming a foundational infrastructure akin to the internet, with a focus on creativity and judgment [13]
Roblox CEO感叹AI研究进展:曾博览群书的自己都快看不懂了
Sou Hu Cai Jing· 2025-12-08 11:28
巴祖基 2005 年创立 Roblox。创业初期,他几乎读遍了从物理模拟到图形渲染的各类研究,而且都能理解。然而 AI 时代的到来改变了一切。他称如今的研究浪潮"规模巨大、速度惊人",从 Transformer 到扩散模型,再到世界 模型,"内容多到让人难以完全掌握"。 IT之家 12 月 8 日消息,AI 研究更新速度飞快,新论文几乎每天出现,技术概念也越来越复杂,Roblox CEO 大 卫・巴祖基对此深有体会。 据《商业内幕》今日报道,巴祖基透露,自己休假时抽出大量时间系统阅读 AI 研究,却发现过程"发人深省"—— 想真正看懂所有论文"极其困难"。 尽管外界关注焦点集中在算力扩张,OpenAI 联合创始人伊利亚・苏茨克维却认为,真正决定 AI 走向的仍是"研 究本身":"我们重新回到研究时代,只不过现在用的是更大的计算机。" 而对于 Roblox 而言,巴祖基的结论是:AI 在"三维世界"里仍然处于非常初期的阶段。他指出,AI 依赖的是人类 制造出来的文本和图像,"我们在用自己创造的内容训练 AI,而不是用真实世界的三维原始数据"。 随着 AI 从学界扩展到国家战略高度,Meta、微软等公司纷纷建立自 ...