量子位
Search documents
谷歌192亿买他回来,现在只想让他闭嘴
量子位· 2025-11-11 11:11
Core Viewpoint - The controversy surrounding Noam Shazzer's statements at Google highlights the ongoing tension between talent retention and adherence to company values, particularly regarding inclusivity and free speech within the organization [4][9][19]. Group 1: Incident Overview - Noam Shazzer, a key figure in the development of the Transformer model, sparked significant internal debate at Google with his controversial remarks on gender issues [6][5]. - The internal forum discussions quickly polarized employees into two opposing camps, with many arguing that Shazzer's comments were provocative and challenged Google's established norms on inclusivity [7][9]. - Google's management intervened by deleting some of Shazzer's comments, which escalated the controversy rather than resolving it, leading to accusations of suppressing free speech [8][9]. Group 2: Noam Shazzer's Contributions - Shazzer is recognized as one of the eight authors of the Transformer model and is credited with making the most significant contributions, including rewriting the project code to enhance its capabilities [20]. - His return to Google was seen as a strategic move, with estimates suggesting that his work on the Gemini project alone is valued at $2.5 billion [14]. - The company invested $2.7 billion to bring Shazzer back, which many consider a worthwhile investment given his pivotal role in AI advancements [28]. Group 3: Historical Context - The current situation draws parallels to the 2017 James Damore incident, where another Google employee was fired for similar issues related to gender discussions [12][19]. - Historical patterns at Google show a recurring theme of conflicts between high-profile employees and management over issues of academic freedom and corporate values, as seen in the cases of Timnit Gebru and Jeff Dean [29][31].
杨植麟回复:Kimi K2训练用的H800!但“只花了460万美元”嘛…
量子位· 2025-11-11 11:11
Core Insights - The Kimi K2 Thinking model reportedly cost only $4.6 million to train, which is lower than the $5.6 million for DeepSeek V3, raising questions about the valuation of closed-source giants in Silicon Valley [13][14]. - The Kimi K2 model is causing a migration trend in Silicon Valley as it offers superior performance at a lower cost compared to existing models [5][6]. - The Kimi K2 model utilizes innovative engineering techniques, including a self-developed MuonClip optimizer, which allows for stable gradient training without human intervention [18]. Training Cost and Performance - The training cost of Kimi K2 is claimed to be $4.6 million, significantly lower than other models, prompting reflection within the industry [13][14]. - Investors and companies are migrating to Kimi K2 due to its strong performance and cost-effectiveness, with reports of it being five times faster and 50% more accurate than closed-source models [8][6]. Technical Innovations - Kimi K2 has optimized its architecture by increasing the number of experts in the MoE layer from 256 to 384 while reducing the number of active parameters during inference from approximately 37 billion to 32 billion [16]. - The model employs Quantization-Aware Training (QAT) to achieve native INT4 precision inference, which enhances speed and reduces resource consumption by about 2 times [21]. Community Engagement and Future Developments - The team behind Kimi K2 engaged with the developer community through a three-hour AMA session, discussing future architectures and the potential for a next-generation K3 model [22][24]. - The team revealed that the unique writing style of Kimi K2 results from a combination of pre-training and post-training processes, and they are exploring longer context windows for future models [26][27].
看图写代码,3毛钱开发一个网页!字节AI Coding新模型真卷麻了
量子位· 2025-11-11 06:59
Core Viewpoint - Volcano Engine has launched a new code model, Doubao-Seed-Code, optimized for Agentic programming tasks, showcasing significant advancements in performance, pricing, and migration costs [2][4][7]. Group 1: Performance - Doubao-Seed-Code achieves state-of-the-art (SOTA) performance, integrating deeply with the TRAE development environment, and ranks at the top of the SWE-Bench Verified leaderboard with a resolution rate of 78.80% [4][63]. - The model is capable of handling multimodal software issues, including those described with images, indicating its versatility in problem-solving [5][64]. - It demonstrates strong capabilities in coding tasks, efficiently completing basic functions and complex interactions, as evidenced by its performance in various coding tests [13][20][28]. Group 2: Pricing - Volcano Engine offers the lowest calling prices in the domestic market, with a subscription plan starting at just 9.9 yuan, making it accessible for developers [6][58]. - The overall usage cost has been reduced by 62.7% compared to industry averages, with Doubao-Seed-Code costing approximately 0.34 yuan for the same token volume that costs 4.05 yuan with Claude Sonnet 4.5 [55][56]. Group 3: Migration Costs - Doubao-Seed-Code is natively compatible with the Anthropic API, allowing for seamless migration with virtually zero configuration costs, making it easy for developers to switch from other models [7][56]. Group 4: Technical Advancements - The model supports visual understanding capabilities, allowing it to generate code from UI design drafts or screenshots, a feature that sets it apart in the domestic market [43][56]. - Doubao-Seed-Code is built on a robust training library with over 100,000 container images and utilizes end-to-end reinforcement learning for efficient optimization [66][67]. Group 5: Market Position - Volcano Engine's Doubao-Seed-Code is positioned as a competitive player in the AI coding landscape, emphasizing performance, affordability, and user-friendly migration, which are critical in the current market [52][74].
iPhone Air卖不动,库克挥泪砍产线…这也就刚卖了一个月
量子位· 2025-11-11 04:24
Core Viewpoint - The iPhone Air has faced significant challenges since its launch, leading to production cuts and a lack of consumer interest, ultimately resulting in its withdrawal from the market [5][25][31] Market Response - The iPhone Air's initial sales were disappointing, with only over 50,000 activations in its first week, which is less than one-tenth of the iPhone 17 Pro Max's performance [5] - In major online sales channels like JD and Tmall, the iPhone Air's sales lagged behind both the iPhone 17 series and the older iPhone 16 models, failing to make it into the top ten of small-screen bestsellers [6] - On Amazon, the iPhone Air received a rating of only 4.4, with many users citing serious battery and performance issues compared to the Pro Max [8] Product Positioning - Apple aimed to create a "non-Pro flagship" by simplifying features, removing high-refresh screens and advanced camera capabilities, while retaining the A-series chip and main camera [12] - However, the iPhone Air's stripped-down features did not meet consumer expectations, leading to a perception of it being underwhelming and overpriced compared to the iPhone 17 Pro, which is only $100 more but offers significantly better specifications [15][16] User Experience Issues - The removal of the physical SIM card slot in favor of eSIM led to complications for users in regions where eSIM is not widely supported, causing frustration and connectivity issues [19][20] - The iPhone Air lacked unique features or configurations that could stimulate accessory development or software adaptation, making it less appealing to consumers [22][23] Competitive Landscape - The withdrawal of the iPhone Air has created a gap in the "light flagship" market, which domestic brands like Xiaomi, OPPO, and Honor have already capitalized on, offering competitive products with strong ecosystems [26][27] - Huawei has quickly responded by launching the Mate 70 Air, which is thinner and lighter at a lower price point, indicating a swift shift in market dynamics following Apple's retreat [28][29] Future Prospects - Although the iPhone Air 2 project has been removed from the main production schedule, internal development continues, with plans for improvements in weight, battery capacity, and camera performance [31][32] - The potential for a future return of the iPhone Air remains, though it may not happen soon [33]
卧底硅谷AI独角兽60天:没有KPI,自觉996,不接受远程办公
量子位· 2025-11-11 04:24
鹭羽 发自 凹非寺 量子位 | 公众号 QbitAI 没有logo、没有招聘JD、没有万恶之源KPI。 反之,全员主动996,写代码互相来找茬,甚至公司最有压力的不是程序员,而是厨子…… 我重生了,重生在 Cursor 的"草根"时期。 成立不到两年,估值就超百亿美元,一经推出全网引爆"氛围编程",不只是让写代码更快,更是重新定义了写代码这事。 | 构建软件的新方式。 | | | | --- | --- | --- | | 前后两批次的效果判若云泥,采用率从个位数飙升至 | 迄今为止我付费使用、毫无疑问最有用的 AI 工具就是 | 最出色的 LLM 应用都有一个"自主性滑杆":你可以决 | | 80%以上。它像野火般迅速蔓延,最顶尖的开发者都 | Cursor。它速度快、在你需要的时机和位置智能补 | 定给 Al 多大的自主权。在 Cursor 中,你可以用 Tab 自 | | 在使用 Cursor。 | 全,括号处理得当,键盘快捷键设计合理,支持自带 | 动补全、用 Cmd+K 做定向编辑,或者直接放手交给全 | | 模型 各方面都打磨得非常到位。 | | 自主代理模式来处理。 | | Diana Hu 胡 ...
打破数据质量鸿沟!清华腾讯Bee项目发布1500万高质量数据集,刷新MLLM全栈开源SOTA
量子位· 2025-11-11 04:24
Core Insights - The article discusses the launch of the Bee project by Tsinghua University and Tencent's Mixuan team, aimed at bridging the performance gap between fully open-source multimodal large language models (MLLMs) and their closed or semi-open counterparts [2][5][26]. Group 1: Background and Motivation - The current MLLM landscape exhibits a three-tier structure: (1) top-tier closed-source models (e.g., Gemini 2.5, GPT-5), (2) semi-open models with private data (e.g., Qwen2.5-VL), and (3) significantly underperforming fully open-source models [5]. - The core bottleneck is identified as the "data quality gap" rather than model architecture [2][10]. Group 2: Key Contributions of the Bee Project - **Honey-Data-15M**: A high-quality SFT dataset comprising 15 million samples, enhanced through a dual-layer Chain of Thought (CoT) approach [6][16]. - **HoneyPipe & DataStudio**: An open-source, end-to-end data enhancement pipeline that provides a transparent and reproducible methodology for data cleaning and CoT augmentation [6][12]. - **Bee-8B**: A new 8 billion parameter model trained on Honey-Data-15M, achieving state-of-the-art (SOTA) results in various benchmarks, rivaling or surpassing mainstream semi-open models [6][21][26]. Group 3: Data Quality Issues - Existing open-source datasets suffer from two main issues: pervasive noise (e.g., factual inaccuracies, mismatched images) and a lack of complex reasoning data [11][14]. - The Bee project emphasizes that the most viable path for the open-source community is to focus on "data quality" rather than merely increasing "data quantity" [11][26]. Group 4: HoneyPipe Process - The HoneyPipe process involves a meticulous "filter-enhance-validate" workflow that produces high-quality datasets [15][18]. - The process includes three stages: noise and irrelevance filtering, short CoT enhancement and validation, and long CoT enhancement for complex queries [18]. Group 5: Performance of Bee-8B - Bee-8B demonstrates superior performance across various benchmarks, including MathVerse and LogicVista, where it achieved scores of 67.0 and 61.3, respectively, outperforming semi-open models [28]. - In general VQA tasks, Bee-8B achieved excellent SOTA scores in multiple benchmarks, including MMStar and CountBench [28]. Group 6: Conclusion - The Bee project effectively addresses the core data quality issues hindering the development of fully open-source MLLMs, advocating for a methodology that prioritizes data quality over sheer volume [26].
从“给答案”到“教动脑”:这届小学生被AI教会主动思考
量子位· 2025-11-11 04:24
Core Viewpoint - The article discusses the evolution of AI in education, highlighting the transition from traditional tutoring methods to advanced AI-driven personalized learning experiences, exemplified by the "Xueersi Learning Machine T4" and its "Xiao Si AI 1-on-1" feature, which aims to enhance student engagement and understanding through interactive and adaptive teaching methods [2][38]. Group 1: Current AI Education Landscape - Various AI education products are emerging, including ChatGPT's learning mode and Google's "Learn Your Way" tool, indicating a growing trend in AI integration within education [2][4]. - Many existing AI education tools focus on efficiency, providing quick answers without addressing deeper understanding, leading to a cycle of rote learning and superficial engagement [2][10]. Group 2: Features of Xiao Si AI 1-on-1 - The "Xiao Si AI 1-on-1" feature represents a significant advancement, functioning as an interactive AI tutor that guides students through problem-solving rather than simply providing answers [4][10]. - It utilizes multimodal perception capabilities to understand both written and verbal inputs, creating a more immersive learning experience [5][10]. - The AI encourages students to write out problem-solving steps, providing real-time feedback and corrections, which fosters critical thinking and deeper comprehension [11][14]. Group 3: Personalized Learning Approach - Xiao Si adapts its teaching strategies based on individual student performance, adjusting the pace and methods to ensure effective learning [21][22]. - It generates dynamic learning profiles for each student, allowing for tailored educational experiences that move away from a one-size-fits-all approach [22][27]. Group 4: Technological Infrastructure - The integration of hardware and software is crucial for achieving low-latency, multimodal interactions, which are essential for creating a native AI teaching experience [30][31]. - The "Nine Chapters Model" (MathGPT) is employed for comprehensive subject tutoring, having received high-level certifications for its capabilities [34][36]. Group 5: Future of AI in Education - The industry is moving towards a model where AI can serve as a complete educational companion, potentially replacing traditional tutoring roles [39][42]. - The article outlines a framework for evaluating AI teachers, suggesting that current AI capabilities are approaching the L3 stage, indicating significant progress in personalized and interactive learning [41][44].
最后一周!人工智能年度榜单申报即将截止。
量子位· 2025-11-11 04:24
本次评选已经从 企业 、 产品 、 人物 三大维度,设立五类奖项。欢迎企业抓住最后时间,尽快报名! 让我们共同见证年度之星,点亮未来的方向。 组委会 发自 凹非寺 量子位|公众号 QbitAI 「2025人工智能年度榜单」申报 已进入倒计时阶段。 今年是量子位 「2025人工智能年度榜单」评选报名 的 第8年。 八年来,我们见证了技术的突破与落地,产业的融合与重塑,也见证了一批 又一批推动时代前行的企业、人物与产品。 企业榜 产品榜 人物榜 2025 人工智能年度 焦点人物 报名方式 本次评选将于 2025年11月17日 截止。评选结果将于量子位主办的 MEET2026智能未来大会 上正式公布。 扫描二维码即可报名评选: 网页端链接:https://wj.qq.com/s2/23740133/iso8/ 如对本次评选有其他疑问,请联系量子位工作人员。添加微信18801103170,或邮件发送至linyu@qbitai.com,并备注「评选-企业-姓 名」。 详细评选标准及报名方式如下。 2025 人工智能年度领航企业 将面向中国人工智能领域,评选出最具综合实力的企业, 参选条件 : 评选标准 : 2025 人 ...
李飞飞最新长文火爆硅谷
量子位· 2025-11-11 00:58
Core Viewpoint - Spatial intelligence is identified as the next frontier for AI, with the potential to revolutionize creativity, robotics, scientific discovery, and more [2][4][10]. Group 1: Definition and Importance of Spatial Intelligence - Spatial intelligence is described as a foundational aspect of human cognition, enabling interaction with the physical world and driving reasoning and planning [20][21]. - The evolution of spatial intelligence is linked to the development of perception and action, which are crucial for understanding and interacting with the environment [12][13][14]. - Historical examples illustrate how spatial intelligence has driven significant advancements in civilization, such as Eratosthenes' calculation of the Earth's circumference and the invention of the spinning jenny [18][19]. Group 2: Current Limitations of AI - Current AI models, including multimodal large language models (MLLMs), have made progress in spatial perception but still fall short of human capabilities [23][24]. - AI struggles with tasks involving physical representation and interaction, lacking the holistic understanding that humans possess [25][26]. Group 3: World Models as a Solution - The concept of "world models" is proposed as a new generative model that can surpass the limitations of current AI by understanding, reasoning, generating, and interacting with complex virtual or real worlds [28][30]. - World models should possess three core capabilities: generative, multimodal, and interactive [31][34][38]. - The development of world models is seen as a significant challenge that requires innovative methodologies to coordinate semantic, geometric, dynamic, and physical aspects [39][41]. Group 4: Applications and Future Potential - The potential applications of spatial intelligence span various fields, including creativity, robotics, science, healthcare, and education [56][57]. - In creativity, platforms like World Labs' Marble are enabling creators to build immersive experiences without traditional design constraints [52][53]. - In robotics, achieving spatial intelligence is essential for robots to assist in various environments, enhancing productivity and human collaboration [60][62]. Group 5: Vision for the Future - The vision for the future emphasizes the importance of AI enhancing human capabilities rather than replacing them, with spatial intelligence playing a crucial role in this transformation [47][50]. - The exploration of spatial intelligence is framed as a collective effort that requires collaboration across the AI ecosystem, including researchers, innovators, and policymakers [51][63].
《麦肯锡2025 AI报告》|附下载
量子位· 2025-11-11 00:58
Core Insights - The report by McKinsey highlights that while 88% of organizations are using AI, only 39% have seen substantial financial returns from it [10][16]. Group 1: AI Adoption and Impact - A majority of enterprises are utilizing AI in at least one business function, indicating that AI has become a standard practice [4][10]. - Despite widespread adoption, less than 40% of organizations are effectively monetizing their AI investments [5][11]. - The report reveals that only high-performing companies are reaping the benefits of AI, with 50% of these companies planning transformative changes driven by AI in the next three years, compared to just 14% of average companies [41][42]. Group 2: AI Agent Utilization - There is a growing interest in AI Agents, with 62% of organizations experimenting with such applications, yet less than 10% have fully integrated them into their operations [22][23]. - Successful implementation of AI Agents is primarily seen in departments with clear processes and high standardization, such as IT and knowledge management [24][25]. - The deployment of AI Agents requires significant restructuring of processes and organizational frameworks, which many companies have yet to achieve [28][29]. Group 3: Financial Performance and AI - While 64% of organizations feel more innovative since adopting AI, only 36% report improved profitability, and just 33% have seen revenue growth [32][35]. - The most significant financial impacts from AI are observed in efficiency-driven roles, such as software engineering and IT management, rather than in revenue-generating functions like marketing and finance [18][36]. Group 4: Talent and Organizational Changes - AI roles are consuming recruitment budgets, with traditional positions being replaced by roles focused on AI capabilities, such as data engineers and AI product managers [53][56]. - The disparity in AI talent acquisition is widening, with large companies hiring AI-related positions at twice the rate of small and medium enterprises [58][59]. - Organizations are experiencing a restructuring of their workforce, with a notable decline in roles that are repetitive and low in creativity [55][56]. Group 5: Risk Management and Governance - High-performing AI organizations are more proactive in addressing risks associated with AI, such as inaccuracies and compliance issues [62][66]. - These organizations deploy AI in critical tasks, indicating a higher tolerance for risk and a focus on efficiency [70][71]. - The report emphasizes that successful AI implementation requires a shift in perspective, viewing AI as a business transformation engine rather than merely a cost-saving tool [72].