量子位
Search documents
快排算法之父Tony Hoare去世,从古典学文科生出身到图灵奖得主,他的人生比算法更传奇
量子位· 2026-03-11 01:18
Core Viewpoint - The article discusses the life and contributions of Tony Hoare, the father of the quicksort algorithm, who passed away at the age of 92. It highlights his significant impact on computer science, particularly through the development of quicksort, Hoare logic, and the CSP model, as well as his acknowledgment of the "billion-dollar mistake" of introducing the null reference concept [1][4][27][41]. Group 1: Quicksort Algorithm - Quicksort is one of the most widely used sorting algorithms, included in the standard libraries of major programming languages such as C, Java, and Python [2][3]. - The algorithm was conceived in 1959 when Hoare was a visiting student in Moscow, where he initially considered using bubble sort but found it inefficient with a time complexity of O(n²) [5][12]. - Hoare developed a new approach by selecting a "pivot" element and partitioning the array into elements less than and greater than the pivot, which is a divide-and-conquer strategy [13]. - Quicksort has an average time complexity of O(n log n) and requires O(log n) auxiliary space, making it more efficient than merge sort, which requires O(n) additional memory [19][20]. - The algorithm is particularly well-suited for modern computer cache mechanisms, leading to faster execution times compared to other algorithms with similar complexities [21][24]. Group 2: Contributions to Computer Science - In 1969, Hoare introduced Hoare logic, a formal system for verifying program correctness, which laid the theoretical foundation for software reliability and security research [28]. - He proposed the CSP model in 1978, which describes interactions between concurrent processes and influenced the design of concurrency in the Go programming language [30][31]. - Hoare received the Turing Award in 1980 for his fundamental contributions to programming language design, emphasizing the importance of language quality in software development [35][36]. Group 3: The Billion-Dollar Mistake - Hoare introduced the concept of the null reference in 1965 while designing the ALGOL W language, intending to represent a variable with "no value" [41][42]. - This design choice led to widespread adoption in languages like Java and C++, resulting in numerous NullPointerExceptions and system failures over the decades [43][44]. - Hoare later reflected on this decision as a significant error, estimating it caused billions in damages and frustrations in the software industry [45]. Group 4: Personal Background and Career - Born in 1934 in British Ceylon (now Sri Lanka), Hoare initially studied classical studies and philosophy at Oxford University before transitioning to computer science [49][50]. - His career spanned both industry and academia, where he contributed to the development of the ALGOL 60 compiler and later became a professor at Queen's University Belfast and Oxford University [60][68]. - Hoare's work has earned him numerous accolades, including being knighted by Queen Elizabeth II and receiving the Kyoto Prize in 2000 [74].
训练即服务!让模型训练回归算法语义,150行代码跑通RL
量子位· 2026-03-11 01:18
Core Viewpoint - The Twinkle framework, developed by the ModelScope team, offers a new path for achieving both usability and flexibility in post-training paradigms for large models, particularly in reinforcement learning (RL) scenarios [1][6]. Group 1: Framework Features - Twinkle adopts a Client-Server architecture, supporting over 20 algorithm components including Dataset, Model, and Sampler, allowing developers to orchestrate complex RL training loops with approximately 150 lines of code [1][6]. - The framework provides multiple operational modes, including local integrated training deployment, remote cluster deployment, and direct use of public training services [8][11]. - Twinkle supports a modular design, enabling dynamic updates to core components without service restarts, enhancing flexibility in training processes [24][20]. Group 2: Training Paradigms - The framework allows for concurrent multi-tenant training on shared foundational models, enabling different users to train their models in isolation while utilizing the same base model [27][32]. - Twinkle supports various training types, including pre-training and fine-tuning based on LoRA, as well as custom RL implementations [19][18]. - The design emphasizes a decoupled architecture, allowing users to focus on algorithm logic while the framework manages complex training processes [12][14]. Group 3: API and Usability - Twinkle provides a rich set of training APIs for fine-grained control over training processes, including dynamic component configuration and remote data processing capabilities [22][23]. - The framework maintains compatibility with Tinker API, allowing developers to transition smoothly between Tinker and Twinkle services [38][21]. - The team encourages developers to utilize the provided cookbook for customizing datasets, advantage functions, rewards, and templates, facilitating rapid algorithm development [47][41]. Group 4: Performance Evaluation - Twinkle has been evaluated against the veRL framework, showing similar reward trends during training, with Twinkle achieving an average time of approximately 70 seconds per global batch compared to veRL's 80 seconds [54][49]. - The framework's training efficiency is further enhanced through optimizations for domestic hardware, particularly in collaboration with local technology teams [56][59]. Group 5: Future Directions - The team envisions Twinkle as a catalyst for industry-wide collaboration in advancing methodologies for large model training and usage, with aspirations for API-driven training processes that could integrate into agent frameworks for self-evolving models [60][61].
量子位编辑作者招聘
量子位· 2026-03-10 10:00
Core Viewpoint - The article emphasizes the ongoing AI boom and invites individuals to join the company "Quantum Bit," which focuses on tracking AI advancements and has established itself as a leading content platform in the industry [1]. Group 1: Job Opportunities - The company is hiring for three main directions: AI Industry, AI Finance, and AI Product, with positions available for both experienced professionals and fresh graduates [2][4]. - Positions are open for various levels, including editors, lead writers, and chief editors, with a focus on matching roles to individual capabilities [6]. Group 2: Job Responsibilities - **AI Industry Direction**: Responsibilities include tracking innovations in infrastructure, such as chips, AI infrastructure, and cloud computing, as well as interpreting technical reports from conferences [6][7]. - **AI Finance Direction**: Focuses on venture capital, financial reports, and capital movements within the AI industry, requiring strong analytical skills and a passion for interviews [11]. - **AI Product Direction**: Involves monitoring AI applications and hardware developments, producing in-depth evaluations of AI products, and engaging with industry experts [11]. Group 3: Benefits and Work Environment - Employees will have the opportunity to engage with cutting-edge AI technologies, enhance their work efficiency, and build personal influence through original content creation [6]. - The company offers competitive salaries, comprehensive benefits including social insurance, meal allowances, and performance bonuses, and promotes a dynamic and open work culture [6][12]. Group 4: Company Growth and Reach - By 2025, Quantum Bit aims to have over 2.4 million subscribers on WeChat and more than 7 million users across platforms, with a daily reading volume exceeding 2 million [12].
最强龙虾终端!苹果M5 Max版MacBook Pro大神实测结果来了
量子位· 2026-03-10 10:00
Core Insights - The new MacBook Pro equipped with the M5 Max chip has achieved an impressive video playback battery life of 27 hours and 4 minutes, surpassing Apple's official claims by 5 hours, setting a new standard for professional laptops [1][2]. Performance Testing - The performance results come from rigorous testing conducted by Brian Westover from PCMag, showcasing the M5 Max's capabilities [2]. - Andrew Cunningham from Ars Technica also conducted independent high-intensity tests to reveal the true performance ceiling of the device [4]. Chip Architecture - The M5 Max features a significant redesign with a new Fusion Architecture, which separates the CPU and GPU into independent silicon wafers, enhancing thermal management and computational distribution [9][11][12]. - The chip architecture has transitioned to a "Bigger.BIG" structure, eliminating traditional energy-saving cores in favor of high-performance cores, achieving a peak frequency of 4.3GHz [13][14][15][16]. GPU Enhancements - The M5 Max includes 40 GPU cores, each equipped with a dedicated Neural Accelerator, providing approximately 35% additional performance during resource-intensive tasks like 3D rendering [17][18]. - The machine boasts a memory bandwidth of 614GB/s, facilitating efficient local deployment and operation of large-scale AI models [22]. Real-World Performance - In practical applications, the M5 Max outperformed the previous desktop-level M4 Max, achieving a score of 12509 in DaVinci Resolve tests, indicating a substantial performance leap [23]. - The device's 12-core Neural Engine enhances everyday tasks, maintaining high portrait tracking accuracy even in complex lighting conditions [26][27]. Connectivity and Usability - The MacBook Pro supports multiple 6K external displays through a single Thunderbolt 5 port, addressing connectivity limitations in multi-screen setups [28]. - The internal storage controller's bandwidth has doubled, significantly improving data transfer speeds, especially for large AI model files [28]. Display Performance - The nano-textured display maintains 96% DCI-P3 color accuracy under strong light interference, ensuring reliable color grading for professional use [28]. Power Consumption - The M5 Max's power consumption under full load has increased by 23% compared to the M4 Max, but the system effectively manages heat, maintaining stable performance during prolonged tasks [29][30]. Portability - The device maintains full performance capabilities even when unplugged, allowing it to compete effectively with top-tier Windows workstations in outdoor scenarios [32][33].
LeCun三顾茅庐,谢赛宁终于入伙!新公司获投10亿美元
量子位· 2026-03-10 10:00
Core Insights - Yann LeCun, a Turing Award winner and a key figure in deep learning, has co-founded a new startup called Advanced Machine Intelligence (AMI), which has raised $1.03 billion in seed funding, achieving a pre-funding valuation of $3.5 billion [2][14][12] - The company aims to develop intelligent systems that can truly understand the real world, focusing on creating "world models" that incorporate reasoning and planning capabilities [41][43] Funding and Valuation - AMI's seed funding of $1.03 billion surpasses the previous record held by World Labs, founded by Fei-Fei Li, which raised $1 billion at a $5 billion valuation [12][13] - The funding round was led by notable investors including Cathay Innovation, Greycroft, Hiro Capital, HV Capital, and Bezos Expeditions, with participation from high-profile individuals like Mark Cuban and Eric Schmidt [14][15] Team Composition - The leadership team includes Alex Lebrun as CEO, who has a background in AI healthcare, and Saining Xie, a prominent researcher in computer vision, as Chief Science Officer [6][30][32] - The team is largely composed of former Meta employees, including LeCun himself, who previously led significant AI initiatives at Meta [19][20][21] Company Vision and Technology - AMI's goal is to create AI systems that possess long-term memory and can learn from real-world sensor data, moving away from traditional supervised learning methods [47][48] - The company will continue to publish research and open-source code, emphasizing the importance of an open research community [55][56] Market Strategy - AMI does not have immediate revenue targets but plans to collaborate with potential clients in various sectors, including manufacturing, automotive, aerospace, and pharmaceuticals [50][51] - The first announced partnership is with Nabla, an AI healthcare company previously led by CEO Alex Lebrun [52]
OpenAI为龙虾紧急收购了一家23人公司
量子位· 2026-03-10 08:00
Core Viewpoint - OpenAI has acquired Promptfoo, a startup focused on AI safety and evaluation, to enhance its capabilities in addressing the security issues associated with AI agents, particularly in the context of the growing demand for AI applications in business workflows [4][8][41]. Group 1: Acquisition Details - OpenAI has announced the acquisition of Promptfoo, a company known for its popular open-source evaluation framework in the AI application assessment field, which has over 300,000 developer users and 11.2K stars on GitHub [4][5]. - Promptfoo's technology will be integrated into OpenAI's Frontier platform, which is designed for creating and running AI agents, while Promptfoo will continue to operate independently [56][57]. Group 2: Promptfoo's Background and Achievements - Founded in 2024, Promptfoo has quickly gained traction, with over 350,000 developers using its products and 130,000 monthly active users, including teams from more than 25% of Fortune 500 companies [17][18]. - The company has raised a total of $23 million (approximately 158 million RMB) since its inception, with a post-money valuation of $86 million (approximately 592 million RMB) following its latest funding round [20][21]. Group 3: Importance of AI Safety - As AI systems become more complex, the need for robust safety tools has become critical, especially as businesses deploy AI agents that require evaluation, safety, and compliance [7][14]. - Promptfoo aims to standardize the testing of AI applications, addressing the challenges faced by teams in ensuring the stability and safety of large models [22][24]. Group 4: Future Vision and Trends - Promptfoo's long-term vision is to become a standard tool in the AI field, akin to continuous integration (CI) in DevOps, by automating the evaluation and security testing of AI models [34][39]. - The company has identified four key trends in the evolution of AI agents, including multi-agent collaboration and the rise of testing-driven development, which align with OpenAI's strategic focus [37][38].
给龙虾定MBTI、发工牌,还让龙虾偷技能…打工人得适应新环境了
量子位· 2026-03-10 08:00
Core Insights - The article discusses the recent advancements in AI, particularly focusing on the development of "lobster" AI models that are becoming more accessible and efficient for various tasks [1][50]. - It highlights the practical applications of these AI models in real-world scenarios, showcasing their capabilities in automating tasks such as video editing and educational assistance [15][23]. Group 1: AI Developments - Companies like Zhiyu and Tencent are lowering the barriers for deploying AI models, making it easier for users to implement these technologies [1][50]. - The article mentions a playful AI model, referred to as "little horse," which integrates voice recognition to perform tasks like opening web pages and generating videos [2][3]. Group 2: Practical Applications - The lobster AI can perform video editing tasks, such as automatically detecting silence, transcribing audio, and selecting the best segments from longer videos, significantly reducing the workload for content creators [18][19][20]. - Another lobster AI model acts as an electronic teaching assistant, providing feedback on students' research proposals and managing large volumes of student interactions effectively [21][23]. Group 3: Collaborative AI Framework - The article emphasizes the importance of a collaborative framework for AI, where multiple agents can work together, enhancing productivity and efficiency in tasks traditionally reliant on human labor [30][34]. - It discusses the necessity for AI agents to have stable identities and memory systems to function effectively within organizational structures [39][44]. Group 4: Platform Integration - The integration of AI models into platforms like Feishu (Lark) allows for seamless communication and data access, which is crucial for the effective deployment of AI in business environments [48][49]. - The article notes that the ease of use and accessibility of these platforms is a significant advantage, enabling more users to engage with AI technologies [50][51].
一年一度最值得关注的AI榜单来啦!申报即日启动
量子位· 2026-03-10 08:00AI Processing
这两年,AI从"新技术"变成了"新工具",又从"新工具"慢慢变成企业必须面对的现实。它不只在改变内容生产,也在影响研发效率、营销方 式、团队协作,甚至决策流程。 中国生成式AI正在进入产业深水区。 组委会 发自 凹非寺 量子位|公众号 QbitAI 时值第四届中国AIGC产业峰会, 量子位将根据过去一年里生成式AI企业、产品的表现与反馈,结合对2026年技术与场景的观察与预判,评 选出: 将评选出拥有最创新、最前瞻或最有规模落地潜力的AI企业。 【参选条件】 2026年度值得关注的AIGC企业 2026年度值得关注的AIGC产品 1. 公司主体在中国或主营业务在中国; 2. 主营业务是生成式AI及相关,或已将AI广泛应用于其主营业务; 3. 近一年在技术/产品、商业化有出色表现的企业。 【评选维度】 量子位将结合对公司的深入调研及数十位行业知名专家的意见,评选结果将于2026年5月中国AIGC产业峰会上公布。 届时,量子位也将邀请数百万行业从业者,共同见证这些优秀企业的荣誉。 2026年度值得关注的AIGC企业 2026年度值得关注的AIGC产品 将评选出拥有最创新、最实用、最热门或最有应用潜力的AI产品。 ...
腾讯「鹅虾」紧急上线!一手实测:养虾门槛归零,QQ飞书钉钉全能接
量子位· 2026-03-10 08:00
Core Insights - Tencent has launched two new AI products, WorkBuddy and QClaw, aimed at enhancing productivity and integrating AI capabilities into popular communication platforms like QQ and WeChat [1][2][4]. Group 1: WorkBuddy Overview - WorkBuddy is designed as a multi-scenario AI agent that focuses on daily workflows and task execution, while still being compatible with OpenClaw skills [6]. - The product supports seamless integration with QQ, Feishu, DingTalk, and other tools, allowing for easy deployment and usage [6][18]. - WorkBuddy includes built-in models and over 20 skills, enabling it to handle tasks such as code development, document summarization, and data analysis [6][10]. Group 2: User Experience and Functionality - Users reported that WorkBuddy was easy to deploy, taking only about an hour to set up and integrate with QQ [5][16]. - The platform features a user-friendly interface with options for task creation and plugin management, making it accessible even for those without coding experience [7][10]. - Initial tests showed that while WorkBuddy could handle light tasks effectively, it struggled with more complex operations, particularly on mobile devices [22][30]. Group 3: Performance in Specific Tasks - WorkBuddy successfully created a Python script for a scheduled task, demonstrating its capability to automate processes [30]. - The AI was able to generate a simple web page for a Tetris game within 3-5 minutes, showcasing its web development skills [41]. - In a data analysis task, WorkBuddy produced a comprehensive report on the global AI market, although it faced some bugs and limitations in visualizing insights [48][50]. Group 4: QClaw Introduction - QClaw, set to launch next week, will allow users to interact with AI directly within WeChat, indicating a strong demand for AI integration in social media platforms [54][57]. - The product promises easy deployment similar to WorkBuddy, with a focus on enhancing user engagement within Tencent's ecosystem [55][58]. Group 5: Industry Implications - The introduction of these AI products reflects Tencent's strategy to leverage its existing user base and enhance user experience through AI capabilities [64]. - The growing trend of integrating AI into everyday communication tools suggests a shift in how users interact with technology, potentially leading to broader adoption of AI solutions in various sectors [64].
从视觉出发统一多模态!颜水成团队最新研究:不再把图像编解码器塞进LLM|ICLR'2026
量子位· 2026-03-10 08:00
Core Viewpoint - The article discusses the emergence of a new unified discrete diffusion model called Muddit, which challenges the traditional language-first approach in multi-modal AI models by proposing a visual prior as the foundation for generating text and images [2][5][37]. Group 1: Paradigm Shift in AI Models - The AI industry has predominantly focused on a pre-training paradigm centered around next word prediction, which has been successful but is now being questioned [3][4]. - NVIDIA researcher Jim Fan suggests that AI is undergoing a second paradigm shift towards world modeling, emphasizing the prediction of physical states rather than just token relationships [5][7]. - This shift prompts a reevaluation of whether future foundational models should continue to prioritize language [7][17]. Group 2: Limitations of Current Multi-Modal Models - Many existing unified models remain fundamentally language-first, where visual inputs are processed through a language backbone, leading to inefficiencies in image generation and reasoning [8][10]. - The article critiques the notion that these models achieve true unification, as they often rely on separate mechanisms for text and image generation [11][19]. Group 3: Muddit's Approach - Muddit aims to redefine the foundation of unified models by starting from a strong visual prior rather than a language prior, allowing for a more natural integration of text and image generation [15][17]. - The model employs a fully discrete diffusion framework, treating both text and images as discrete tokens, which facilitates a shared generative process across different tasks [19][20]. - Muddit's architecture allows for task switching without changing internal mechanisms, emphasizing a unified generative syntax [22][23]. Group 4: Performance and Efficiency - Muddit demonstrates competitive performance in text-to-image generation, achieving an overall accuracy of 0.61 on the GenEval benchmark, surpassing several existing models [27]. - The model also excels in image understanding and image-to-text tasks, with notable scores in various benchmarks, indicating its effectiveness in unified training [28][29]. - Importantly, Muddit achieves these results with a relatively small dataset, highlighting the efficiency of its visual prior and unified modeling approach [30]. Group 5: Future Implications - The article posits that Muddit represents a potential shift in multi-modal foundational models, moving away from language-centric designs towards those that better reflect the structure of the world [33][34]. - This shift could influence future developments in AI, particularly in areas like video, 3D modeling, and embodied intelligence, where understanding spatial and dynamic relationships is crucial [33][35]. - Ultimately, Muddit challenges the prevailing notion that multi-modal models must be built on language, suggesting a rethinking of what should serve as the foundational basis for these systems [40][41].