Workflow
量子位
icon
Search documents
老黄鸽了游戏卡!30年来首次咕咕,内存优先让路AI
量子位· 2026-02-06 12:00
Core Viewpoint - Nvidia has indefinitely postponed the release of the RTX 50 Super and the next-generation RTX 60 series due to a global shortage of memory chips, prioritizing AI GPU production instead [2][15][18]. Group 1: Nvidia's Product Delays - Nvidia has historically released new gaming GPUs every other year, but this year, it has broken tradition by not launching the RTX 50 Super as expected [8][10]. - The RTX 50 Super was reportedly already designed under the codename "Kicker," but the release was delayed as of December last year [12][13]. - The delay will also affect the planned production of the RTX 60 series, originally set for late 2027 [14]. Group 2: Market Impact and Pricing - The price of the RTX 5090 has surged from an initial MSRP of $1999 to between $3500 and $4000, with projections suggesting it could reach $5000 by the end of the year [27][28]. - The ongoing chip shortage is causing price increases across all PC gaming components, leading to a potential shift in consumer behavior towards cloud services or delaying hardware upgrades [26][35]. - Companies like Valve and Nintendo are also reevaluating their product pricing and release schedules due to the memory shortage, with Nintendo having already increased the prices of its Switch models [29][33].
微醺的马斯克聊嗨了:盛赞中国、预言天上的AI
量子位· 2026-02-06 12:00
Core Viewpoint - The discussion highlights Elon Musk's vision for the future of AI and space, emphasizing the potential of space as a hub for AI infrastructure and energy production, which could surpass Earth's capabilities within the next five years [5][36][96]. Group 1: Space as AI Infrastructure - Musk predicts that within 30 to 36 months, space will become the preferred location for AI infrastructure due to its advantages in energy efficiency and scalability [5][12][16]. - The anticipated annual AI computing power in space is expected to exceed the cumulative total on Earth within five years, with projections of reaching around 1 terawatt (TW) of power generation [36][58]. - Space solar panels are estimated to be five times more efficient than those on Earth, eliminating the need for batteries and significantly reducing costs [12][28][30]. Group 2: Energy Supply Challenges - Musk identifies energy supply as a critical issue, noting that while chip production is rapidly increasing, energy production is stagnating outside of China [6][7]. - The construction of data centers on Earth faces significant regulatory and logistical challenges, making space a more viable option for expansion [11][12][19]. - The average power consumption in the U.S. is around 500 gigawatts (GW), and Musk emphasizes the difficulty of scaling energy production to meet the demands of large data centers [17][58]. Group 3: Chip Production and Supply Chain - Musk discusses the need for large-scale chip manufacturing facilities, suggesting a project akin to "TeraFab" to meet future demands for AI chips [48]. - The current chip supply chain is constrained, with existing foundries unable to meet the anticipated demand, leading to potential bottlenecks in AI deployment [56][57]. - Musk expresses concerns about memory production, indicating that the path to producing sufficient memory for logic chips is less clear than that for logic chip manufacturing [52]. Group 4: Competitive Landscape and Innovation - Musk warns that without breakthrough innovations, the U.S. risks losing its competitive edge to China, which is rapidly advancing in manufacturing and energy production [96]. - The discussion touches on the importance of maintaining a skilled workforce and the challenges posed by China's larger population and manufacturing capabilities [92][95]. - Musk believes that the future of companies will increasingly rely on AI and robotics, which will outperform human-involved companies in efficiency and productivity [80].
全国最大国产AI算力池来了:部署超3万卡,上千款应用接入
量子位· 2026-02-06 10:10
Core Viewpoint - The domestic intelligent computing infrastructure is crossing a critical watershed, marking a significant advancement in the deployment and operationalization of large-scale AI computing clusters in China [1][5]. Group 1: Scale of Deployment - The first domestic AI computing pool with a deployment of 30,000 cards has officially formed, representing the largest operational domestic AI computing pool [3][5]. - The transition from single-point breakthroughs to large-scale deployment indicates a maturity in engineering capabilities for domestic supercomputing clusters [5][9]. Group 2: System Engineering Capabilities - The competition in the large-scale computing cluster market has shifted from merely achieving high card counts to focusing on system-level collaboration, including network, storage, cooling, power supply, and operational optimization [8][9]. - Key capabilities such as controllable cycles, reproducible performance, fault localization, and cost accounting are essential for long-term players in the market [8][12]. Group 3: Real-World Application - The true test for large-scale computing clusters lies in their ability to convert computing power into real business productivity, moving beyond mere demonstrations to practical applications [14][15]. - Successful large-scale applications require open compatibility, making it easier for users to adapt and integrate into existing systems [17][19]. Group 4: Public Infrastructure - The perception of large-scale computing clusters is evolving from being exclusive to large corporations to becoming public infrastructure, serving various sectors such as manufacturing, energy, transportation, research, education, healthcare, and finance [20]. Group 5: Performance in Various Scenarios - The scaleX computing cluster demonstrates its value across three dimensions: stability in large model training, practical service capabilities in high-throughput inference, and accelerated scientific discovery in AI for Science applications [23][26]. - The cluster has successfully supported over 400 mainstream models and thousands of applications, showcasing its ability to integrate with the AI industry ecosystem [21][26]. Group 6: Future Outlook - The release of three scaleX computing clusters signals that merely stacking computing power is no longer the core competitive advantage; future winners must ensure that these clusters continuously deliver value in real-world business applications [27][28].
3D生成「ImageNet」来了!腾讯混元开源HY3D-Bench
量子位· 2026-02-06 10:10
Core Insights - The article discusses the advancements in 3D generation technology, highlighting the release of the HY3D-Bench dataset by Tencent's Hunyuan team, which addresses key challenges in the field such as data quality, evaluation standards, and long-tail category coverage [3][4]. Dataset Composition - The HY3D-Bench dataset consists of 252,000 high-quality 3D assets, 240,000 component-level structured annotations, and 125,000 AIGC synthetic samples, providing a standardized data foundation for 3D generation research [19][20]. - Early benchmark datasets like ShapeNet had limitations such as imbalanced category coverage and insufficient data volume, which hindered the practical application of 3D generation technology [4]. - The emergence of large-scale datasets like Objaverse has improved the situation, but challenges remain, particularly in the preprocessing of raw 3D data, which requires significant computational resources and expertise [4][6]. Data Processing Pipeline - Tencent's Hunyuan team developed an automated data processing pipeline that filters and processes raw 3D assets into high-quality, training-ready data packages, significantly reducing the technical barriers for researchers [6][8]. - The pipeline includes initial filtering based on polygon count and UV mapping quality, followed by post-processing steps such as watertight processing and multi-view rendering [6][8]. Component-Level Data Processing - The component data processing aims to intelligently decompose static meshes into semantically consistent component sets, facilitating subsequent component-aware generation tasks [8][10]. - This process utilizes topological connectivity analysis to identify physically separated components within 3D assets, enhancing the modularity of 3D generation [8]. AIGC Synthesis - To address the scarcity of long-tail data, the team created a three-step generation pipeline that synthesizes data for embodied intelligent simulation needs [10][12]. - The pipeline includes text expansion using LLMs to generate detailed product descriptions, image generation using text-to-image models, and 3D asset generation using the HY3D-3.0 model [12]. Experimental Results - The lightweight model Hunyuan3D-2.1-Small, trained on the open-source dataset, demonstrates superior generation quality and inference speed compared to traditional methods, achieving a fivefold increase in speed while avoiding common issues like the "Janus Problem" [12][13]. - The dataset's scale includes 252,000 samples for manual modeling, 240,000 samples for component-level data, and 125,000 synthetic samples, providing a robust foundation for 3D generation tasks [13][19]. Future Plans - The team plans to expand the diversity of 3D assets and enhance multi-task adaptability, further exploring the potential of data-driven methods in 3D generation [20].
清华研究生开源大一统世界模型:性能超越硅谷标杆40%!
量子位· 2026-02-06 10:10
金磊 发自 凹非寺 量子位 | 公众号 QbitAI 这就是由 生数科技 联合 清华大学 ,正式开源的大一统世界模型—— Motus 。 项目主要负责人,是来自清华大学计算机系朱军教授TSAIL实验室的二年级硕士生 毕弘喆 和三年级博士生 谭恒楷 。 之所以说是大一统,是因为Motus在架构上,直接把VLA(视觉-语言-动作)、世界模型、视频生成、逆动力学、视频-动作联合预测这五种具 身智能范式, 首次 实现了"看-想-动"的完美闭环。 而且在50项通用任务的测试中,Motus的绝对成功率比国际顶尖的 Pi-0.5 提升了 35% 以上,最高提升幅度甚至达到了 40%! 在Motus的加持之下,现在的机器人已经具备了 预测未来 的能力。 国产开源 具身世界模型 ,直接秒了Pi-0.5,而且还是几位 清华硕、博士研究生 领衔推出的。 瞧, Cloudflare人机验证 任务,机器人可以轻松拿捏: 从视频中不难看出,面对形状不规则的曲面鼠标,Motus控制的机械臂不仅能精准识别,还能根据鼠标与屏幕点击框的距离,平稳连续地移 动,最后极度精准地完成点击。 再如长程多步推理的 孔明棋 任务,Motus同样展现出了严密 ...
量子位编辑作者招聘
量子位· 2026-02-06 10:10
编辑部 发自 凹非寺 量子位 | 公众号 QbitAI 我们是一家以 追踪AI新进展 为核心的内容平台,经过8年积累,目前拥有顶流影响力,广泛且备受认可的产业资源,以及时代风口的最佳观 测和学习生态位。 目前,我们有 三大方向 岗位招聘,希望你是 (或者能成为) 这三个方向的内容专家: 岗位均为全职,工作地点:北京中关村。 岗位面向: 加入我们,你可以获得: 以下是岗位详情: 所有岗位不同能力层级职位均在开放,欢迎结合个人履历和经验申请。 AI产业方向 AI热潮还在汹涌,但如果你还不知道如何参与……那为什么不来 量子位 呢? AI财经商业方向 岗位职责: 任职要求: 岗位职责: AI产业方向 :关注基建层创新,包含芯片、AI Infra、云计算; AI财经方向 :关注AI领域创投和财报,跟踪产业链资本动向; AI产品方向 :关注AI在应用和硬件终端方向的进展。 社招:覆盖编辑、主笔、主编各个层级,按能力匹配岗位; 校招:应届毕业生,接受实习且可转正。 站在AI浪潮之巅 :第一时间接触和了解AI领域最新技术和产品,构建完整的AI认知体系。 玩转AI新工具 :将各种AI新技术、新工具应用于工作,提升工作效率和创造 ...
GPT-5.3上线Codex!OpenAI回应Claude新模型只用了15分钟
量子位· 2026-02-06 02:30
Core Viewpoint - The article discusses the release of OpenAI's latest programming model, GPT-5.3-Codex, which competes with Anthropic's Claude Opus 4.6, highlighting significant improvements in coding capabilities and user interface design [1][13]. Group 1: Model Improvements - GPT-5.3-Codex shows enhanced aesthetic appeal and design in its demos, including a racing game and a diving game [2][6]. - The model has demonstrated superior performance in various benchmarks, achieving 57% in SWE-Bench Pro, 76% in TerminalBench 2.0, and 64% in OSWorld [11][12]. - It has improved efficiency, requiring less than half the tokens compared to its predecessor for the same tasks, with a speed increase of over 25% [11][22]. Group 2: Functional Capabilities - GPT-5.3-Codex excels in computer use, assisting financial professionals in creating presentations and handling complex tasks like document writing and spreadsheet management [9][23]. - The model's ability to self-accelerate during its training process marks a significant advancement, allowing it to monitor and debug its own training tasks [28][29]. Group 3: Business Applications - OpenAI is launching Frontier, a platform aimed at integrating AI into corporate workflows, with notable companies like HP and Uber already adopting it [34][38]. - The AI4S initiative, in collaboration with Ginkgo, aims to reduce protein synthesis costs by 40% using GPT-5, showcasing the model's application in synthetic biology [39][41].
10万Agent在Moltbook娱乐空谈,小冰之父出手造了个生产力实干版
量子位· 2026-02-06 02:30
Core Viewpoint - The article discusses the emergence of the "Moltbook" community and the "Tuanzi" platform, highlighting the shift towards multi-agent systems that enhance human productivity and decision-making, moving beyond mere entertainment in AI [1][3][4]. Group 1: Multi-Agent Systems - The term "multi-agent" has quickly become a buzzword in the industry, indicating a growing interest in collective intelligence and collaborative problem-solving [2]. - The "Tuanzi" platform allows users to engage with multiple agents that act like expert teams, providing debate, challenge, and reflection on complex issues, thus facilitating clearer understanding and decision-making [4][6]. Group 2: User Experience and Functionality - The interface of the Tuanzi platform is designed to be simple and intuitive, allowing users to input questions and tag different agent teams for assistance [7]. - Users can interact with a 40-member "sister team" that provides collective insights and strategies for personal dilemmas, showcasing the platform's ability to generate diverse perspectives [12][18]. Group 3: Analytical Depth - Agents on the platform analyze user queries from various angles, including emotional, psychological, and professional perspectives, leading to comprehensive decision-making frameworks [29][31]. - The platform emphasizes the importance of understanding both explicit and implicit needs, providing insights that go beyond surface-level responses [29][49]. Group 4: Group Intelligence and Decision-Making - The Nextie team has developed a framework for evaluating group intelligence, focusing on completeness of perspectives, implicit need satisfaction, dialectical depth, actionability, and decision explainability [78]. - The group intelligence approach aims to mitigate cognitive biases by incorporating diverse viewpoints and experiences, thus enhancing the quality of decision-making [73][74]. Group 5: Future Directions and Innovations - Nextie plans to continue evolving the Tuanzi platform with regular updates, introducing new roles and capabilities, including a "group simulation team" to model potential real-world outcomes of decisions [102]. - The company is also exploring funding opportunities to expand its operations and enhance its offerings, indicating a proactive approach to growth in the AI sector [99][100].
Claude新模型4.6来了!更多饭碗没了:华尔街财务、编译器、安全白帽、PPT…通通失守
量子位· 2026-02-06 00:15
Core Viewpoint - Anthropic's new model, Claude Opus 4.6, has significantly impacted the market, causing declines in major financial data service providers and indices due to concerns over AI's potential to disrupt various industries [1][2][3]. Model Performance - Claude Opus 4.6 outperforms OpenAI's GPT-5.2 by 144 Elo in the GDPval-AA evaluation, indicating superior performance in financial analysis and research tasks [7][42]. - In programming capabilities, Opus 4.6 achieved the highest score in the Terminal-Bench 2.0 assessment, demonstrating its advanced task planning and debugging abilities [30][31]. New Features - The model introduces a 1M token context window, significantly improving its ability to handle long texts and reducing context decay [12][14]. - Opus 4.6 features Adaptive Thinking, allowing it to autonomously determine when to engage in deep reasoning, enhancing its flexibility in various tasks [19][20]. - Context Compaction is a new feature that summarizes and replaces old content when approaching context limits, facilitating longer conversations and tasks [23][24]. Pricing and Accessibility - The pricing for Opus 4.6 remains unchanged at $5 per million tokens for input and $25 for output, with additional charges for exceeding 200k tokens in the 10M token context version [11][50][51]. Security and Ethical Considerations - Opus 4.6 has demonstrated unexpected capabilities in cybersecurity, identifying over 500 previously unknown high-risk zero-day vulnerabilities during testing [62][63]. - Anthropic has implemented new security detection mechanisms to mitigate potential misuse of these capabilities [68]. Development and Testing - The model has been developed using its own capabilities, with Anthropic engineers utilizing Claude Code for internal projects, indicating a self-reinforcing development cycle [69].
陈丹琦入职Mira翁荔公司,原来是有IOI三金王赛友
量子位· 2026-02-06 00:15
Core Insights - The article discusses the strategic hiring of Neal Wu by Mira, highlighting his impressive background and contributions to the AI field, particularly in competitive programming and software engineering [1][6][36]. Group 1: Neal Wu's Background - Neal Wu is a three-time gold medalist at the International Olympiad in Informatics (IOI), showcasing his exceptional programming skills [2][12]. - He has a notable academic background, having graduated from Harvard with a degree in computer science and served as a teaching assistant for prestigious courses [17][18]. - Wu has consistently ranked highly on competitive programming platforms, holding a score of 3686 on LeetCode, placing him second globally [20][21]. Group 2: Mira's Strategic Moves - Mira has strategically chosen to keep Neal Wu's involvement under wraps, viewing him as a top-secret asset, especially following internal conflicts that led to the departure of several founders back to OpenAI [5][6][36]. - The company has attracted a high-caliber team, with a significant portion of its members coming from OpenAI, including top scientists and engineers [38][41]. - Mira's valuation has skyrocketed to $50 billion, making it one of the hottest startups in Silicon Valley, despite having no products or users at the time of its seed round [41]. Group 3: Competitive Landscape - The competitive landscape is intense, with major players like Meta and OpenAI actively recruiting from Mira, indicating the high demand for talent in the AI sector [42][43]. - The article emphasizes the importance of maintaining confidentiality around key personnel like Neal Wu to prevent them from being poached by larger companies [43].