量子位

Search documents
 Cursor发布首个编程大模型!代码生成250tokens/秒,强化学习+MoE架构
 量子位· 2025-10-30 01:06
 Core Insights - Cursor has officially released its first in-house coding model, named Composer, as part of the Cursor 2.0 update [1][2] - Composer is reported to complete complex tasks in just 30 seconds, achieving a speed increase of 400% compared to competitors [3][12]   Model Features - The new Cursor 2.0 includes a native browser tool that allows the model to test, debug, and iterate code autonomously until achieving correct results [4] - Voice code generation enables users to convert their thoughts into code without typing [5] - The interface has shifted from a file-centric to an agent-centric model, allowing multiple agents to run simultaneously without interference [6][7]   Performance Metrics - Composer generates code at a speed of 250 tokens per second, which is approximately twice as fast as the current leading models like GPT-5 and Claude Sonnet 4.5 [19][20] - The model demonstrates enhanced reasoning and task generalization capabilities, comparable to mid-tier leading models [21]   Training Methodology - Composer's performance is attributed to reinforcement learning, which allows the model to learn from real programming tasks rather than static datasets [22][26] - The training process involves the model working directly within a complete codebase, utilizing production-level tools to write, test, and debug code [27][28]   Practical Application - Cursor 2.0 is designed to provide a practical AI system that aligns closely with developers' daily workflows, enhancing its usability in real-world scenarios [35][36] - The model has shown emergent behaviors, such as running unit tests and autonomously fixing code format errors [31]   Transparency and Model Origin - There are concerns regarding the transparency of Composer's foundational model, with questions about whether it is based on pre-existing models or entirely self-trained [37][40] - Cursor has previously developed an internal model named Cheetah, which was used for testing speed and system integration [42]
 Sora连更三大新功能!一键打造IP形象,限时免注册码抢占安卓市场
 量子位· 2025-10-30 01:06
 Core Insights - Sora has introduced three major new features: Character Cameo, video stitching, and community leaderboard [1][12][13] - The app has temporarily lifted the invitation code requirement in the US, Canada, Japan, and South Korea to facilitate direct registration [2][17] - The motivation behind the limited-time opening is attributed to insufficient computing power [3]   Feature Summaries - **Character Cameo**: This upgraded feature allows users to maintain consistency with non-human cameo characters, including pets or animated figures, enhancing user engagement [6][9] - **Video Stitching**: Users can now combine two videos if they find the generated content too short, increasing the versatility of video creation [12] - **Community Leaderboard**: This feature categorizes the most used cameo characters and the most remixed videos, fostering community interaction [13]   Market Strategy - The temporary removal of the invitation code requirement coincides with the launch of Sora's Android version, aiming to rapidly expand the user base and capture market share [18] - Initially, Sora employed a viral marketing strategy where each activated account could share four invitation codes, creating significant buzz but also a gray market for codes [15][16]
 再创历史!英伟达市值一夜突破5万亿美元,今年涨幅56%,黄仁勋晋升全球第八富豪
 量子位· 2025-10-30 01:06
但不断创纪录的英伟达,相比之下创始人、CEO黄仁勋就"委屈"一些了,即便不断创造股价和市值神话,但 老黄的个人财富水涨船高之后, 也才以1792亿美元的持股价值跃居福布斯全球富豪榜第八 。 衡宇 发自 凹非寺 量子位 | 公众号 QbitAI 见证历史,英伟达市值突破5万亿美元大关! 轻轻松松 成为全球第一家市值突破5万亿美元的公司 。而且英伟达的纪录英伟达破,当初率先突破4万亿美元的也是英伟达——追赶英伟达的 只有英伟达了。 10月29日,美股开盘不久,英伟达股价一度上涨5.44%,盘中触及212.19美元/股,收盘价稳定在207.04美元/股。 随之而来的,是一个足以载入科技史的新数字: 5.03万亿美元 。 还有一个醒目的数字值得关注。 从2025开年至今10个月时间里,英伟达股价已暴涨了56% ——放眼全球,还有什么标的有这样体量巨大又 增长迅猛的案例?还有谁? 根据道琼斯市场数据,英伟达现在的市值已经超过了AMD、Arm Holdings、ASML、博通、英特尔、泛林集团、美光科技、高通和台积电的 总和。 与此同时,其市值还超过了标普500指数中的整个行业板块,包括公用事业、工业和消费品。 相比特斯 ...
 14000人原地被裁!亚马逊今日:打工人水深,AI机器人火热
 量子位· 2025-10-29 09:30
Jay 发自 凹非寺 量子位 | 公众号 QbitAI 亚马逊前几天宣布的「3万人血裁计划」,正式开刀了。 刚刚, 1.4万名亚马逊员工被宣布没了工作 ,而另一边,公司正在开足马力部署AI和机器人上岗。 流程还没开始走,电脑账号直接锁死,很多人甚至还来不及备份自己的文件。 更狠的是见谁砍谁,管你业绩怎么样。 3年内从L4升到L6,本以为自己是高质量人类,这下看来我也只是个可被AI取代的螺丝钉…… 我立刻就失去了对所有东西的访问权限 :( 离谱的是,消息一出, 亚马逊股票当天立即上涨了1% 。 部分老员工苦笑道: 如果裁员能继续推高股价,那我的股票收入就算是失业金了。 这边亚马逊估计更是开心坏了,不仅能少开一大笔工资,还有投资人送钱。 跟量子位来看看,这到底是怎么一回事。 1.4万人被裁 10月28日,约1.4万名亚马逊员工在一封「致员工的信」中收到了噩耗。 亚马逊高级副总裁Beth Galetti在信中遗憾地宣布,亚马逊将启动新一轮裁员——在35万名公司全职员工中, 约有4%的同事要准备收拾行李 走人 。 值得注意的是,这1.4万是净裁员人数——那些通过内部调岗成功「自救」的员工,并不算在内。 据悉,本轮裁员 ...
 全球首个具身智能开放平台来了!让大模型长出“身体”,像人一样自然表达交互
 量子位· 2025-10-29 09:30
henry 发自 凹非寺 量子位 | 公众号 QbitAI 具身智能赛道的想象力,远比眼前的机器人要辽阔。 当大家还在琢磨怎么把大模型塞进机器人里时, 数字人 也和 具身智能 关联上了。 就在今天, 魔珐科技 发布了面向开发者的具身智能基础设施—— 「魔珐星云」具身智能3D数字人开放平台 。 这也是全球首个。 在魔珐星云的驱动下,不仅大语言模型能够"长出身体",实体机器人也能像人一样拥有动作、表情,实现自然表达。 它可以根据文本,实时生成3D数字人的语音、表情、眼神、手势和身体动作,让任何屏幕、应用、终端都实现自然、流畅的多模态交互。 魔珐星云主要驱动3D具身数字人的三大应用方向。 首先,魔珐星云可以 为大模型和AI智能体提供身体和表达能力 ,让原本只能文字交流的模型,通过语音、表情和动作与人类进行自然互动。 其次,它可以 让手机、平板、电视、车载屏幕等各种终端升级为具身智能界面 ,让每一块屏幕都能"能说、会动",从被动的信息载体转变为 主动的服务者或信息提供者。 最后,魔珐星云还能 驱动人形机器人实现自然沟通 。 凭借低于 1.5秒 的端到端延迟、 千万级 并发能力,以及 百元级算力 即可运行的架构,人机对 ...
 不好美国要捧杀了!新研究:中国正在成为全球科学领导者
 量子位· 2025-10-29 09:30
 Core Viewpoint - The article discusses a recent study published in the Proceedings of the National Academy of Sciences, which indicates that China is emerging as a global leader in science, particularly in collaboration with the United States [2][4].   Group 1: Research Findings - The study analyzed 6 million papers using machine learning to assess the leadership roles of Chinese scientists in international collaborations, revealing that as of 2023, Chinese leaders in US-China collaborations have increased to 45% and are expected to reach parity by 2027-2028 [4][21]. - By 2030, China is projected to achieve equal leadership status with the US in strategic fields such as AI, semiconductors, energy, and materials science [5][6].   Group 2: Methodology - The research employed a three-step approach to quantify "leadership" in scientific collaborations, defining leadership roles and scoring scientists based on nine predictive features [12][18]. - The nine dimensions for scoring included past citation counts, research overlap with keywords, self-citation rates, years of academic experience, total publications, cumulative citations, unique keyword counts, author order, and institutional academic ranking [15][17].   Group 3: Implications - The findings suggest a significant shift in the global scientific leadership landscape, with China rapidly increasing its share of leadership roles in international collaborations [20][21]. - The study's results have sparked discussions in the West about the potential decline of Western dominance in science, as exemplified by recent funding issues faced by prominent scientists in the US [26][27].
 人工智能年度榜单火热报名中!五大奖项,寻找AI+时代的先锋力量
 量子位· 2025-10-29 09:30
组委会 发自 凹非寺 量子位|公众号 QbitAI 为了让更多从业者感受智能浪潮的跃迁,也为了给予更多同行同路人掌声与鼓舞,我们将正式启动 「2025人工智能年度榜单」评选报名 。 本次评选将从 企业 、 产品 、 人物 三大维度,设立五类奖项。欢迎企业踊跃报名! 让我们共同见证年度之星,点亮未来的方向。 企业榜 产品榜 详细评选标准及报名方式如下。 2025 人工智能年度领航企业 将面向中国人工智能领域,评选出最具综合实力的企业, 参选条件 : 2025 人工智能年度 领航企业 2025 人工智能年度 潜力创业公司 2025 人工智能年度 杰出产品 2025 人工智能年度 杰出解决方案 1、注册地在中国,或主营业务主要面向中国市场; 2、主营业务属于人工智能及相关产业,或已将人工智能广泛应用于主营业务,并在细分领域居于行业领先地位; 评选标准 : 人物榜 2025 人工智能年度 焦点人物 2025 人工智能年度潜力创业公司 聚焦于中国人工智能领域创新创业力量,将评选出最具投资价值和发展潜力的AI创业公司, 参选条件 : 评选标准 : 3、具备成熟的产品或服务,已获得实际客户应用及市场认可; 4、近一年在技术 ...
 阿里新研究:统一了VLA和世界模型
 量子位· 2025-10-29 09:30
 Core Insights - WorldVLA is a unified framework that integrates Visual Language Action Models (VLA) with World Models, proposed by Alibaba DAMO Academy, Lake Lab, and Zhejiang University [1][4] - Experimental results indicate that WorldVLA significantly outperforms independent action models and world models, showcasing a mutual enhancement effect [2]   Model Overview - The framework combines three independent tokenizers for encoding images, text, and actions, utilizing a VQ-GAN model for image tokenization with a compression ratio of 16 and a codebook size of 8192 [8] - The action tokenizer discretizes continuous robot actions into 256 intervals, representing actions with 7 tokens [8]   Model Design - WorldVLA employs a self-regressive action world model to unify action and image understanding and generation [4] - The model addresses limitations of existing VLA and world models by enhancing action generation accuracy through environmental physical understanding [5][14]   Training and Performance - WorldVLA is jointly trained by integrating data from both action models and world models, enhancing action generation capabilities [13] - The model's performance is positively correlated with image resolution, with 512x512 pixel resolution showing significant improvements over 256x256 [21][23]   Benchmark Results - WorldVLA demonstrates superior performance compared to discrete OpenVLA models, even without pre-training, validating its architectural design [19] - The model's ability to generate coherent and physically plausible states in various scenarios is highlighted, outperforming pure world models [31][32]   Mutual Enhancement - The world model enhances the action model's performance by predicting environmental state changes based on current actions, crucial for tasks requiring precision [25] - Conversely, the action model improves the visual understanding of the world model, supporting better visual generation [17][30]
 美国AI公司们,开始青睐Made in China的大模型
 量子位· 2025-10-29 08:00
 Core Viewpoint - The article discusses the increasing adoption of Chinese AI models, such as GLM and Qwen3, by American companies, highlighting a shift towards cost-effective and efficient solutions in the AI industry [1][14][44]   Group 1: Adoption of Chinese AI Models - Windsurf, a leading AI programming product, recently integrated a mysterious model that turned out to be GLM from China [2][7] - Vercel, a company valued at $9.3 billion, announced a partnership with Zhipu to provide GLM-4.6 API services, indicating a trend of American companies utilizing Chinese models [17][19] - Other platforms, such as Featherless, have also begun supporting Chinese models, showcasing a broader acceptance in the AI landscape [22][24]   Group 2: Reasons for Adoption - The primary reasons for the shift towards Chinese models are performance and cost-effectiveness, with many companies finding that Chinese models can deliver comparable or superior performance at a lower price [26][27] - Chamath Palihapitiya, founder of Social Capital, noted that while models from OpenAI and Anthropic are good, they are too expensive, making Chinese models a more viable option for scaling businesses [30][34] - The competitive pricing strategies of Chinese AI companies, such as offering significant token allocations and discounts, further enhance their attractiveness to American firms [36][39]   Group 3: Industry Implications - The trend indicates a transition in the AI industry from a focus on technical superiority to practical applications, where cost, speed, and scalability are paramount [40][41] - The choices made by companies like Vercel and Social Capital challenge the notion that only the most powerful models are suitable for commercial use, emphasizing the importance of high cost-performance ratios [42][44] - This shift may signal the onset of a more diverse and competitive global AI landscape, where the value of Chinese models continues to rise [47]
 单条演示即可抓取一切:北大团队突破通用抓取,适配所有灵巧手本体
 量子位· 2025-10-29 05:11
 Core Insights - The article discusses the challenges of traditional reinforcement learning (RL) in high-dimensional action spaces for robotic grasping tasks and introduces the DemoGrasp framework as a solution [1][2][4].   Group 1: DemoGrasp Framework - DemoGrasp is a simple and efficient learning method for general robotic grasping, initiated from a single successful demonstration trajectory [2][4]. - The framework transforms multi-step Markov Decision Processes (MDP) into a single-step MDP by editing demonstration trajectories, enhancing learning efficiency and performance transfer to real robots [4][7].   Group 2: Learning Process - The learning process involves editing the robot's actions in the demonstration trajectory to adapt to different objects and poses, focusing on wrist and finger adjustments [9][16]. - DemoGrasp utilizes a simulation environment with thousands of parallel worlds to train the policy network, which outputs editing parameters based on observations [10][11].   Group 3: Training Efficiency - The training efficiency is notable, with a single RTX 4090 GPU achieving over 90% success rate in just 24 hours on a compact action space [12]. - The framework can adapt to various robotic hands without adjusting training hyperparameters, achieving an average success rate of 84.6% across 175 objects [20].   Group 4: Performance Metrics - DemoGrasp outperforms existing methods in the DexGraspNet dataset, achieving a visual policy success rate of 92% with minimal generalization gap [17][18]. - In real-world tests, DemoGrasp successfully grasped 110 unseen objects, maintaining over 90% success rates for regular objects and 70% for challenging flat and small objects [21][22].   Group 5: Future Directions - The framework aims to support more complex tasks such as functional grasping and tool usage, with potential for real-time adjustments and error recovery in future research [25][26]. - DemoGrasp can integrate with multimodal large models for autonomous grasping in open environments [27].










