Workflow
DeepSeek
icon
Search documents
DeepSeek又出手了?一个神秘的AI模型引起全球开发者热议
凤凰网财经· 2026-03-18 13:21
Core Viewpoint - The article discusses the emergence of a new AI model named "Hunter Alpha," which has sparked speculation about its connection to the upcoming DeepSeek V4 model due to its impressive performance metrics and anonymous release [3][4][6]. Group 1: Performance Metrics - Hunter Alpha boasts a parameter scale of 1 trillion, placing it among the leading models in the industry [4]. - The model claims to have a context window of up to 1 million tokens, significantly surpassing most commercial models, allowing it to handle longer texts and more complex tasks [4]. - As of the latest statistics, Hunter Alpha has processed over 160 billion tokens, indicating rapid adoption among developers [5]. Group 2: Connection to DeepSeek - The model's self-identification as a "Chinese AI model trained primarily in Chinese" and its knowledge cutoff date of May 2025 align with the specifications of DeepSeek's existing models [6]. - Some developers suggest that the reasoning style of Hunter Alpha may reveal its "heritage," with its scale and memory capacity matching expectations for DeepSeek V4 [7]. - Despite the similarities, some analysts remain cautious about definitively linking Hunter Alpha to DeepSeek V4, noting differences in token behavior and architectural patterns [9][10]. Group 3: Industry Practices - The anonymous release of AI models for real feedback has become a standard practice in the industry, with platforms like OpenRouter facilitating testing across multiple AI systems [8]. - Notifications on Hunter Alpha's profile indicate that all prompts and completions are recorded for model improvement, a common practice in the field [9].
Wall Street Breakfast Podcast: The AI No One Claims
Seeking Alpha· 2026-03-18 10:55
Jonathan Kitchen/DigitalVision via Getty Images Listen below or on the go via Apple Podcasts and Spotify What is “Hunter Alpha”? AI model fuels talk of new system at DeepSeek. (00:14) lululemon athletica (LULU) beats top- and bottom-line but sets disappointing guidance. (01:39) Amazon (AMZN) plans drastic cut in packages it sends through US Post Office: report. (02:58) This is an abridged transcript. An artificial intelligence model that surfaced anonymously on a developer platform last week is said ...
新共识!特斯拉Optimus V3发布时间
Robot猎场备忘录· 2026-03-18 07:54
温馨提示 : 点击下方图片,查看运营团队最新原创报告(共260页) 说明: 欢迎约稿、刊例合作、行业交流 , 行业交流记得先加入 "机器人头条"知识星球 ,后添加( 微信号:lietou100w )微信; 若有 侵权、改稿请联系编辑运营(微信:li_sir_2020); 正文: 3月,已非Optimus V3发布最佳节点,关注T链们左侧机会,静待右侧机会! 对于T链们而言,3月催化点是Optimus V3亮相,关键点是V3表现超预期;值的注意的是,目前关于Optimus V3 发布时间已有分歧,市场口径多改为3月底或4月初。 马斯克在3月12日最新访谈中, 针对Optimus V3发布及量产问题,也仅表示" Optimus V3已处于最后完成阶 段,今年夏季开始生产,预计明年实现大规模量产 "。 —— 针对V3具体发布时间,目前有不少卖方老师观点是:特斯拉将在3月底发布V3视频,并在4月初举办专项发布 会; 小编视角: 两者可能性皆较小,外围因素影响下,3月已非V3发布最佳节点;同时,停办AI Day以来,特斯 拉未举办单个产品专项发布会且V3定位是自家工厂。 接下来,聊一下本周T链走势和利好T链标的们: 外 ...
DeepSeek又出手了?一个神秘的AI模型引起全球开发者热议
华尔街见闻· 2026-03-18 04:22
性能参数触发市场敏感点 一个拥有万亿参数的神秘免费AI模型突然上线,关于DeepSeek V4将发布的猜测再一次涌现。 据路透社3月18日报道, 一款名为"Hunter Alpha"的AI模型近日在开发者平台OpenRouter匿名上线,引发全球开发者社区关注 。该模型未标注开发者身份, 但观察其性能参数与时间节点,市场猜测这可能是DeepSeek在正式发布前对其下一代系统进行的秘密测试。 Hunter Alpha于3月11日以"隐身模型"形式发布, 且目前向开发者提供免费访问。测试显示,该系统具备1万亿参数规模和高达100万token的上下文窗口。 在测试中,该模型自称"主要以中文训练的中国AI模型",知识截止时间为2025年5月,这一点与DeepSeek现有模型一致。但当被问及开发者时,其回应:"我 只知道自己的名字、参数规模和上下文长度。" OpenRouter平台未披露模型来源,DeepSeek亦未回应置评请求。 数据截止期与推理风格指向DeepSeek 将Hunter Alpha与DeepSeek联系起来的线索,主要来自其底层数据特征和运行逻辑。 在媒体测试中,该聊天机器人自称是"主要用中文训练的中 ...
融资 1200亿后 Kimi 再扔王牌,新架构爆改 Transformer 老配件,比 DeepSeek 同款还省钱
AI前线· 2026-03-17 07:53
作者 | 允毅 连马斯克、Andrej Karpathy 都纷纷点赞,DeepSeek 和 Kimi 前后脚都盯上的 "残差连接" ,到底是 什么? 最近,Kimi 放出一篇重磅新论文,瞄准一个过去十年几乎没人动过的 Transformer 底层根基: 残差 连接(Residual Connection) 。残差连接由何恺明于 2015 年在 ResNet 论文中提出,此后便成为 深度学习领域的标配。 简单来说,可以把大模型的 Transformer 架构,想象成一支几十人排成长队的"传话小组",那么残差 连接就像一条规定:每个工人听完前面所有人的话后,都往里面再补一句,然后原封不动往后传。 这套规则长这样: 但这会带来一个麻烦:队尾的工人收到的话,是前面几十个工人的内容全堆在一起的,越往后话越 乱、越长,前面工人说的重点被埋住了,后面工人加的内容也没人听得清,AI 就变笨了。这叫"稀释 问题"。 于是,Kimi 想到把 "注意力机制" 引进来解决这一问题,它提出一个新的规则: "注意力残 差"(Attention Residuals) 。如同给工人们配备了"智能筛选器",不用再全盘收下前面堆出来的大 杂烩, ...
梁文锋推迟V4,是为了根治龙虾的健忘症?
虎嗅APP· 2026-03-17 00:08
Core Viewpoint - The article discusses the anticipation surrounding the release of DeepSeek's V4, emphasizing the importance of its Long-Term Memory (LTM) feature, which aims to enhance AI's contextual understanding and memory capabilities, setting it apart from competitors like OpenClaw [7][8][17]. Group 1: V4 Development and Features - DeepSeek's V4 is expected to include a significant architectural overhaul with 1 trillion parameters and native multimodal capabilities, set to be released in April [7][8]. - The core innovation of V4 is the Long-Term Memory (LTM) system, which allows the AI to retain user interactions and preferences over time, improving its contextual understanding [8][11]. - The LTM aims to address the limitations of existing models, particularly OpenClaw, which struggles with memory retention and context management [9][10][22]. Group 2: Challenges and Competitor Analysis - The AI industry is rapidly evolving, with competitors releasing new features and models, putting pressure on DeepSeek to catch up [38]. - DeepSeek currently lacks multimodal capabilities, being primarily a text-based model, while competitors have advanced to support audio and video processing [39][43]. - The company faces challenges in agent capabilities, AI programming, and search functionalities, which are critical for maintaining competitiveness in the market [45][48][51]. Group 3: Memory and Learning Capabilities - Current AI models, including OpenClaw, have significant limitations in memory management, leading to issues with context retention and task continuity [18][30]. - Research indicates that many leading models struggle to learn effectively from context, highlighting a gap in their ability to utilize information dynamically [32][34]. - The development of a robust memory system within V4 could potentially transform how AI learns and interacts, making it more adaptable and user-friendly [30][35].
Optimus V2.5集体亮相,V3发布恐要推迟!
Robot猎场备忘录· 2026-03-16 00:02
Core Viewpoint - The article discusses the recent unveiling of multiple Optimus V2.5 robots in Austin, Texas, and anticipates the release of Optimus V3, which is expected to be the most advanced robot in the world, with production starting in summer and large-scale manufacturing anticipated next year [2][3]. Summary by Sections Optimus V2.5 and V3 Release - Multiple units of Optimus V2.5 were showcased in Austin, engaging with the public and demonstrating features like autonomous charging [2] - There is a divergence in market expectations regarding the release date of Optimus V3, with predictions shifting to late March or early April [3] - Elon Musk indicated that Optimus V3 is in the final stages of completion, with production set to begin in summer and large-scale production expected next year [3] Market Reactions and T-Chain Performance - The T-chain market has shown weak performance since March, with a notable downturn except for a brief rally on March 10, attributed to sector rotation rather than official Tesla news [4] - The article highlights that the upcoming Optimus V3 reveal is a key catalyst for the T-chain, with a focus on companies that have signed Power Purchase Agreements (PPAs) [3][4] Notable T-Chain Developments - New core suppliers such as a linear actuator supplier (Z) and a motor supplier (H) have gained attention, indicating a preference for newly confirmed entities in the market [5] - Several T-chain companies are set to embark on new North American tours and are signing PPAs, indicating a tightening focus on core suppliers [6] - Recent developments include a core harmonic reducer supplier (S) and a hand motor supplier (D) making progress with Tesla, with some products already having signed PPAs [8] Future Outlook - The article emphasizes the importance of the V3 production expectations and the ability of T-chain companies to secure their share of the market, with ongoing updates to be provided in the knowledge community [10] - The T-chain is currently viewed as a "pejorative term," with a call to focus on core, reliable suppliers as the market awaits the Optimus V3 reveal [10]
暴力上涨的token背后是裁员
小熊跑的快· 2026-03-15 13:14
Al Model Rankings Based on real usage data from millions of users accessing models through OpenRouter. 00 Top Models Weekly usage of models across OpenRouter 18T 9T 4.5T 2025年3月17日 11月24日 7月21日 | | & LLM Leaderboard | This Week | | --- | --- | --- | | | MiniMax M2.5 | 1.82T tokens | | | by minimax | 10% | | 2. | Step 3.5 Flash (free) | 1.3T tokens | | | by stepfun | 193% | | 3. | Gemini 3 Flash Pre ... | 1.01T tokens | | | by google | J4% | | | DeepSeek V3.2 | 1.01T tokens | | | by deepseek | 125% | | | Cla ...
ByteDance suspends launch of video AI model after copyright disputes, The Information reports
Yahoo Finance· 2026-03-14 16:13
Core Viewpoint - ByteDance has paused the global launch of its AI video generator Seedance 2.0 due to copyright disputes with major Hollywood studios and streaming platforms [1] Group 1: Legal Issues - ByteDance is facing legal threats from U.S. studios, including Disney, regarding unauthorized use of intellectual property in Seedance 2.0 [2] - Disney accused ByteDance of using its characters to train Seedance 2.0 without permission, leading to a cease-and-desist letter [2][3] - ByteDance's legal team is actively working to identify and resolve potential legal issues related to the model [5] Group 2: Product Features and Market Position - Seedance 2.0 is designed for professional film, e-commerce, and advertising use, capable of processing text, images, audio, and video simultaneously to lower content production costs [3] - The model has garnered attention for its ability to generate cinematic storylines, drawing comparisons to competitors like DeepSeek [4] - ByteDance had planned to launch Seedance 2.0 globally in mid-March but has since suspended these plans [4]
英伟达豪掷260亿美元下场造AI模型,直接叫板OpenAI
硬AI· 2026-03-12 09:04
Core Viewpoint - Nvidia is transitioning from a hardware giant to a full-stack AI company by investing $26 billion over the next five years in developing open-source AI models, directly challenging the market positions of OpenAI, Anthropic, and DeepSeek [2][3][4]. Group 1: Investment and Strategic Shift - Nvidia's significant investment of $26 billion has been confirmed by company management, marking a strategic shift towards competing directly with top AI laboratories [3][4]. - The launch of the Nemotron 3 Super model, which boasts 128 billion parameters, signifies Nvidia's commitment to advancing its AI capabilities [6]. Group 2: Model Performance and Benchmarking - The Nemotron 3 Super achieved a score of 37 in the Artificial Intelligence Index, surpassing OpenAI's GPT-OSS score of 33, indicating its competitive performance in the AI model landscape [6]. - Nvidia's model participated in the PinchBench benchmark test, ranking first in evaluating control capabilities, further showcasing its advanced performance [6]. Group 3: Hardware and Software Integration - Nvidia's strategy involves a deep integration of hardware and software, with future AI models designed not only for chip development but also for optimizing supercomputing data center architectures [10]. - The open-source strategy is expected to foster a developer network around Nvidia's hardware ecosystem, enhancing market stickiness for its chips [10]. Group 4: Industry Reception and Significance - The research community has reacted positively to Nvidia's strategic move, with experts highlighting its milestone significance in the open-source AI landscape [12]. - Nvidia's investment is viewed as a historic statement of commitment to openness in AI, positioning the company at the forefront of both open and closed AI projects [12].