DeepSeek
Search documents
Deepseek、智谱、MiniMax,集体宣布上新
21世纪经济报道· 2026-02-11 15:28
Core Insights - The article highlights significant advancements in China's AI model landscape, particularly focusing on the launch of new models like GLM-5 by Zhiyu and Seedance 2.0 by ByteDance, which are set to revolutionize video generation and production capabilities [1][3]. Group 1: AI Model Developments - Zhiyu's new model GLM-5 has topped the popularity charts on OpenRouter and is now available on chat.z.ai [1]. - DeepSeek has updated its model to support a context length of up to 1 million tokens, a significant increase from the previous version's 128,000 tokens [1]. - MiniMax is set to officially launch its M2.5 model, which is currently in beta testing in overseas markets [1]. Group 2: Seedance 2.0 Capabilities - Seedance 2.0 can generate high-quality video content with minimal input, allowing users to create cinematic experiences with simple prompts [3]. - The model demonstrates advanced understanding of context and can synchronize audio and visuals effectively, addressing previous challenges in AI video production [3][4]. - Despite some minor flaws, the potential for ordinary users to create professional-level videos is highlighted, indicating a shift in video production dynamics [4]. Group 3: Industry Implications - The emergence of Seedance 2.0 marks the end of traditional production cost structures, enabling video creation to become more accessible and interactive for the general public [4]. - The rapid decrease in video production costs and increase in efficiency could lead to a content explosion across various sectors, including short films, animation, advertising, and gaming [4]. - Concerns about the blurring lines between reality and fabrication due to the realism of AI-generated content are raised, prompting discussions on the implications for authenticity in media [5].
Deepseek、智谱、MiniMax,集体宣布上新
2 1 Shi Ji Jing Ji Bao Dao· 2026-02-11 15:27
Group 1 - The core message highlights the rapid advancements in AI models, particularly the launch of new models like GLM-5 by Zhiyu and updates from DeepSeek and MiniMax, indicating a competitive landscape in AI technology [1] - Zhiyu's GLM-5 model has gained significant attention by topping the popularity charts on OpenRouter, showcasing its market impact [1] - DeepSeek has expanded its context length support to 1 million tokens, a substantial increase from the previous 128,000 tokens, reflecting the trend towards more complex AI capabilities [1] Group 2 - ByteDance's Seedance 2.0 has gained immense popularity, demonstrating the ability to generate high-quality videos with minimal input, marking a significant leap in AI video technology [2] - The model allows users to create cinematic experiences with simple prompts, indicating a shift towards democratizing video production [2][4] - Seedance 2.0's capabilities in synchronizing audio and visual elements have set a new standard in AI-generated content, addressing previous challenges in the field [2] Group 3 - The evolution of Seedance 2.0 signifies the end of traditional production costs and methods, paving the way for widespread video creation and interaction [4] - The anticipated explosion of content creation will transform industries such as short films, animation, advertising, and gaming, leading to unprecedented levels of content availability [4] - Concerns about the blurring lines between reality and fabrication due to the realism of AI-generated videos have emerged, prompting restrictions on the use of real human faces in content creation [4] Group 4 - The arrival of AI technology signifies a shift in the video production landscape, where the barriers to creating high-quality content are lowered, leading to a more democratized production environment [5] - Industry professionals are urged to reconsider what constitutes a competitive advantage in a world where everyone can produce cinematic-quality videos [6] - The unique aspects of storytelling and aesthetic judgment remain critical, as technology can only replicate human imagination to a certain extent [6]
盖茨谈再次访华:谈中国创新,谈AI愿景,直面“爱泼斯坦争议”
Di Yi Cai Jing· 2026-02-11 14:34
作者 | 第一财经 钱小岩 时隔两年半后再次访华的盖茨,对中国和世界又有了新的认识。 时隔两年半后,盖茨基金会主席比尔·盖茨再次回到中国。 盖茨此次行程首先抵达中国海南,参观、了解中国在农业领域的创新进展,随后抵达上海,关注于传染 病防控和母婴健康,上述三大领域都是盖茨基金会(以下简称"基金会")关注的重点。2月11日傍晚, 盖茨还意外在上海张江现身,出席一场名为"行动创造希望"的活动。 在行程间隙,盖茨接受了第一财经记者的专访。在专访中,盖茨就对华合作、人工智能潜力乃至围绕其 个人声誉的争议,都进行了坦诚和系统的回应。这是盖茨在2023年6月到访北京后的首次访华。在过去 30年中,盖茨到访中国20多次。 始终看好中国创新 "中国在过去三代人的时间里,积累了大量的农业发展和教育改善的经验,其实可以问问你们的祖父 母,在(中国)取得这些成功前,是什么样子的",在专访伊始,盖茨就以此称赞中国在短期内实现现 代化的成绩。 盖茨对第一财经表示,中国正在大幅提升在农业研究方面的投入,对于在海南的参观,他对无融合生殖 杂交水稻印象深刻,"取得了难以置信的成果"。 2019年,美国科学家和中国农科院中国水稻研究所水稻生物学 ...
来了!DeepSeek新模型 | 附体验入口
Xin Lang Cai Jing· 2026-02-11 13:22
Core Insights - DeepSeek has released an updated model, enhancing its capabilities significantly [1][3] Model Enhancements - The context capacity has been upgraded to 1 million tokens from the previous 128,000, allowing for the processing of extensive content such as the entire "Three-Body Problem" trilogy [9][11] - The knowledge base has been updated to May 2025, indicating a new foundational model, potentially referred to as DeepSeek V4 [9][14] Performance Improvements - The frontend and coding capabilities have seen substantial improvements, now comparable to top competitors like Gemini 3 Pro and K2.5 [10][12] - The language style has become more lively and authentic, reducing inaccuracies and enhancing user interaction [10][13] Limitations - The model remains a pure text model and does not support visual understanding, focusing solely on text and voice inputs [14][15]
智谱开源OCR!测完我把手机里的扫描软件都卸了......
量子位· 2026-02-11 12:49
Core Insights - The article discusses the capabilities and performance of the GLM-OCR model, highlighting its competitive edge in the OCR technology landscape, particularly in complex scenarios like handwriting and table recognition [1][39]. Performance Comparison - GLM-OCR outperforms several competitors in various OCR tasks, achieving a document parsing accuracy of 94.6% on OmniDocBench V1.5, surpassing PaddleOCR and others [2]. - In text recognition, GLM-OCR achieves 94.0% accuracy, significantly higher than some competitors like Deepseek-OCR2, which only reaches 34.7% [2]. - For formula recognition, GLM-OCR scores 96.5%, indicating strong performance in recognizing mathematical expressions [2]. - The model also excels in table recognition, with an accuracy of 85.2% on PubTabNet, outperforming many alternatives [2]. Practical Applications - GLM-OCR is particularly effective for structured documents such as Word, PPT, and academic papers, as well as for recognizing clear handwriting, receipts, and scanned contracts [3][4]. - The model demonstrates strong capabilities in recognizing handwritten forms, achieving an accuracy of 86.1% [4]. - It can accurately extract information from various documents, including meeting minutes and whiteboard notes, making it suitable for everyday work scenarios [3][4]. User Experience - Users report a generally positive experience with GLM-OCR in standard document parsing tasks, although challenges remain with unclear handwriting and complex layouts [4][12]. - The model's ability to handle low-quality inputs is commendable, with a recognition accuracy of around 96% for mixed content, although some errors were noted in specific cases [13][29]. Structural Extraction - GLM-OCR is capable of structured information extraction, producing outputs in standard JSON format from various documents, which is beneficial for applications like invoicing and identification [36][38]. - The model's performance in structured extraction improves significantly when clear prompts are provided, indicating its adaptability to user requirements [38]. Industry Trends - The OCR technology market is rapidly evolving, with new models like GLM-OCR emerging to meet increasing demands for efficiency and accuracy [39][40]. - The trend towards smaller model parameters (0.07B to 0.9B) is making deployment easier and more cost-effective for users [51]. - Enhanced output quality and reduced processing times are becoming standard expectations in the OCR industry, benefiting users across various sectors [51].
DeepSeek更新新模型,支持最高1M百万Token上下文长度
Xin Lang Cai Jing· 2026-02-11 11:35
Core Viewpoint - DeepSeek has released a version update that supports a maximum context length of 1 million tokens, but it has not yet enabled multimodal capabilities [1][2]. Group 1: Version Update - The recent update for DeepSeek on both web and app platforms allows for a context length of up to 1 million tokens [1][2]. - As of now, the updated version does not support multimodal capabilities [1][2]. Group 2: Future Developments - Reports suggest that a minor update for the V3 series model is expected to be released around the Spring Festival [1][2]. - The next flagship model from DeepSeek is anticipated to be a trillion-parameter foundational model, but the significant increase in scale has slowed down the training speed, causing delays in the release process [1][2].
DeepSeek新模型来了?
Hua Er Jie Jian Wen· 2026-02-11 11:21
Core Insights - DeepSeek is advancing its new model version with a grayscale test, potentially the final version before the official V4 launch [1] - The V4 model is expected to be released in mid-February 2026, and it will not replicate the global AI computing demand panic seen during the V3 launch [2] - The core value of V4 lies in driving the commercialization of AI applications through underlying architectural innovations rather than disrupting the existing AI value chain [2] Model Enhancements - The context length of the model has been expanded from 128K to 1M, nearly a tenfold increase, and the knowledge base has been updated to May 2025 [1] - V4 is expected to introduce two innovative technologies, mHC and Engram, which aim to overcome computing chip and memory bottlenecks [2][8] - Initial internal tests indicate that V4 outperforms models like Anthropic Claude and OpenAI's GPT series in programming tasks [2] Technical Innovations - mHC (Manifold Constraint Hyperconnection) addresses the bottlenecks in information flow and training instability in deep Transformer models, enhancing the richness and flexibility of communication between neural network layers [4] - Engram is a "conditional memory" module that decouples memory from computation, allowing static knowledge to be stored in a sparse memory table, thus freeing up expensive GPU memory for dynamic calculations [6] Cost Efficiency and Market Impact - The introduction of mHC and Engram is expected to significantly reduce training and inference costs, stimulating downstream application demand and initiating a new cycle of AI infrastructure development [8] - The report suggests that Chinese AI hardware manufacturers may benefit from increased demand and investment due to these cost optimizations [8] Market Dynamics - The market landscape has shifted from a dominant player to a more fragmented competition, with DeepSeek's market share declining as more players enter the field [9][11] - The efficiency in computing management and performance improvements from DeepSeek are accelerating the development of Chinese large language models and applications, altering the global competitive landscape [11] Opportunities for Software Companies - Major global cloud service providers are actively pursuing general artificial intelligence, and the capital expenditure race continues [12] - If V4 can maintain high performance while significantly lowering training and inference costs, it will help developers convert technology into revenue more quickly, alleviating profit pressures [12] - Enhanced capabilities of V4 are expected to create more powerful AI agents, transforming them from mere conversational tools to capable assistants that can handle complex tasks [12]
DeepSeek V4 Is Coming This Month. Why It Could Rattle the Markets, Again.
Yahoo Finance· 2026-02-11 11:20
It was a little more than a year ago that tech stocks fell sharply and briefly due to concerns that an artificial intelligence (AI) chatbot from a Chinese-based company, DeepSeek, could offer significant competition to ChatGPT and other models. While that led to a brief decline for Nvidia (NASDAQ: NVDA) and other tech stocks, they did end up recovering. But the concern around heavy spending on AI continues to weigh on investors' minds these days. And those fears may reach new heights as DeepSeek may be abo ...
DeepSeek更新新模型 可一次性处理超长文本
Xin Lang Cai Jing· 2026-02-11 11:13
Core Insights - DeepSeek has updated its web and app versions to support a maximum context length of 1 million tokens, significantly enhancing its ability to process long texts [1][2] - The previous version, DeepSeek V3.1, had a context length of 128,000 tokens, indicating a substantial improvement in the latest update [1] - DeepSeek successfully processed a document of over 240,000 tokens, demonstrating its capability to recognize and handle extensive content [2] - There are indications that a minor update for the V3 series was expected around the Spring Festival, but the major advancements are still forthcoming [2] - The next flagship model from DeepSeek is anticipated to be a trillion-parameter foundational model, although the increase in scale has slowed down the training speed and delayed the release timeline [2]
春节见?DeepSeek下一代模型:“高性价比”创新架构,助力中国突破“算力芯片和内存”瓶颈
硬AI· 2026-02-11 08:40
Core Viewpoint - Nomura Securities believes that DeepSeek's upcoming next-generation model V4 may further reduce training and inference costs through innovative architectures mHC and Engram technology, accelerating the innovation cycle of China's AI value chain [2][4][5]. Group 1: Innovation in Technology Architecture - The report indicates that computing chips and memory have been bottlenecks for China's large models, and V4 is expected to introduce two key technologies—mHC and Engram—to optimize these constraints from both algorithmic and engineering perspectives [7]. - mHC, or "Manifold Constraint Hyperconnection," aims to address the bottleneck of information flow and training instability in deep Transformer models, enhancing the communication between neural network layers [8]. - Engram is a "conditional memory" module designed to decouple "memory" from "computation," allowing static knowledge to be stored in a sparse memory table, which can be quickly accessed during inference, thus freeing up expensive GPU memory for dynamic calculations [11]. Group 2: Impact on AI Development - The combination of these two technologies is significant for China's AI development, as mHC provides a more stable training process to compensate for potential shortcomings in domestic chips, while Engram smartly manages memory to bypass HBM capacity and bandwidth limitations [13]. - Nomura emphasizes that the most direct commercial impact of V4 will be a further reduction in the training and inference costs of large models, stimulating demand and benefiting Chinese AI hardware companies through an accelerated investment cycle [13][14]. Group 3: Market Dynamics and Competition - Nomura believes that major global cloud service providers are still in a race for general artificial intelligence, and the capital expenditure competition is far from over, suggesting that V4 is unlikely to create the same level of shockwaves in the global AI infrastructure market as last year [15]. - However, global large model and application developers are facing increasing capital expenditure burdens, and if V4 can significantly lower training and inference costs while maintaining high performance, it will serve as a strong boost for these players [15][16]. - The report reviews the market landscape one year after the release of DeepSeek's V3 and R1 models, noting that these models accelerated the development of Chinese LLMs and applications, altering the competitive landscape and increasing attention on open-source models [16]. Group 4: Software Evolution - On the application side, the more powerful and efficient V4 is expected to give rise to more capable AI agents, transitioning from "dialogue tools" to "AI assistants" that can handle complex tasks [20][21]. - This shift will require more frequent interactions with underlying large models, increasing token consumption and thereby raising computing demand [21]. - Consequently, the enhancement of model efficiency is not expected to "kill software," but rather create value for leading software companies that can leverage the capabilities of the new generation of large models to develop disruptive AI-native applications or agents [22].