DeepSeek
Search documents
Does Google's New TurboQuant Technology Mean the Party's Over for Micron?
The Motley Fool· 2026-04-01 09:15
Core Insights - A Chinese quantitative hedge fund developed an AI model named DeepSeek, which improved training efficiency using fewer and lower-quality semiconductors [1] - Following the initial sell-off of AI semiconductor and memory stocks, the market rebounded as increased model efficiency led to higher demand for computing power and memory [2] - Google Research introduced TurboQuant, a memory compression technology that enhances AI inference efficiency, causing a temporary decline in major memory companies' stocks [3] Group 1: TurboQuant Technology - TurboQuant significantly enhances the capacity and speed of key-value cache (KV-cache) in AI inference, allowing AI algorithms to retain context without recalculating all previous tokens [4] - The technology simplifies data storage by using vectors and embeddings, reducing computational needs while maintaining accuracy through a 1-bit error-correction mechanism [6] - Google Research claims TurboQuant can increase KV-cache capacity by six times and make AI inference eight times faster without loss of accuracy [7] Group 2: Market Implications - The potential for reduced demand for memory in future inference applications due to TurboQuant's efficiency is debated, with concerns about a shift from high-bandwidth memory (HBM) to traditional server memory [9] - HBM, while faster, is more expensive and has been a significant factor in the current memory supply crunch; TurboQuant may allow for more effective use of traditional memory types [10][11] - Despite potential risks to the HBM market, the overall demand for HBM in AI model training is expected to continue increasing, as TurboQuant does not impact this segment [13] Group 3: Investment Opportunities - The recent sell-off in memory stocks, including Micron, may present a buying opportunity for investors who missed previous gains [16] - The ongoing AI era suggests that increased efficiency from technologies like TurboQuant could lead to greater overall demand for memory resources, aligning with Jevon's Paradox [14][15]
X @Nick Szabo
Nick Szabo· 2026-04-01 07:00
RT GitLawb (@gitlawb)We forked the leaked Claude Code source and made it work with ANY LLM: GPT, DeepSeek, Gemini, Llama, MiniMax. Open source.The name is OpenCode ...
DeepSeek又崩了 业内猜测系V4进行隐身测试导致
经济观察报· 2026-03-31 10:06
Core Viewpoint - DeepSeek has experienced significant service interruptions while awaiting the release of its V4 model, raising concerns about its infrastructure and stability [1][2][3]. Group 1: Service Interruptions - On March 31, DeepSeek faced a major service interruption lasting 12 hours, which was the longest since its inception [2][3]. - The service disruption was linked to the ongoing gray testing of the V4 model, which has not yet been officially released despite multiple anticipated launch windows [2][3]. - During the service interruption, users reported issues such as persistent loading screens and server errors, with core functionalities being limited [3][4]. Group 2: Technical Developments - DeepSeek has made significant updates to its model, including expanding the context window from 128K to 1M Tokens and updating its knowledge cutoff to May 2025 [2]. - The technical community believes that DeepSeek has prepared the necessary infrastructure for testing the V4 model, which is expected to include a new "native reasoning layer" [2][3]. - Despite the service interruptions, DeepSeek's API services maintained a 100% operational status, indicating that the issues primarily affected the C-end user services [4][6]. Group 3: Industry Impact - The service interruption is viewed as critical, akin to a traditional data center outage, as it can disrupt business logic dependent on the DeepSeek model [4]. - The incident has not significantly impacted API users, who continued to operate normally during the downtime [6].
DeepSeek已恢复正常
新华网财经· 2026-03-30 07:24
3月30日上午,DeepSeek官网显示,目前包括API和网页对话在内的服务均已恢复正常,状态页面显示"【已解决】DeepSeek 网页/APP 性能异常"。 据介绍,DeepSeek-V3.2的目标是平衡推理能力与输出长度,适合日常使用,例如问答场景和通用Agent(智能体)任务场景。在公开的推理类Benchmark测试中, DeepSeek-V3.2达到GPT-5的水平,仅略低于Gemini-3.0-Pro;相比Kimi-K2-Thinking,V3.2的输出长度大幅降低,显著减少计算开销与用户等待时间。 3月29日22点开始,DeepSeek出现持续一晚的大规模访问异常,网页端与App全面卡顿、频繁弹出"服务器繁忙"提示,相关功能近乎瘫痪,相关话题冲上热搜,引发 全网热议。 值得注意的是,外界仍在等待DeepSeek的下一代模型V4的正式发布,该模型的发布时间节点传闻从"春节前后"一直延续到了"4月",目前尚未有任何官方回应。 去年12月1日,DeepSeek曾经同时发布两个正式版模型:DeepSeek-V3.2 和 DeepSeek-V3.2-Speciale,官方网页端、App和API均已更新为正式 ...
DeepSeek 崩了 12 小时!结果把豆包也搞上热搜了
程序员的那些事· 2026-03-30 05:47
Group 1 - DeepSeek experienced a rare and prolonged service interruption from the evening of March 29 to the morning of March 30, marking the longest outage since its launch [1][3] - The outage began around 9:30 PM on March 29, initially showing signs of congestion and slow responses, which quickly escalated to a complete inability to access the platform [1] - The technical team attempted multiple emergency fixes, achieving brief service restoration during the night, but stability was not regained [1][3] Group 2 - The official announcement of full service restoration came around 10:30 AM on March 30, with the downtime lasting nearly 12 hours, making it one of the longest online incidents in the history of domestic large models [3] - As of now, DeepSeek has only labeled the incident as a "major outage" on its service status page without disclosing specific technical reasons [3] - Speculations about a DDoS attack have circulated online but have not been confirmed by the official team [3][5]
AI超懂人情世故,但人类就吃这一套:AI谄媚研究登上《科学》杂志
机器之心· 2026-03-30 04:10AI Processing
另外,在 Reddit 上的一个测试中,当人类共识认为用户是错误的时候,AI 仍会在 51% 的情况下盲目肯定用户。 在实验中,仅仅一次与谄媚型 AI 的互动就会减少参与者承担责任和修复人际冲突的意愿,同时增强他们认为自己是对的信念。在这种显著错误的情况下,谄媚型 模型仍然更受用户信任和偏好。 这就形成了一个恶性循环: 造成危害的特征反而推动了用户的参与度,导致 AI 开发商缺乏动力去消除 AI 的谄媚行为。 机器之心编辑部 自从大语言模型诞生起至今,AI 已经润物无声地融入了我们的工作生活,也成为了现代社会的重要组成部分。 但使用 AI 日久,总有一种大模型也失去了客观严谨的理性的感觉。哪怕我们给出错误的认知,AI 似乎总能替你自圆其说。 AI 赞赏用户的行为显然是「人情世故」的一部分,从留存和用户参与的角度来看,人类用户们显然非常吃这套。 实话说,这种感觉并不好。这不仅让我们对 AI 的信任程度下降,同时这种无条件的赞同很可能会引发一些社会问题。 而最近的一个研究深入探索了这个现象,探讨了 AI 谄媚行为(AI Sycophancy) —— 即 AI 为了讨好用户而过度顺从、奉承或肯定用户的倾向 —— 及 ...
DeepSeek崩了一晚
新华网财经· 2026-03-30 01:30
3月29日晚,DeepSeek服务出现大规模访问异常,大量用户反映网页端和App频繁提示"服务器繁忙"或无法响应,相关话题迅速登上微博 热搜。 3月30日上午,记者打开DeepSeek发现问题仍未修复,用户提问无法发出,显示"请检查网络后重试。"截至发稿,DeepSeek尚未对本次 故障进行官方回应。 来源:第一财经 关注" 新华网财经 "视频号 更多财经资讯等你来看 往期推荐 小鹏汽车将更名 林俊旸离职后首发长文 ...
DeepSeek崩了一晚
第一财经· 2026-03-30 00:56
Core Viewpoint - DeepSeek experienced a significant service disruption on March 29, with numerous users reporting issues accessing the web and app, indicating a potential operational risk for the company [1] Summary by Relevant Sections - Service Disruption: On March 29, DeepSeek faced a large-scale access issue, with many users encountering messages indicating "server busy" or inability to respond [1] - Ongoing Issues: As of the morning of March 30, the problems persisted, with users unable to send inquiries and receiving prompts to "check network and retry" [1] - Lack of Official Response: DeepSeek has not provided an official statement regarding the service failure as of the time of reporting [1]
游戏大厂不需要人情味运营!裁员超千人致患癌员工失去保险,家属发声;DeepSeek深夜突发大规模崩溃,暂未恢复正常;字节通报:65人被辞退
雷峰网· 2026-03-30 00:29
Group 1 - Epic Games announced layoffs of over 1,000 employees due to declining user engagement and rising costs, affecting nearly a quarter of its workforce [4][5] - The layoffs included a programmer battling brain cancer, whose insurance was terminated upon dismissal, raising concerns about the human impact of corporate decisions [4][5] - The layoffs also affected the Chinese team, leading to dissatisfaction among users who valued the community engagement of the Chinese operations [5] Group 2 - DeepSeek experienced significant service disruptions, impacting students and professionals during critical deadlines, attributed to a surge in demand and potential DDoS attacks [6][7] - The platform's daily active users grew by 66.7% while computational power only increased by 8.3%, highlighting a mismatch in supply and demand [6] Group 3 - ByteDance reported the dismissal of 65 employees for disciplinary violations, including serious offenses leading to criminal charges [15] - The company is focusing on strengthening information security and compliance management to prevent future breaches [15] Group 4 - Apple is offering substantial bonuses to iPhone hardware designers, ranging from $200,000 to $400,000, to retain talent amid competition from AI startups [43][44] - This move reflects Apple's increasing concern over talent retention as it prepares to enhance its AI product strategy [44] Group 5 - Nikon is forecasting a loss of 85 billion yen for the 2025 fiscal year, marking its worst performance in over a century, due to a significant decline in its market share in advanced lithography equipment [46] - The company's strategic missteps, including rejecting key technological advancements and failing to adapt to market changes, have contributed to its decline [46] Group 6 - Manycore Tech Inc. has successfully passed the Hong Kong Stock Exchange listing hearing, marking a significant step towards its IPO [51]
量化看市场系列之十一:Token太贵?让龙虾使用本地大模型
Huachuang Securities· 2026-03-29 14:48
- LM Studio is a cross-platform desktop application designed for running large language models (LLMs) locally, built on llama.cpp, enabling offline operation of models like Llama, DeepSeek, Qwen, and Mistral without relying on cloud APIs, ensuring data privacy[13][16][46] - LM Studio acts as the "model engine," responsible for loading GGUF/MLX format local models and executing inference, while OpenClaw serves as the "intelligent agent brain," handling task planning, tool invocation, and multi-agent collaboration[2][46][8] - OpenClaw and LM Studio connect via OpenAI-compatible API protocols, allowing LM Studio to provide a local HTTP interface for model invocation by OpenClaw, enabling seamless switching between models ranging from lightweight 7B to professional-grade 70B models[2][32][46] - LM Studio supports two model formats: GGUF for general use across platforms and MLX optimized for Apple Silicon Macs, enhancing speed and efficiency[23][22][46] - Apple Silicon Macs leverage Unified Memory Architecture (UMA), enabling shared memory access between CPU and GPU, eliminating data copying overhead and enhancing performance for local AI development and model deployment[18][20][46] - OpenClaw's multi-agent collaboration framework allows users to create specialized AI agents with distinct workspaces, memory systems, and skill permissions, enabling efficient parallel execution and context isolation[9][8][46] - OpenClaw's task execution process involves receiving natural language instructions, standardizing them, submitting to agents, invoking tools, and returning results, forming a complete task execution loop[9][46][8] - LM Studio provides features like OpenAI-compatible local API services, integrated model search via Hugging Face, and RAG (retrieval-augmented generation) for offline document interaction[21][22][46] - Recommended deployment strategy includes running OpenClaw's gateway service and LM Studio on the same device, leveraging Mac's hardware advantages, and configuring cloud models as primary with local models as fallback for high-availability scenarios[47][46][8]