Workflow
Seek .(SKLTY)
icon
Search documents
早盘消息0820| T 链 Gen3 技术路线重塑供应链、DeepSeek 模型升级到V3.1…
Xin Lang Cai Jing· 2025-08-20 05:17
Group 1: Photovoltaic Industry - The Ministry of Industry and Information Technology (MIIT) is actively coordinating between power generation companies and local industries to enhance price transmission from manufacturing to power stations, emphasizing a market-oriented and legal approach to eliminate outdated production capacity [1] - The average bidding price for components from China Resources and China Huadian has increased by 5-8% month-on-month, while silicon material companies have proactively limited production, leading to a 10% decrease in silicon wafer inventory over two weeks [1] - The investment sequence indicates a tight supply of silicon materials in Q3, a premium for BC battery technology in Q4, and a simultaneous increase in both volume and price of auxiliary materials such as glass and adhesive films [1][2] Group 2: Solid-State Battery Technology - A breakthrough in solid-state battery technology has been achieved with the introduction of 5μm vapor-deposited lithium anodes, significantly reducing dendrite risk and achieving over 500 cycles with a capacity retention rate above 90% [3] - The cost of 5μm vapor-deposited lithium is projected to drop to 2 million yuan per GWh, compared to 4 million yuan for 20μm rolled lithium foil, indicating a substantial cost reduction in the industry [3] - The solid-state battery market could reach 50-100 billion yuan by 2030, driven by the demand for 100GWh of global solid-state battery production [3] Group 3: Robotics Industry - The T-Link Gen3 technology is reshaping the supply chain with a focus on lightweight materials, energy efficiency, and sensor integration, leading to a re-tendering of motors, reducers, and lead screws [4] - The use of PEEK materials has reduced costs by 30% compared to imports, and the new harmonic magnetic field motors have achieved a 50% reduction in size while doubling power density [5] - The 3D vision solution from Orbbec has a single machine value of 200 USD, and the company has passed factory audits [6] Group 4: Semiconductor and AI Models - The DeepSeek model has been upgraded to V3.1, expanding the context length from 64K to 128K, which is expected to increase demand for GPU memory and HBM [7] - The need for larger training clusters is anticipated to rise by 30%, benefiting semiconductor and storage manufacturers such as Cambricon, Haiguang, and Lanke [7] Group 5: Pharmaceutical Industry - Rongchang Biotech has licensed its ophthalmic drug RC28-E to Japan's Santen Pharmaceutical, marking a shift in domestic innovative drug licensing from popular fields like oncology to specialized areas with differentiated advantages [8] - This collaboration model provides a clear path for value realization in less popular biotech sectors through upfront payments, milestones, and sales sharing, enhancing cash flow and leveraging established commercialization channels [8] Group 6: High-Speed Rail Industry - The China National Railway Group has initiated its second batch of high-speed train tenders for the year, with 210 sets, marking a recent high and exceeding market expectations [9] - This move reinforces the trend of sustained railway investment recovery, with new construction and maintenance peaks positively impacting the performance certainty of core companies in the industry [9]
DeepSeek 开源新模型 V3.1:上下文长度拓展至 128K
Huan Qiu Wang Zi Xun· 2025-08-20 04:54
来源:环球网 【环球网科技综合报道】8月20日消息,DeepSeek日前在Hugging Face上开源了新模型 V3.1-Base。 此外,日前DeepSeek 还发布通知称,线上模型版本已升级至 V3.1,上下文长度拓展至 128k,可通过官 方网页、App、小程序测试,API 接口调用方式保持不变。 就在8月14日,DeepSeek App发布了1.3.0版本,此次更新在修复已知问题、优化文本操作体验的基础 上,首次引入"对话内容生成分享图"功能,为用户提供更便捷、个性化的内容传播方式。(思瀚) ...
DeepSeek V3.1 Base突袭上线,击败Claude 4编程爆表,全网在蹲R2和V4
3 6 Ke· 2025-08-20 03:52
Core Insights - The newly released DeepSeek V3.1 model features 685 billion parameters and supports various precision formats, from BF16 to FP8 [1] - The model demonstrates exceptional programming capabilities, achieving a score of 71.6% in the Aider programming benchmark, surpassing Claude Opus 4 [1][11] - V3.1 introduces native search token support, enhancing search functionalities [1] - The architecture has been innovated by removing the "R1" designation, indicating a potential shift towards a hybrid architecture in future models [1][10] - The cost for a complete programming task is only $1.01, significantly lower than proprietary systems, which are 60 times more expensive [1][13][16] Performance Metrics - DeepSeek V3.1 has 671 billion parameters activated with a context length of 128K tokens, ranking fourth on Hugging Face's trend list even before the model card was released [2] - The model's programming performance is 1% higher than Claude 4, with a cost reduction of 68 times [16] - In the SVGBench benchmark, V3.1 ranks just below GPT-4.1-mini, outperforming its predecessor, DeepSeek R1 [17] User Engagement - The DeepSeek community has grown to over 80,000 followers, indicating strong interest and anticipation for future releases [4] - Users have reported significant improvements in understanding and output speed, particularly in context length tests [21][25]
DeepSeek有点含蓄了,实测V3.1有进步,编程等个别场景硬刚GPT-5
3 6 Ke· 2025-08-20 03:03
Core Insights - The article discusses the advancements in the DeepSeek V3.1 model, highlighting improvements in code capabilities and front-end aesthetics, as well as the model's ability to handle complex tasks and logic reasoning [4][21][23] Group 1: Model Enhancements - The model size has reached 685 billion parameters, supporting various tensor types such as BF16, F8_E4M3, and F32, which balance computational precision and efficiency [4] - Significant improvements in code generation and front-end design aesthetics have been noted, with V3.1 performing well in code testing scenarios [4][6] - The model's context length has been expanded to 128K tokens, enhancing its processing capabilities compared to previous versions [2] Group 2: Product Design and Features - A proposed product design combines calendar and to-do list functionalities, allowing users to categorize tasks with color coding, manage short-term tasks, and visualize long-term tasks effectively [5] - The design includes features for marking tasks as completed, handling overdue tasks with visual prompts, and displaying long-term tasks across multiple days [5] Group 3: Logic Reasoning Improvements - The V3.1 model shows progress in logical reasoning, as demonstrated by its performance on a specific prediction problem involving multiple individuals and their statements about selection outcomes [21] - Despite being a non-reasoning model, V3.1 has made strides in understanding and processing logical scenarios [21] Group 4: Future Expectations - The article mentions ongoing anticipation for the DeepSeek R2 update, despite delays, indicating that the company continues to make steady improvements with each release [23]
刚刚,DeepSeek新模型开源,五大能力变化明显,附一手体验
3 6 Ke· 2025-08-20 00:14
Core Insights - DeepSeek has upgraded its online model to DeepSeek V3.1, expanding the context window from 64k to 128k, available across web, app, and mini-program platforms [3][21] - The new model has shown improvements in various capabilities, including programming, understanding physical laws, creative writing, and mathematical problem-solving [4][19] Model Upgrade - The model is now open-sourced on Hugging Face, with only the Base version available for download, showing no significant changes in parameters or tensor types compared to DeepSeek-V3-0324 [2] - Initial experiences with DeepSeek V3.1 indicate enhanced performance in web development, with more complex and aesthetically pleasing outputs compared to the previous version [4][6] Performance Enhancements - In web development tasks, DeepSeek V3.1 produced longer code with improved completion and aesthetics, demonstrating better layout and content planning [4][6] - The model successfully recreated a simple game similar to the Chrome dinosaur game, although some aspects of the game were not accurately rendered [8] Question Answering and Interaction - DeepSeek V3.1 provided more detailed and factually accurate responses to niche historical questions, showing a reduction in "hallucination" compared to its predecessor [10][12] - The model's tone has shifted to be more conversational and nuanced, using conditional statements and emphasizing complexity in its answers [13] Creative Outputs - The model demonstrated its creative capabilities by generating poetry and engaging in playful comparisons between notable figures in AI, showcasing a balanced approach in its responses [17][16] Mathematical Abilities - DeepSeek V3.1 displayed a mixed performance in basic arithmetic, initially providing incorrect answers before correcting itself [18] User Engagement - Users have quickly adopted the new model, with feedback highlighting improvements in physical simulations and creative outputs [19][21]
AI进化速递 | DeepSeek线上模型版本升级
Di Yi Cai Jing· 2025-08-19 13:19
Group 1 - DeepSeek has upgraded its online model to version 3.1, expanding the context length to 128k [1] - Alibaba's Tongyi Qianwen has launched an image editing model called Qwen-Image-Edit [1] - Nvidia has released a small language model named Nemotron-Nano-9B-v2 [1] Group 2 - Xiaopeng Motors' chairman He Xiaopeng announced that humanoid robots and L4-supported vehicles are expected to be mass-produced by 2026 [3] - Nvidia is collaborating with Foxconn to develop a humanoid robot, which is expected to debut in November [1] - Figure's founder indicated that Helix is set to undergo a significant upgrade [1] - Arm has hired Amazon's AI chip head Rami Shino to support its self-developed chip initiatives [1]
DeepSeek线上模型版本升级
Di Yi Cai Jing· 2025-08-19 13:17
Core Viewpoint - The online model version of DeepSeek has been upgraded to V3.1, expanding the context length to 128K, while maintaining the same access methods through the official website, app, mini-program, and API calls [1] Group 1 - DeepSeek's online model has been upgraded to version 3.1 [1] - The context length has been expanded to 128K [1] - Access methods remain unchanged, including the official website, app, mini-program, and API [1]
DeepSeek新版本突袭上线,R2发布时间仍未明确
Feng Huang Wang· 2025-08-19 12:20
Group 1 - The core focus of the recent update is the expansion of context length, which has been increased to 128k, allowing for improved memory and processing capabilities [1][3] - Users have reported enhancements in front-end coding capabilities following the update [1][3] - There are rumors regarding the potential release of DeepSeek R2 in late August, but no official release date has been confirmed [3]
DeepSeek线上模型版本升级至V3.1
Mei Ri Jing Ji Xin Wen· 2025-08-19 11:43
Core Viewpoint - DeepSeek has upgraded its online model to version 3.1, expanding the context length to 128k [1] Company Summary - The upgrade to version 3.1 indicates a significant enhancement in the capabilities of DeepSeek's online model, allowing for a larger context length which can improve the model's performance and usability [1]
核心模型被曝蒸馏DeepSeek?前女友一纸控诉,曝出欧版OpenAI塌房真相
3 6 Ke· 2025-08-18 12:12
Core Viewpoint - Mistral AI, once hailed as "Europe's OpenAI," is embroiled in a scandal involving allegations of plagiarism, specifically that its core technology is derived from DeepSeek, misleadingly presented as an original RL achievement [1][3][21]. Group 1: Allegations and Scandal - A former female employee of Mistral revealed in a personal letter that the company distilled DeepSeek's technology and misrepresented it as their own, using OpenAI's data while distorting benchmark results [3][4][21]. - The scandal gained traction online, with notable figures in the AI community, such as DeepMind researcher Susan Zhang, publicly condemning Mistral's unethical practices [4][21]. - The former employee expressed her frustrations about being sidelined and ignored when she raised concerns about the company's practices, leading to her eventual dismissal [6][7]. Group 2: Technical Comparisons - An industry insider, Sam Paech, had previously noted similarities between Mistral's Small 3.2 model and DeepSeek, suggesting that Mistral's outputs closely mirrored those of DeepSeek [9][10]. - Further analysis revealed that Mistral-small-3.2 and DeepSeek-v3 exhibited strikingly similar characteristics, indicating a lack of originality in Mistral's model [12][21]. Group 3: Historical Context and Achievements - Mistral AI was once celebrated for its rapid rise, achieving a valuation of $6.2 billion within just over a year of its establishment, positioning itself as a significant player in the European AI landscape [24][34]. - The company had previously launched successful products, including the Le Chat application, which topped the charts in France, and was supported by French President Macron as a key player in the national AI strategy [26][28][34].