DeepSeek V3.1 Base突袭上线,击败Claude 4编程爆表,全网在蹲R2和V4
Seek .Seek .(US:SKLTY) 3 6 Ke·2025-08-20 03:52

Core Insights - The newly released DeepSeek V3.1 model features 685 billion parameters and supports various precision formats, from BF16 to FP8 [1] - The model demonstrates exceptional programming capabilities, achieving a score of 71.6% in the Aider programming benchmark, surpassing Claude Opus 4 [1][11] - V3.1 introduces native search token support, enhancing search functionalities [1] - The architecture has been innovated by removing the "R1" designation, indicating a potential shift towards a hybrid architecture in future models [1][10] - The cost for a complete programming task is only $1.01, significantly lower than proprietary systems, which are 60 times more expensive [1][13][16] Performance Metrics - DeepSeek V3.1 has 671 billion parameters activated with a context length of 128K tokens, ranking fourth on Hugging Face's trend list even before the model card was released [2] - The model's programming performance is 1% higher than Claude 4, with a cost reduction of 68 times [16] - In the SVGBench benchmark, V3.1 ranks just below GPT-4.1-mini, outperforming its predecessor, DeepSeek R1 [17] User Engagement - The DeepSeek community has grown to over 80,000 followers, indicating strong interest and anticipation for future releases [4] - Users have reported significant improvements in understanding and output speed, particularly in context length tests [21][25]