Seek .(SKLTY)
Search documents
PriceSeek提醒:铝锭现货价格普遍下跌
Xin Lang Cai Jing· 2025-12-08 12:25
Group 1 - The core viewpoint of the article indicates that the spot prices of aluminum ingots (AL99.70) from China Aluminum Corporation have decreased across various regions on December 8, 2025, with specific price drops noted [1][5] - In the East China market, the price is quoted at 21,920 yuan/ton, in South China at 21,810 yuan/ton, in Southwest China at 21,840 yuan/ton, and in Central China at 21,770 yuan/ton, reflecting declines of 170 yuan/ton, 160 yuan/ton, 160 yuan/ton, and 170 yuan/ton respectively [1][5] - The overall decline in prices represents a decrease of approximately 0.7-0.8%, suggesting that market supply is ample or demand is weak, which generally exerts a bearish influence on spot prices [2][6] Group 2 - The pricing model used by the company is based on big data and a pricing model to generate transaction guidance prices, known as the Business Society Price [3][7] - This pricing model can be utilized to determine two types of transaction settlement prices: one for a specified date and another for an average settlement price over a specified period [3][7] - The settlement price formula is defined as: Settlement Price = Business Society Benchmark Price × K + C, where K is an adjustment coefficient and C includes various cost factors [3][7][8]
DeepSeek双模型发布:一位是“话少助手” 一位是“偏科天才”
Ke Ji Ri Bao· 2025-12-08 10:03
Core Insights - DeepSeek has released two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which have garnered attention for their performance in comparison to leading models like OpenAI's GPT-5 and Google's Gemini3 Pro [1][2] Model Features - DeepSeek-V3.2 is designed as a high-efficiency assistant with strong reasoning and agent capabilities, aimed at automating complex tasks such as report generation and coding [2] - DeepSeek-V3.2-Speciale focuses on solving high-difficulty mathematical problems and academic research, pushing the limits of open-source model reasoning [2] Technological Innovations - The new models incorporate two significant breakthroughs: Domain-Specific Architecture (DSA) and Thinking Tool Invocation technology [2] - DSA enhances efficiency by allowing the model to retrieve only the most relevant information, reducing resource consumption [2] - Thinking Tool Invocation enables multi-round reasoning and tool usage, allowing the model to think, execute, and iterate on tasks autonomously [2] Market Positioning - The release of these models aims to bridge the performance gap between open-source and closed-source models, providing a competitive edge for open-source development [3][4] - DeepSeek's focus on practicality and generalization is intended to create pressure on closed-source vendors, transforming aspirations into competitive realities [4] Community Engagement - DeepSeek has updated its official web platform, app, and API to the new version, while the Speciale version is currently available only as a temporary API for community evaluation [4]
外媒关注中国发布“全球首款AI手机”:会是第二个“DeepSeek时刻”吗?
Huan Qiu Shi Bao· 2025-12-07 22:51
该产品在市场上引起热烈反响。据报道,这款原型机在中国一经发布便迅速售罄。虽然厂商并未透露总 销量,但其转售价格已在市场飙升约43%。美国科技媒体Wccftech报道称,该产品让人联想到2025年初 DeepSeek引发的轰动,当时全球集体震惊于中国以极低的计算成本提供的顶级推理模型,如今中国科 技公司再次推出全球首款真正具备智能代理功能的AI手机。 《印度快报》报道称,目前全球尚没有其他手机能够达到豆包手机如此高的自主性,虽然商业化进程还 有待观察,但是已清晰地展示了智能手机未来将如何改变我们的生活。同时,这款手机的问世也表明, 首款真正意义上的智能体手机或许并非来自硅谷,而是来自中国融合人工智能和移动技术的生态系统。 尽管这款产品目前只是豆包方面发布的"技术预览版",不过,将语言大模型植入到操作系统层面,也引 发业界关于数据授权、隐私、系统安全等问题的激烈争议。中关村信息消费联盟理事长项立刚告诉《环 球时报》记者,"将大模型与操作系统进行深入融合确实存在很大的争议,其商业推广也阻力重重。但 是如果要让AI Agent更加强大,必须深入到手机硬件和操作系统的底层,才能充分释放AI的能力。"项 立刚认为,"这肯 ...
开源和闭源模型的差距在拉大:这是DeepSeek论文揭示的残酷真相
3 6 Ke· 2025-12-06 00:03
Core Insights - DeepSeek's V3.2 technical report indicates that the performance gap between open-source models and closed-source models is not narrowing but rather widening, based on extensive empirical data [1][2]. Performance Comparison - In benchmark tests, DeepSeek V3.2 scored 85.0 in MMLU-Pro, while GPT-5 scored 87.5 and Gemini 3.0 Pro achieved 90.1. In the GPQA Diamond test, the scores were 82.4 for DeepSeek, 85.7 for GPT-5, and 91.9 for Gemini 3.0 Pro [2][3]. - The most significant gap was observed in the HLE test, where DeepSeek V3.2 scored 25.1, compared to GPT-5's 26.3 and Gemini 3.0 Pro's 37.7, indicating a substantial performance disparity [3][4]. Structural Issues Identified - The report identifies three structural issues limiting the capabilities of open-source models in complex tasks: 1. **Architectural Limitations**: Open-source models rely on traditional vanilla attention mechanisms, which are inefficient for long sequences, hindering scalability and effective post-training [6]. 2. **Resource Investment Gap**: The post-training budget for DeepSeek V3.2 exceeds 10% of its pre-training costs, while most open-source models allocate less than 1%, leading to significant performance differences [7]. 3. **AI Agent Capability Lag**: Open-source models show inferior generalization and instruction-following abilities in real-world applications, as evidenced by lower scores in key agent evaluation benchmarks [8]. DeepSeek's Strategic Innovations - DeepSeek has implemented fundamental technical innovations across three core dimensions: 1. **Architectural Changes**: Introduction of the DSA (DeepSeek Sparse Attention) mechanism, which reduces computational complexity from O(L²) to O(L×k), significantly lowering inference costs while maintaining performance [10]. 2. **Increased Resource Allocation**: DeepSeek has made an unprecedented decision to allocate substantial resources for post-training, training expert models in six key areas with a total of 943.7 billion tokens during the pre-training phase [12]. 3. **Enhanced Agent Capabilities**: Development of a systematic task synthesis process, creating over 1,800 diverse environments and 85,000 complex prompts, which has improved performance in agent-related tests [13]. Conclusion - DeepSeek V3.2 demonstrates a viable path for open-source AI to compete with closed-source models through innovative architecture and strategic resource allocation, suggesting that technological innovation may be the key to survival in the competitive AI landscape [14].
DeepSeek-V3.2巨「吃」Token,竟然是被GRPO背刺了
3 6 Ke· 2025-12-04 10:38
Core Insights - The release of DeepSeek-V3.2 has generated significant attention in the industry, highlighting both its capabilities and areas needing improvement, particularly in token efficiency and output verbosity [1][2][5]. Token Efficiency - DeepSeek-V3.2 Speciale exhibits poor token consumption efficiency, requiring 77,000 tokens for complex tasks compared to Gemini's 20,000 tokens, indicating over three times the token usage for similar quality outputs [1][5]. - Users have noted that if the token generation speed of DeepSeek-V3.2 Speciale could be improved from approximately 30 tokens per second to around 100 tokens per second, the overall usability and experience would significantly enhance [5]. Output Quality - The Speciale version has been criticized for producing lengthy and verbose outputs, often resulting in incorrect answers, which is attributed to inherent flaws in the GRPO algorithm [2][14]. - The technical report from DeepSeek acknowledges the increased token consumption during inference, with the Speciale version consuming 86 million tokens in benchmark tests, up from 62 million in the previous version [7][14]. Algorithmic Issues - The GRPO algorithm, which has been a standard in reinforcement learning, is identified as a source of bias leading to longer and incorrect responses. This includes length bias, where shorter correct responses receive greater updates, and longer incorrect responses face weaker penalties [18][21]. - While the difficulty bias has been optimized in DeepSeek-V3.2, the length bias remains, potentially contributing to the excessive token consumption observed in the Speciale version [18][21].
谷歌掀“美国版DeepSeek冲击”,投资人拆解算力赛道前景|华尔街观察
Di Yi Cai Jing Zi Xun· 2025-12-04 10:09
Core Insights - Concerns over Google's advancements in AI have led to a significant market value loss for Nvidia, exceeding $100 billion [1] - Morgan Stanley's report predicts Google's TPU production will reach approximately 5 million and 7 million units by 2027 and 2028, respectively, resulting in an estimated revenue increase of $13 billion and an EPS boost of $0.40 [1] - Google's stock has surged nearly 70% year-to-date, with its market capitalization approaching $4 trillion and a PE ratio nearly doubling from 14 to 28 [1] Group 1: Google's AI Developments - Google is considered the closest company to achieving AGI, with advantages in computational power and extensive data resources [4] - The launch of Gemini 3, trained on Google's TPUs, highlights the cost and efficiency benefits over Nvidia's GPUs [1][4] - Major investment firms have begun to position Google as a core holding, with Berkshire Hathaway disclosing a $4.3 billion stake in Alphabet [4] Group 2: Competitive Landscape - The competition is shifting from who has the smartest chatbot to who has the most integrated ecosystem, with Google holding advantages in both areas [5] - OpenAI may struggle in a costly multi-modal competition against Google, which is integrating Gemini into its extensive user base [5] - Concerns about AI investment efficiency are rising, but historical technological revolutions suggest long-term profitability [6] Group 3: Nvidia's Position - Despite Google's valuation reassessment, Nvidia remains a key player in AI, particularly in addressing emotional intelligence through its GPUs [7] - Nvidia's GPUs are essential for achieving breakthroughs in human-like thinking, while TPUs have advantages in specific scenarios [7] - The AI sector is still in its early stages, with Nvidia's valuation remaining reasonable despite market fluctuations [8][9] Group 4: Future Investment Opportunities - Investors are increasingly focusing on AI applications, which are seen as the true beneficiaries of AI advancements [10] - Vertical applications in sectors like education, healthcare, and creative industries are expected to yield significant opportunities [11] - Chinese companies, such as Bilibili, are gaining attention for their user experience and growth potential, supported by a large user base [11]
DeepSeek-V3.2被找出bug了:疯狂消耗token,答案还可能出错,研究人员:GRPO老问题没解决
3 6 Ke· 2025-12-04 02:21
Core Insights - DeepSeek-V3.2 has gained significant attention but still exhibits bugs, particularly in token efficiency, which has been a longstanding issue [1][4]. Group 1: Performance Issues - The Speciale version of DeepSeek-V3.2 consumes a higher number of tokens for complex tasks, requiring 77,000 tokens compared to Gemini's 20,000 tokens for the same problem [4]. - The model has a "length bias," where longer incorrect answers are penalized less, leading to the generation of verbose but incorrect responses [8][11]. Group 2: Algorithmic Biases - The GRPO algorithm has two hidden biases: length bias and difficulty bias. The length bias results in longer incorrect answers being favored, while the difficulty bias causes the model to focus excessively on overly simple or overly difficult questions, neglecting those of medium difficulty which are crucial for skill improvement [8][9]. - The core author of the research, Zichen Liu, noted that while the new advantage value calculation has corrected the difficulty bias, the length bias remains unaddressed [10][11]. Group 3: Token Efficiency and Cost - DeepSeek's official report acknowledges that token efficiency is still a challenge for V3.2, as the new models require generating longer trajectories to match the output quality of Gemini-3.0-Pro [14]. - Despite the high token consumption, DeepSeek-V3.2 is priced at only 1/24th of GPT-5, making it relatively acceptable in terms of cost [14].
AI三国杀:OpenAI狂卷,DeepSeek封神,却被Mistral偷了家?
3 6 Ke· 2025-12-03 11:55
Core Insights - Mistral has launched two significant products: the Mistral Large 3 model and the Ministral 3 series, both of which are open-source, multimodal, and designed for practical applications [1][3]. Mistral Large 3 - Mistral Large 3 features a MoE architecture with 41 billion active parameters and 675 billion total parameters, showcasing advanced image understanding and multilingual capabilities, ranking 6th among open-source models [3][6]. - It has achieved a high ELO score, placing it in the top tier of open-source models, comparable to Kimi K2 and slightly behind DeepSeek v3.2 [6][10]. - The model performs on par with larger models like DeepSeek 37B and Kimi K2 127B across various foundational tasks, indicating its competitive strength [8][10]. - Mistral has partnered with NVIDIA to enhance the model's stability and performance by optimizing the underlying inference pathways, making it faster and more cost-effective [10][16]. Ministral 3 Series - The Ministral 3 series includes models of 3B, 8B, and 14B sizes, all capable of running on various devices, including laptops and drones, and optimized for performance [11][18]. - The instruct versions of the Ministral 3 models show significant improvements in performance, with scores of 31 (14B), 28 (8B), and 22 (3B), surpassing the previous generation [11][29]. - The 14B version of Ministral has demonstrated superior performance in reasoning tasks, outperforming competitors like Qwen 14B in multiple benchmarks [25][28]. Strategic Positioning - Mistral aims to address enterprise needs by providing customizable AI solutions that are cost-effective and reliable, contrasting with the high costs associated with proprietary models from competitors like OpenAI and Google [29][33]. - The company is evolving into a platform that not only offers models but also integrates various functionalities such as code execution and structured reasoning through its Mistral Agents API [33][37]. - Mistral's approach reflects a shift towards a more decentralized AI model, emphasizing accessibility and usability across different devices and environments, which could reshape the global AI landscape [37][39].
朱啸虎:DeepSeek对人类历史的改变被低估了 |未竟之约
Xin Lang Cai Jing· 2025-12-03 10:40
新浪声明:所有会议实录均为现场速记整理,未经演讲者审阅,新浪网登载此文出于传递更多信息之目 的,并不意味着赞同其观点或证实其描述。 责任编辑:梁斌 SF055 由新浪财经 、微博着力打造,微博财经 × 语言即世界工作室联合出品的泛财经人文对话栏目《未竟之 约》首期深度访谈即将上线。主持人张小珺对话金沙江创投主管合伙人朱啸虎,直面AI浪潮下的激流 与暗礁。 朱啸虎:DeepSeek对人类历史的改变被低估了。 由新浪财经 、微博着力打造,微博财经 × 语言即世界工作室联合出品的泛财经人文对话栏目《未竟之 约》首期深度访谈即将上线。主持人张小珺对话金沙江创投主管合伙人朱啸虎,直面AI浪潮下的激流 与暗礁。 朱啸虎:DeepSeek对人类历史的改变被低估了。 新浪声明:所有会议实录均为现场速记整理,未经演讲者审阅,新浪网登载此文出于传递更多信息之目 的,并不意味着赞同其观点或证实其描述。 责任编辑:梁斌 SF055 ...
老外傻眼,明用英文提问,DeepSeek依然坚持中文思考
3 6 Ke· 2025-12-03 09:14
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which show significant improvements in reasoning capabilities, with DeepSeek-V3.2 competing directly with GPT-5 and Speciale performing comparably to Gemini-3.0-Pro [1] - There is a notable phenomenon where even when queries are made in English, the model sometimes reverts to using Chinese during its reasoning process, leading to confusion among overseas users [3][5] - The prevalent belief is that Chinese characters have a higher information density, allowing for more efficient expression of the same textual meaning compared to English [5][9] Model Performance and Efficiency - Research indicates that using non-English languages for reasoning can lead to a 20-40% reduction in token consumption without sacrificing accuracy, with DeepSeek R1 showing token reductions ranging from 14.1% (Russian) to 29.9% (Spanish) [9] - A study titled "EfficientXLang" supports the idea that reasoning in non-English languages can enhance token efficiency, which translates to lower reasoning costs and reduced computational resource requirements [6][9] - Another study, "One ruler to measure them all," reveals that English is not the best-performing language for long-context tasks, ranking sixth among 26 languages, with Polish taking the top spot [10][15] Language and Training Data - The observation that Chinese is frequently used in reasoning by models trained on substantial Chinese datasets is considered normal, as seen in the case of the AI programming tool Cursor's new version [17] - The phenomenon of models like OpenAI's o1-pro occasionally using Chinese during reasoning is attributed to the higher proportion of English data in their training, which raises questions about the language selection process in large models [20] - The increasing richness of Chinese training data suggests that models may eventually exhibit more characteristics associated with Chinese language processing [25]