Workflow
大模型算法
icon
Search documents
道通科技:持续对包括ChatGPT、DeepSeek等国内外前沿基础大模型算法进行合作应用
Zheng Quan Ri Bao· 2025-09-10 14:07
Core Viewpoint - The company is actively collaborating with leading foundational large model algorithms such as ChatGPT, DeepSeek, Llama, and Qwen to enhance its model capabilities and develop targeted domain-specific models based on business needs [2]. Group 1 - The company is engaged in algorithm innovation and application development, focusing on deepening its capabilities in the field of artificial intelligence [2]. - The collaboration with both domestic and international advanced foundational large model algorithms indicates a strategic approach to leverage cutting-edge technology [2].
Z Event|大厂的同学下班一起聊AI?8.28北京和新加坡线下开饭
Z Potentials· 2025-08-26 04:16
Group 1 - The events are designed for professionals from large companies and startups in product/technology sectors, focusing on topics like large model algorithms and AI agents [1][3] - The gatherings are limited in size, with 8-10 participants in Beijing and 6-8 in Singapore, promoting an intimate networking environment [1][3] - Registration for the events closes the night before at 8 PM, emphasizing the limited availability and first-come, first-served basis [3] Group 2 - The company is actively recruiting a new cohort of interns, targeting creative individuals from the post-2000 generation [6][7] - The initiative is part of a broader effort to find innovative young entrepreneurs, likened to a Chinese version of Y Combinator [8]
Z Event|大厂的同学下班一起聊AI?8.28北京和新加坡线下开饭
Z Potentials· 2025-08-24 11:51
Group 1 - The events are designed for professionals from large companies and startups in product/technology sectors, focusing on topics like large model algorithms and AI agents [1][3] - The gatherings are limited in size, with 8-10 participants in Beijing and 6-8 in Singapore, promoting an intimate networking environment [1][3] - Registration for the events is on a first-come, first-served basis, with a deadline set for 8 PM the night before each event [3] Group 2 - The company is actively recruiting a new cohort of interns, targeting creative individuals from the post-2000 generation [6][8] - The initiative is part of a broader effort to identify and nurture young entrepreneurial talent in the AI sector, likened to a Chinese version of Y Combinator [8]
中国石油长庆油田:井下作业“智变”带来“质变”
Core Viewpoint - The introduction of an intelligent downhole operation system at the Changqing Oilfield is set to revolutionize shale oil development by significantly enhancing operational efficiency and decision-making speed, addressing challenges in the industry [3][4]. Group 1: Technological Advancements - The intelligent system, referred to as the "shale oil brain," allows for historical data and fault diagnosis to be accessed in under 10 seconds, a drastic reduction from over 40 minutes previously [3]. - The system aims to achieve zero time in decision response, zero delays in collaboration, zero blind spots in fault handling, and zero deviations in evaluation systems [3]. - The accuracy of fault identification has reached 90%, showcasing the system's effectiveness in improving operational efficiency [3]. Group 2: Industry Context - Shale oil is a crucial component of unconventional oil and gas resources, with its efficient development being vital for national energy security [3]. - The Changqing Oilfield accounts for over 50% of China's total shale oil production, yet faces challenges such as low operational efficiency and insufficient precise decision-making due to various factors [3][4]. Group 3: Innovation and Collaboration - The company emphasizes the need for innovative thinking to optimize industrial structure and leverage digital technologies for business model reconstruction and management transformation [4]. - Collaboration with specialized digital companies has led to the development of a diagnostic expert knowledge base, enhancing the entire operational process from perception to evaluation [4]. Group 4: Operational Efficiency - The intelligent system integrates over 20 parameters, including downhole pressure, temperature, and flow, enabling real-time monitoring and preemptive alerts for operational issues [4][6]. - The transition from "post-repair analysis" to "pre-repair warning" has significantly reduced response times for downhole diagnostics from hours to seconds [4]. - The application of the intelligent model is expected to lower maintenance costs and improve decision-making efficiency, with average fault handling times previously exceeding three days [6].
DeepSeekGRM带来新的推理Scaling路径
HTSC· 2025-05-07 07:25
Investment Rating - The industry rating is "Overweight" indicating that the industry stock index is expected to outperform the benchmark [22] Core Insights - The introduction of the Self-Principle Critique Tuning (SPCT) method by the DeepSeek team enhances the efficiency and performance of generalist reward modeling during the inference phase, suggesting a new scaling method for inference [2][3] - The DeepSeek GRM model, with 27 billion parameters, achieves performance comparable to the existing R1 model with 671 billion parameters, indicating significant advancements in model efficiency [4] - The SPCT method improves model generation quality and scalability, outperforming existing models in benchmark tests, and demonstrates that inference phase scaling strategies are more advantageous than merely increasing model parameters during training [4][5] - The GRM model reduces hardware requirements significantly, with training costs being only one-sixth of the R1 model, and inference energy consumption at approximately 17% of the R1 model, making it favorable for edge deployment [5] - The upcoming release of the DeepSeek R2 model is anticipated within 1-2 months, with the GRM model serving as a precursor to further algorithmic innovations [6] Summary by Sections Inference Scaling - SPCT enhances the adaptability and scalability of models during inference, addressing challenges in obtaining accurate reward signals in general domains [3] - The new method provides insights for further iterations of large model algorithms [3] Model Performance - DeepSeek GRM-27B outperforms existing models, achieving results comparable to R1 and GPT-4o, while utilizing a dual-loop structure for real-time evaluation and correction [4] - The research indicates that new exploration in the inference phase can expand model boundaries despite a slowdown in scaling laws during pre-training [4] Hardware Efficiency - The GRM model's hardware requirements are significantly lower, allowing for potential deployment on consumer-grade GPUs, thus expanding the performance-cost boundary [5] Future Developments - The anticipated release of the DeepSeek R2 model is expected to bring further algorithmic innovations, with a focus on optimizing training and inference efficiency [6]