线性规划
Search documents
DeepSeek悄悄开源LPLB:用线性规划解决MoE负载不均
3 6 Ke· 2025-11-20 23:53
Core Insights - DeepSeek has launched a new code repository called LPLB on GitHub, which aims to address the bottlenecks of correctness and throughput in model training [1][4] - The project currently has limited visibility, with fewer than 200 stars on GitHub, indicating a need for more attention [1] Project Overview - LPLB stands for Linear-Programming-Based Load Balancer, designed to optimize load balancing in machine learning models [3] - The project is still in the early research phase, with performance improvements under evaluation [7] Mechanism of LPLB - LPLB implements dynamic load balancing through three main steps: dynamic reordering of experts, constructing replicas, and solving optimal token allocation for each batch [4] - The mechanism utilizes a built-in linear programming solver and NVIDIA's cuSolverDx and cuBLASDx libraries for efficient linear algebra operations [4][10] Comparison with EPLB - LPLB extends the capabilities of EPLB (Expert Parallel Load Balancer) by focusing on dynamic fluctuations in load, while EPLB primarily addresses static imbalances [8] Key Features - LPLB introduces redundant experts and edge capacity definitions to facilitate token redistribution and minimize load imbalance among experts [9] - The communication optimization leverages NVLINK and NVSHMEM to reduce overhead compared to traditional methods [10] Limitations - Current limitations include ignoring nonlinear computation costs and potential delays in solving optimization problems, particularly for small batch sizes [11][12] - In extreme load imbalance scenarios, LPLB may not perform as well as EPLB due to its allocation strategy [12] Typical Topologies - LPLB allows for various topological configurations, such as Cube, Hypercube, and Torus, to define the distribution of expert replicas [13] Conclusion - The LPLB library aims to solve the "bottleneck effect" in large model training, where the training speed is limited by the slowest GPU [14] - It innovatively employs linear programming for real-time optimal allocation and utilizes NVSHMEM technology to overcome communication bottlenecks, making it a valuable resource for developers working on MoE architecture training acceleration [14]
DeepSeek悄悄开源LPLB:用线性规划解决MoE负载不均
机器之心· 2025-11-20 15:13
Core Insights - DeepSeek has launched a new code repository called LPLB (Linear-Programming-Based Load Balancer) on GitHub, which aims to optimize the workload distribution in Mixture of Experts (MoE) models [2][5]. - The project is currently in the early research stage, and its performance improvements are still under evaluation [8][15]. Project Overview - LPLB is designed to address dynamic load imbalance issues during MoE training by utilizing linear programming algorithms [5][9]. - The load balancing process involves three main steps: dynamic reordering of experts based on workload statistics, constructing replicas of experts, and solving for optimal token distribution for each batch of data [5][6]. Technical Mechanism - The expert reordering process is assisted by EPLB (Expert Parallel Load Balancer), and real-time workload statistics can be collected from various sources [6][11]. - LPLB employs a lightweight solver that uses NVIDIA's cuSolverDx and cuBLASDx libraries for efficient linear algebra operations, ensuring minimal resource consumption during the optimization process [6][11]. Limitations - LPLB currently focuses on dynamic fluctuations in workload, while EPLB addresses static imbalances [11][12]. - The system has some limitations, including ignoring nonlinear computation costs and potential delays in solving optimization problems, which may affect performance under certain conditions [11][12]. Application and Value - The LPLB library aims to solve the "bottleneck effect" in large model training, where the training speed is often limited by the slowest GPU [15]. - It introduces linear programming as a mathematical tool for real-time optimal allocation and leverages NVSHMEM technology to overcome communication bottlenecks, making it a valuable reference for developers researching MoE architecture training acceleration [15].
一个运行了80年的算法,我们现在才真正理解它?
机器之心· 2025-10-19 03:48
Core Insights - The article discusses the significance of the simplex method, a mathematical optimization technique developed by George Dantzig in 1947, which has been widely used for resource allocation and logistics decisions for nearly 80 years [4][6][10]. Group 1: Historical Context - George Dantzig, a prominent mathematician, created the simplex method after solving two unsolved problems during his graduate studies, which later became foundational for his doctoral thesis [2][3]. - The U.S. military's interest in optimization problems post-World War II led to the development of the simplex method to efficiently allocate limited resources in complex scenarios [5][6]. Group 2: Theoretical Developments - Despite its practical efficiency, the simplex method faced theoretical challenges, particularly regarding its potential exponential time complexity with increasing constraints, as proven by mathematicians in 1972 [7][10]. - Recent research by Sophie Huiberts and Eleon Bach has addressed these theoretical concerns, demonstrating that the feared exponential running time does not manifest in practice [10][26]. Group 3: Methodological Insights - The simplex method operates geometrically by finding the shortest path through a multi-dimensional space defined by constraints, aiming to maximize profit or minimize costs [11][19][21]. - The introduction of randomness in the algorithm, as established by earlier researchers, has been shown to significantly improve its performance, ensuring polynomial time complexity rather than exponential [13][17][26]. Group 4: Future Directions - The latest findings suggest that while significant progress has been made in understanding the simplex method, the ultimate goal remains to develop a method that scales linearly with the number of constraints [28]. - Although the research has not yet led to direct practical applications, it provides stronger mathematical support for the reliability of current software based on the simplex method, alleviating concerns about potential exponential complexity [30].
展望未来:炼油与石化行业战略转型已成必选项
麦肯锡· 2025-08-26 10:06
Core Viewpoint - The refining and chemical industries are facing significant challenges due to slowing demand growth, the rise of electric vehicles, and ongoing capacity expansion, leading to a projected decline in refining margins by about 5% to 30% by 2030 [3][4]. Recent Trends and Market Outlook - The refining market is expected to see a notable decline in profit margins, primarily driven by demand slowdown and capacity expansion disrupting supply-demand balance [3]. - The chemical market is also under pressure, with rapid capacity expansion, especially in China, outpacing demand growth, leading to overcapacity and compressed profit margins [3]. Challenges for Asian Refining and Chemical Industries - The evolving market dynamics are reshaping the competitive landscape, necessitating adaptation from companies [4]. - Uncertainties in carbon neutrality policies complicate long-term planning for refining and chemical companies, potentially leading to the exit of outdated capacities and cancellation of planned projects [4]. - Geopolitical tensions and fluctuating trade policies are adding further challenges, with tariffs on key raw materials increasing production costs by approximately 7% [4]. Strategic Pathways for Survival - Companies are focusing on cost reduction, capacity optimization, and digital transformation to navigate the challenges in the refining and petrochemical sectors [5]. - Operational transformation is essential for survival, with companies leveraging various strategies to enhance operations and profitability [5][6]. Production and Optimization - Linear programming (LP) models can provide insights to capture high-value opportunities with minimal investment, potentially increasing capacity by up to 5% [7]. - Advanced process control (APC) is being deployed to optimize operations and improve product yields, with potential cost reductions of $0.3 per barrel [8]. Efficient Maintenance - Effective maintenance strategies can significantly reduce costs and downtime, with potential savings of 5-15% on turnaround costs [10]. - Predictive maintenance is being utilized to monitor equipment health and reduce unplanned downtime [10]. Capital Expenditure (CAPEX) Optimization - Optimizing CAPEX is crucial for addressing tightening capital constraints and ensuring maximum returns while minimizing costs and risks [11]. - Structured methodologies like risk threat prioritization (RTP) are being employed to ensure rigorous evaluation of capital projects, leading to CAPEX reductions of 10-20% [11][13]. Sales Optimization - Optimizing commercial performance is vital for maintaining profitability, with effective sales strategies leveraging data-driven analysis to accelerate revenue growth [14]. - Dynamic pricing models based on customer willingness to pay are being adopted to maximize revenue and profit [15]. Conclusion - The Asian refining and petrochemical industries are entering a period of structural upheaval, with traditional advantages becoming less reliable [16]. - Future winners will be those companies that can adapt quickly to market changes, deeply integrate digital technologies, and optimize costs and product portfolios [16].
杉数科技申请基于线性规划的热能交易市场调控方法专利,优化热能交易市场运作机制
Sou Hu Cai Jing· 2025-05-16 02:45
Core Insights - The article discusses the patent application by multiple subsidiaries of Shanshu Technology, focusing on a method for regulating the thermal energy trading market based on linear programming [1][2][3] Group 1: Patent Application - Shanshu Technology has applied for a patent titled "Method, Device, and Electronic Equipment for Regulating Thermal Energy Trading Market Based on Linear Programming," with publication number CN119990663A, filed on February 2025 [1] - The patent aims to optimize the operation mechanism of the thermal energy trading market, enhancing overall market efficiency and stability [1] Group 2: Company Overview - Shanshu Technology (Beijing) Co., Ltd. was established in 2016, with a registered capital of 13.003 million RMB, and has invested in 10 companies and participated in 70 bidding projects [2] - Shanghai Shanshu Network Technology Co., Ltd., also founded in 2016, has a registered capital of 10 million RMB and has participated in 8 bidding projects [2] - Guangzhou Shanshu Technology Co., Ltd. was established in 2023 with a registered capital of 1 million RMB, focusing on software and IT services [2] - Shanshu Technology (Nanjing) Co., Ltd., founded in 2023, has a registered capital of 3 million RMB and specializes in professional technical services [2][3] - Fifth Paradigm (Shenzhen) Technology Co., Ltd., established in 2018, has a registered capital of 1 million RMB and focuses on software and IT services [3] - Shanshu Technology (Suzhou) Co., Ltd., founded in 2016, has a registered capital of 1 million RMB and provides professional technical services [3]