规模扩展
Search documents
未知机构:FundaAI评论HockTan对CPO的看法铜缆还没死但光学-20260309
未知机构· 2026-03-09 02:15
Summary of Conference Call Notes Industry and Company Involved - The discussion revolves around the CPO (Copper-Powered Optical) industry and the perspectives of Hock Tan, CEO of Broadcom, regarding the future of optical interconnect technology. Core Points and Arguments - Hock Tan's comments do not fundamentally alter the direction of the CPO industry, with the GTC conference being a pivotal event for the sector [1] - Tan believes that optical interconnects are unlikely to become mainstream for intra-rack scaling before 2027, emphasizing that copper cables (DAC) currently hold advantages in latency, power consumption, and cost for short distances [1][2] - Broadcom's SerDes technology is expected to extend the transmission distance of 200G/400G copper cables, reducing the immediate incentive for customers to switch to optical solutions [1] - Tan indicates that significant adoption of optical technology will not occur until beyond 2028, as customers consistently prefer copper cables for current applications [2] - The current architecture, such as the Vera Rubin NVL72, continues to utilize copper for chip-to-chip or XPU-to-XPU connections within racks, with optical technology only being used for larger cluster boundaries [2] - The industry has historically focused on optical scaling discussions around systems like NVL576 and beyond, maintaining a pattern of using copper within racks and optical between racks [2] - A fully optical intra-rack connection may not be realized until future generations, potentially waiting for the Feynman era [2] - The short-term outlook indicates that copper cables are still very much in use and not obsolete [4] Additional Important Insights - The medium-term deployment of scale-out CPO is expected to face challenges in achieving large-scale implementation [5] - Long-term industry direction will hinge on the growth of scale-up domains reaching a critical point, where optical solutions will become necessary in certain applications [5] - Once this tipping point is reached, the value pools that Broadcom is currently trying to protect may be among the first to undergo restructuring [5] - There is a divergence in views between FundaAI and Broadcom, with FundaAI suggesting that Broadcom's recent emphasis on copper reflects its own interests while downplaying the long-term strategy for optical scaling [3]
后训练时代如何延续Scaling Law?这是你该读的LLM后训练综述
机器之心· 2025-05-01 02:11
Core Insights - The article discusses the significance of post-training techniques such as fine-tuning and reinforcement learning (RL) in enhancing the capabilities of large language models (LLMs) [1][2][5]. Summary by Sections Overview of LLM Post-Training - A recent review report on LLM post-training has gained positive feedback, compiling a resource library of related papers and tools that has received over 700 stars [2]. - The review includes contributions from institutions like UAE University of Artificial Intelligence, University of Central Florida, Google DeepMind, and the University of Oxford, covering techniques to enhance LLMs through RL, supervised fine-tuning, and evaluation benchmarks [2]. Challenges in LLMs - Despite advancements, LLMs face issues such as generating misleading content (referred to as "hallucinations") and maintaining logical consistency in longer conversations [5]. - The reasoning capabilities of LLMs are debated, as they operate on implicit statistical patterns rather than explicit logical reasoning, which can lead to difficulties in simple logical tasks [5]. Training Phases of LLMs - The training process of LLMs is divided into two main phases: pre-training and post-training. Pre-training focuses on next-token prediction using large datasets, while post-training involves multiple rounds of fine-tuning and alignment to improve model behavior and reduce biases [6]. Fine-Tuning Techniques - Fine-tuning is essential for adapting pre-trained LLMs to specific tasks, enhancing their performance in areas like sentiment analysis and medical diagnosis. However, it carries risks of overfitting and high computational costs [7][10]. - Efficient techniques like Low-Rank Adaptation (LoRA) and adapters can reduce computational overhead while allowing models to specialize in specific tasks [10]. Reinforcement Learning in LLMs - RL is introduced to improve LLM adaptability through dynamic feedback and optimization of sequential decisions. This differs from traditional RL settings, as LLMs select tokens from a vast vocabulary rather than a limited action space [9][11]. - The feedback in language-based RL is often sparse and subjective, relying on heuristic evaluations rather than clear performance metrics [13]. Scaling Techniques - Scaling is crucial for enhancing LLM performance and efficiency, though it presents significant computational challenges. Techniques like Chain-of-Thought (CoT) reasoning and search-based methods help improve multi-step reasoning and factual accuracy [14][15]. - Despite advancements, challenges such as diminishing returns and increased inference time remain, necessitating targeted strategies for efficient deployment [15]. Evaluation Benchmarks - Various benchmarks have been proposed to assess the performance of LLM post-training, ensuring a comprehensive understanding of their strengths and limitations across different tasks [46]. - These benchmarks play a vital role in improving response accuracy, robustness, and ethical compliance during the post-processing phase [46]. Future Directions - The article highlights the growing interest in RL for optimizing LLMs since 2020, emphasizing the need for interactive methods and robust reward modeling to address challenges like reward hacking [52]. - Key areas for future research include personalized and adaptive LLMs, process versus outcome reward optimization, and the integration of dynamic reasoning frameworks to enhance model performance in complex queries [53].