Core Viewpoint - The article highlights that the concept of Scaling Law was proposed 32 years ago by Bell Labs, not by recent AI advancements, emphasizing the historical significance of this research in machine learning [1][6]. Group 1: Historical Context - The paper titled "Learning Curves: Asymptotic Values and Rate of Convergence" introduced a predictive method for training errors and testing errors converging to the same asymptotic error value as training size increases, following a power-law form [4][6]. - The authors of the 1993 paper included notable figures such as Vladimir Vapnik and Corinna Cortes, who contributed significantly to the field of machine learning [6][25]. Group 2: Methodology and Findings - The research aimed to save computational resources when training classifiers by predicting their performance on larger datasets based on smaller training sets [8][10]. - The study found that as the training set size increases, both training and testing errors converge to a common asymptotic value, denoted as 'a', which typically falls between 0.5 and 1 [10][16]. - The proposed method allows for the estimation of classifier performance on larger datasets without complete training, thus conserving computational resources [10][14]. Group 3: Implications and Applications - The findings indicated that the predictive model was highly accurate for linear classifiers, demonstrating its potential to optimize resource allocation in training models [15][24]. - The research also revealed that the more difficult the task, the higher the asymptotic error and the slower the convergence rate, indicating a relationship between task complexity and learning efficiency [22].
他们在1993年就提出了Scaling Law
量子位·2025-09-02 06:17