Workflow
transformer模型
icon
Search documents
Nature公开谷歌IMO金牌模型技术细节!核心团队仅10人,一年给AI编出8000万道数学题训练
创业邦· 2025-11-14 10:24
Core Insights - Google DeepMind has publicly released the complete technology and training methods behind its new model, AlphaProof, which is designed for mathematical proofs [2][4] - The model utilizes a 3 billion parameter encoder-decoder transformer architecture and incorporates a reinforcement learning environment based on the Lean theorem prover [8][7] Development Process - The AlphaProof team was relatively small, with around 10 core members, and was led by IMO gold medalist Miklós Horváth, who developed a method for creating various problem variants for training [4][5] - Over the course of a year, the team explored various research ideas, integrating successful approaches into the AlphaProof system [5] Training Methodology - AlphaProof transforms the mathematical proof process into a game-like environment, where each mathematical proposition serves as a new game level [7] - The model was pre-trained on approximately 300 billion tokens of code and mathematical text, followed by fine-tuning with around 300,000 manually crafted proofs from the Mathlib library [9][10] - A significant breakthrough was achieved through an automated formalization process that generated about 80 million formalized problems from 1 million natural language math questions [10] Performance at IMO 2024 - AlphaProof demonstrated impressive performance at the 2024 IMO, solving three problems, including the most difficult one, P6, which only 5 out of 609 participants solved completely [15][16] - The model utilized a testing-time reinforcement learning mechanism to generate around 400,000 related problem variants for particularly challenging questions [13][15] Future Directions - Following its success, DeepMind has opened access to AlphaProof for researchers, allowing them to explore its capabilities [19] - While AlphaProof excels in identifying counterexamples and formalizing statements, it faces challenges with custom definitions and relies heavily on the Lean theorem prover [20] - The model's dependency on Lean's evolving environment and the limited availability of unique mathematical problems present ongoing challenges for its broader applicability [20]
泰信基金张海涛:量化策略长期业绩得益于丰富的数据源、因子库以及模型持续迭代
Zhong Zheng Wang· 2025-08-07 14:28
Group 1 - The core viewpoint is that quantitative strategies in investment rely on diverse data sources, including traditional financial reports and non-traditional data such as social media sentiment and supply chain information, to generate forward-looking investment signals [1][2] - The performance of growth factors has been relatively strong in the current year, indicating a favorable market environment for growth-oriented investments [1] - A rich factor library is essential for diversifying sources of returns and enhancing cyclical resilience, necessitating regular updates to the factor pool to include both economically supported and algorithmically derived factors [1] Group 2 - Continuous model iteration and an open attitude towards new technologies, particularly AI, are crucial for improving the efficiency of factor development and constructing stronger predictive signals [2] - The application of AI in quantitative investment processes has become increasingly prevalent, including the use of large models for text data analysis and advanced models like transformers for end-to-end factor mining [2]