Workflow
量化技术
icon
Search documents
AI创业圈又冲出一个288亿独角兽......
Tai Mei Ti A P P· 2025-08-15 03:09
Core Insights - Fireworks AI has emerged as a unicorn with a valuation of $28.8 billion, backed by prominent investors including Nvidia and AMD, indicating strong confidence in its business model and technology [1][14][17] - The founder, Qiaolin, has a robust background in AI and technology, having previously led a large engineering team at Meta, which developed PyTorch into a leading tool for AI developers [2][12] - Fireworks AI aims to simplify AI deployment for startups by providing optimized access to powerful AI models through a pay-per-use API, addressing common pain points in the industry [5][12] Company Overview - Fireworks AI was founded in 2022 by Qiaolin and a team of experts from PyTorch and Google, focusing on AI infrastructure and optimization technologies [2][5] - The company operates as an "AI computing central kitchen," renting Nvidia servers and pre-installing popular open-source models for easy access by clients [5][12] Technology and Innovation - Fireworks AI's competitive edge lies in its proprietary optimization techniques that enhance the speed and cost-effectiveness of AI models, making it more than just a server rental service [6][10] - The company has successfully improved the performance of its client, Cursor, by implementing techniques such as quantization and speculative execution, resulting in a significant increase in processing speed [10][12] Market Position and Competition - Fireworks AI has attracted significant investment from top-tier venture capital firms and tech giants, establishing itself as a key player in the AI infrastructure market [13][14] - The relationship with Nvidia is complex, as Nvidia not only invests in Fireworks AI but also competes in the same space, raising concerns about potential conflicts of interest and market dynamics [15][17] - Qiaolin acknowledges the competitive landscape and the necessity for Fireworks AI to scale quickly to establish a strong market position before facing direct competition from Nvidia [16][17]
让强化学习快如闪电:FlashRL一条命令实现极速Rollout,已全部开源
机器之心· 2025-08-12 09:51
机器之心报道 编辑:冷猫 在今年三月份,清华 AIR 和字节联合 SIA Lab 发布了 DAPO,即 Decoupled Clip and Dynamic sAmpling Policy Optimization(解耦剪辑和动态采样策略优化)。这是 一个可实现大规模 LLM 强化学习的开源 SOTA 系统,使用该算法,该团队成功让 Qwen2.5-32B 模型在 AIME 2024 基准上获得了 50 分, 我们也做了相关报道 。 中国科学技术大学校友,伊利诺伊大学香槟分校博士,微软研究院的首席研究员 刘 力 源 、清华大学校友,加州大学圣地亚哥分校计算机科学与工程学院博士生 姚峰 团队在强化学习的研究中更进一步。 该团队发现,在 DAPO-32B 中, rollout 生成是强化学习训练的主要瓶颈 ,占据了约 70% 的总训练时间。因此,该团队从 rollout 阶段着手,将 8 bit 量化技术应用 于 rollout 生成,并通过 TIS 技术在保持下游性能的同时实现了显著加速。 众所周知,FP8 能让强化学习运行得更快,但往往以性能下降为代价。 刘力源、姚峰团队推出 FlashRL,是首个开源且可用 ...
独家网络研讨会:“美”涨船高之际,如何以量化技术把握美股机遇?
彭博Bloomberg· 2025-07-18 05:43
Core Viewpoint - The article discusses the recent strong performance of the US stock market, particularly the S&P 500 index, which has approached historical highs, and highlights the importance of understanding market dynamics and utilizing quantitative techniques for investment opportunities [1]. Group 1: Market Dynamics - The US stock market has shown a strong upward trend, with the S&P 500 index nearing historical highs as of early July [1]. - Goldman Sachs has raised its target for the index to 6900 points for the second time since May, indicating a positive outlook [1]. Group 2: Key Issues to Address - The article raises critical questions regarding the sources of market optimism and how it may evolve in the future [1]. - It emphasizes the need for systematic exploration and evaluation of investment opportunities, from macroeconomic outlooks to individual stock potentials [1]. Group 3: Investment Strategies - The discussion includes the role of options strategies in risk management and enhancing returns during portfolio adjustments [1]. - It highlights the importance of technical indicators in practical applications for investment analysis [1]. Group 4: Event Details - The article promotes a webinar featuring Bloomberg experts who will provide in-depth analysis of recent trends in the US stock and options markets, as well as insights on using the Bloomberg quantitative platform BQuant Desktop for various analyses [1][4].
ICML 2025 | 注意力机制中的极大值:破解大语言模型上下文理解的关键
机器之心· 2025-05-06 04:11
Core Insights - The article discusses a significant phenomenon in large language models (LLMs) related to the concentration of massive values in the self-attention mechanism, particularly in the query (Q) and key (K) representations, which is crucial for contextual knowledge understanding [1][3][4]. Research Highlights - The study reveals that massive values are highly concentrated in Q and K, which is contrary to the expectation of independent operations in each attention head. This consistency across multiple layers and heads is visually demonstrated [3][4]. - The phenomenon of massive values is specifically observed in models using Rotational Position Encoding (RoPE), such as LLaMA, Qwen, and Gemma, while models without RoPE, like GPT-2 and OPT, do not exhibit this pattern [4]. - The research establishes a direct link between the presence of massive values in Q and K and the ability to understand contextual knowledge [4]. Key Findings 1. **Concentration of Massive Values**: Massive values are found to be highly concentrated in specific regions of each attention head, indicating a surprising level of consistency [3][4]. 2. **Impact on Contextual Knowledge Understanding**: The study shows that the presence of massive values is critical for understanding contextual knowledge, as demonstrated through destructive experiments that reset these values to their average [5][6]. 3. **Quantization Techniques**: Specific quantization methods that address massive values, such as AWQ and SmoothQuant, are shown to better preserve contextual knowledge understanding compared to methods that do not focus on massive values [7]. 4. **Origin of Concentration Phenomenon**: The concentration of massive values is attributed to RoPE, which affects low-frequency regions in Q and K, leading to this phenomenon appearing from the early layers of the model [8]. Experimental Results - The experiments reveal a stark contrast in the impact of massive values on different knowledge tasks: - **Resilience in Parametric Knowledge Retrieval**: Tasks relying on parametric knowledge show a decline of only 15-20% in accuracy when massive values are disrupted, maintaining 76%-88% accuracy [10]. - **Catastrophic Decline in Contextual Knowledge Tasks**: Tasks requiring contextual understanding experience a drastic drop in performance, with accuracy in key retrieval tasks plummeting from 100% to near 0% when massive values are disrupted [11]. - **Control Experiments**: When only non-massive values are disrupted, task performance remains stable, confirming the unique importance of massive values in contextual understanding [12]. Future Directions - The research opens several avenues for further exploration, including enhancing or adjusting the distribution of massive values to improve contextual understanding, examining the universality of this phenomenon across different architectures, and designing targeted quantization methods to protect massive values related to contextual understanding [16].