Nvidia(NVDA)
Search documents
AI 网络市场更新:基于 OFC 2026 展会的核心结论展望-AI networking market update_ Our read-through from OFC 2026
2026-03-26 13:20
Summary of Key Points from the Conference Call Industry Overview - **Industry**: AI Networking and Optical Communication - **Event**: OFC 2026 (Optical Fiber Conference) - **Market Dynamics**: The AI networking market is experiencing a long runway for growth driven by significant investments from major global AI players in large language models (LLMs) and increasing AI inference workloads. This trend is expected to continue through a multi-year technology upgrade cycle [1][2] Core Insights - **Market Growth**: - The optical communication market is at an inflection point with accelerated growth. Existing growth engines, such as pluggable transceivers, represent a market opportunity of USD 50 billion, while new growth engines like Optical Circuit Switch (OCS) and Co-packaged Optics (CPO) could add over USD 20 billion to the total addressable market (TAM) [1] - Lumentum's management projects its TAM will grow from USD 18 billion to USD 90 billion over the next five years, driven by advancements in scale-across, scale-up, and the OCS market [1] - **Technological Advancements**: - Next-generation 3.2T optical transceivers are nearing commercialization, expected to launch in late 2027 or early 2028. Major players are preparing their supply chains, showcasing advanced technologies such as 400G EML chips and optical engines [2] - TeraHop, a subsidiary of InnoLight, has introduced the industry's first 12.8T XPO transceiver module based on Silicon Photonics (SiPh) technology [2] - **Supply Chain Challenges**: - Supply bottlenecks persist in the high-end optical chip segment, with Lumentum's Indium Phosphide (InP) capacity expected to double by the end of 2026 and again by the end of 2027. Despite this, demand is projected to outpace supply, with an estimated demand-supply gap of 25%-30% [3] Emerging Technologies - **CPO and NPO**: - NVIDIA is advancing its CPO roadmap, with significant developments expected in both scale-out and scale-up networks. The scale-up CPO market could be 3 to 4 times larger than the initial scale-out market [6] - Cloud service providers (CSPs) are exploring NPO solutions, which offer more flexibility compared to CPO [5] - **Optical Fiber Innovations**: - High-end optical fibers, including multi-core and hollow-core fibers, are gaining attention due to increasing demands for higher bandwidth and lower latency [7] - Corning and Prysmian showcased innovative fiber solutions at OFC, highlighting the importance of optical fibers in AI networking [7] Market Projections - **OCS Market Growth**: - The global OCS market is projected to grow from approximately USD 400 million in 2025 to over USD 2.5 billion by 2029, reflecting a CAGR of 58% driven by AI demand [34] - Accelink and Eoptolink demonstrated their latest OCS products at OFC, indicating a push towards commercialization [34] Key Players and Products - **Zhongji InnoLight**: - Rated as a "Buy" with a target price of CNY 799.00, the company is recognized for its leadership in the global optical transceiver market and is expected to benefit from upcoming transceiver upgrades [49][50] - **Broadcom**: - Showcased a range of products aimed at AI clusters, including its first 400G/lane optical DSP and advanced optical solutions for both 1.6T and 3.2T transceivers [24][25] - **Other Notable Companies**: - Companies like Semtech, Macom, and Accelink are also making strides in high-speed copper and optical solutions, showcasing their latest technologies at the conference [8][43] Conclusion - The AI networking and optical communication sectors are poised for significant growth, driven by technological advancements and increasing demand for high-speed data transmission. Key players are actively preparing for the next generation of optical transceivers and solutions, while supply chain challenges remain a critical focus for the industry.
全球半导体与半导体设备:你相信埃隆(马斯克)吗?-Global Semiconductors and Semicap Do you believe in Elon
2026-03-26 13:20
Summary of Key Points from the Conference Call Industry Overview - The discussion centers around the **semiconductor industry**, particularly focusing on the ambitious **Terafab project** announced by **Elon Musk** aimed at scaling compute production to **1 terawatt (TW)** per year, which is approximately **50 times** the current global compute supply of **20 gigawatts (GW)** [2][4]. Core Insights and Arguments - **Terafab Project**: - The project will start with an advanced fabrication facility in **Austin**, designed to manufacture components necessary for advanced AI compute, including compute engines, logic, memory, packaging, and mask production [2]. - The focus will be on **edge inference chips** for applications like **Tesla cars** and **Optimus robots**, as well as space-optimized compute chips [2]. - **Manufacturing Requirements**: - To achieve **1 TW** of annual compute, it is estimated that between **7 to 18 million** **300mm wafer starts** per month would be required, primarily driven by **HBM memory** [3][4]. - This translates to needing **140-360 new 50K WSPM factories**, with a capital expenditure of approximately **$5-$13 trillion** at **$35 billion** per fab-equivalent [3][26]. - **Current Capacity Context**: - The required capacity for **1 TW** would exceed the entire current global installed semiconductor capacity, which is around **16 million 300mm equivalent WSPM** [4][28]. - The analysis excludes other semiconductor types outside of HBM, GPU, and CPU, indicating a significant gap in current manufacturing capabilities [4]. - **Market Implications**: - The immediate impact on the semiconductor industry may be limited to hype, but if Musk succeeds, it could lead to increased demand for semiconductor capital equipment (semicap) [4]. - The potential for Musk to produce his own chips could negatively affect current incumbents, but overall, the demand for compute is expected to benefit all players in the industry [4]. Additional Important Insights - **Investment Ratings**: - Various companies in the semiconductor space have been rated based on their performance and future prospects: - **ADI**: Market-Perform, target price **$375.00** [7]. - **AMD**: Market-Perform, target price **$235.00**, with potential growth from a new deal with OpenAI [8]. - **AVGO**: Outperform, target price **$525.00**, with strong AI growth expected [8]. - **NVDA**: Outperform, target price **$300.00**, with significant upside in the datacenter market [10]. - **QCOM**: Outperform, target price **$175.00**, despite memory headwinds [11]. - **AMAT**: Outperform, target price **$425.00**, driven by WFE growth [12]. - **Emerging Domestic Players**: - Companies like **NAURA**, **AMEC**, and **Piotech** are positioned to benefit from domestic WFE substitution in China, indicating a shift in market dynamics [14][15][16]. - **Global Semiconductor Landscape**: - The report highlights the competitive landscape, with established players like **Samsung**, **SK Hynix**, and **Micron** receiving favorable ratings, while others like **KIOXIA** are rated underperform [17][18]. This summary encapsulates the key points discussed in the conference call, providing insights into the semiconductor industry's current state and future potential driven by ambitious projects like Musk's Terafab.
科技未来:AI 数据中心网络入门指南-Future of Tech AI Datacenter Networking Primer
2026-03-26 13:20
Summary of AIDC Networking Conference Call Industry Overview - The focus is on **AI Datacenter (AIDC) networking**, which is becoming a critical component of AI infrastructure as AI workloads scale exponentially [1][10] - The total addressable market (TAM) for AIDC networking chips is projected to reach approximately **USD 100 billion by 2030**, with a compound annual growth rate (CAGR) of around **30%** [2][15] Key Insights - **Demand Surge**: The demand for AIDC networking chips is driven by the compound bandwidth effect, where adding accelerators increases not only point-to-point bandwidth but also multiplies traffic across higher tiers of the cluster [2][23] - **Networking Cost**: Networking components are becoming the second-largest cost in AI datacenters, indicating a faster growth rate for AIDC networking compared to xPUs [2][5] - **Connection Types**: AIDC networking can be categorized into three major connection types: - **DC-DC connections** for wide area bandwidth - **CPU-centric connections** for data flow management - **xPU-to-xPU connections** for high bandwidth and low latency pathways [3][36] Competitive Landscape - **Intense Competition**: The scale-up networking domain is highly competitive, with Nvidia's NVLink setting the performance benchmark, while alternatives like UALink and Ethernet-based architectures are emerging [4][66] - **Regional Variations**: China is developing its own protocols, such as Huawei's Unified Bus (UB), which reflects a strategic emphasis on larger cluster scales [4][52] Market Dynamics - **High Margins**: The sector offers strong industry beta and attractive margins due to high technological and capital barriers, limiting new entrants [5][66] - **Key Suppliers**: Major players include: - **Broadcom**: Dominates the merchant Ethernet switch silicon market and is well-positioned for next-generation AI fabrics [67][68] - **Nvidia**: Holds a leading position in AIDC networking through its vertically integrated AI platform [71][73] - **Marvell**: Focuses on high-performance networking and storage silicon, with a growing emphasis on AI DC networking [74][76] - **Huawei**: Innovates in AI DC networking in China with a proprietary architecture based on its UB protocol [82] Investment Implications - **Stock Ratings**: Companies like Hygon and Cambricon are rated as Outperform, with target prices set at **CNY 280** and **CNY 2,000**, respectively [7] - **Nvidia and Broadcom**: Both companies are expected to benefit significantly from the growing AIDC networking market, with target prices of **$300** and **$525**, respectively [8] Additional Insights - **Technological Evolution**: The architecture of AIDC networks is evolving, with a shift from maximizing individual accelerator performance to optimizing large-scale cluster efficiency [10][11] - **Forecasting Uncertainty**: While the market size is projected to grow, there remains a wide margin of uncertainty in forecasting due to the rapid evolution of AIDC technologies [11][12] - **Bandwidth Growth**: Total bandwidth in AIDC networks is expected to grow faster than accelerator compute capacity, driven by the compound bandwidth effect [23][32] This summary encapsulates the critical points discussed in the conference call regarding the AIDC networking industry, its competitive landscape, market dynamics, and investment implications.
GTC-OFC-年报一季报
2026-03-26 13:20
Summary of Key Points from Conference Call Records Industry Overview - **AI Computing Demand**: The demand for AI computing is expected to continue its explosive growth, with overseas computing power projected to increase by 150%-200% by 2026, while domestic computing power is expected to grow by 30%-50% [1][4]. - **Light Interconnection Sector**: The supply-demand situation in the optical interconnection sector is tight, with a significant increase in demand for 1.6T optical modules anticipated in 2026 [1]. - **Copper Interconnection and PCB Demand**: High growth in copper interconnection and PCB demand is expected, driven by upgrades in PCB layers and the value of optical connectors within cabinets reaching $30,000 to $50,000 [1]. Company-Specific Insights - **NVIDIA's LPU Architecture**: NVIDIA's LPU architecture has exceeded expectations, with SRAM capacity and bandwidth doubling, and the value of a single card's SRAM reaching $500. The V1 version is set for mass shipment in Q3 2026 [1][6]. - **Storage Sector Growth**: The storage sector is experiencing short-term high growth driven by price increases, with companies like 伟仕佳杰 benefiting from breakthroughs in B-end sales and margin improvements [1][4]. - **Consumer Electronics**: Companies like 裕同科技 are entering new markets through acquisitions, indicating a strong valuation margin [1]. Market Dynamics - **Geopolitical Impact on AI Industry**: Geopolitical conflicts may stimulate demand for data center reconstruction, further increasing the need for the AI industry chain. China's stable energy structure and social environment enhance its supply chain reliability [2]. - **Investment Opportunities in Internet and Gaming Sectors**: The valuation of Hong Kong's internet and gaming sectors is at historical lows (PE of 10-15), presenting opportunities for investment in quality assets [1][11]. Technological Developments - **Advancements in AI Hardware**: The GTC and OFC conferences highlighted the strengthening logic of the computing power industry, with advancements in CPU, optical modules, and copper cable interconnection technologies [3]. - **NVIDIA's Future Platforms**: The next-generation platform, Rubin, will integrate LPU but not all models will include it. The demand for GPUs and related hardware is expected to remain strong [8][9]. Financial Performance Expectations - **2026 Earnings Projections**: Overseas computing companies are expected to see significant earnings growth, with estimates of 150%-200% for the year. Domestic computing companies are projected to grow by 30%-50% [4][5]. - **Q1 Earnings Impact**: Q1 earnings for companies like 中际旭创 and 新易盛 may only account for 1/6 to 1/9 of annual profits, but are expected to show over 100% year-on-year growth [5]. Investment Strategies - **Focus on Value Stocks**: In the current market environment, stocks in the internet and gaming sectors, such as 腾讯 and 阿里巴巴, are seen as undervalued and present good investment opportunities [11][12]. - **Long-term View on AI Optical Interconnection**: The AI optical interconnection sector is expected to see significant growth, with a projected fivefold increase in market size from 2025 to 2030 [13][14]. Key Companies to Watch - **AI Optical Interconnection Companies**: Companies like 中际旭创 and 新易盛 are highlighted for their strong fundamentals and low valuations, making them attractive for investment [16][17]. - **Emerging Technologies**: The adoption of CPO and OCS technologies is expected to drive growth in the AI interconnection architecture, with significant market opportunities for companies involved [14][15]. Conclusion - The AI and optical interconnection sectors are poised for substantial growth, driven by technological advancements and increasing demand. Investment opportunities exist in undervalued companies within these sectors, particularly those with strong fundamentals and growth potential.
GTC-OFC总结-光互联-全液冷大时代
2026-03-26 13:20
Summary of Key Points from Conference Call Records Industry Overview - The industry is entering a new era characterized by optical interconnection and full liquid cooling, with the optical communication and liquid cooling sectors being the primary beneficiaries [1][2] Core Insights and Arguments - **Optical Module Demand**: Demand for traditional optical modules is stronger than expected, with Lumentum's 2027 capacity already secured by Google, indicating optimistic market expectations for 800G and 1.6T optical modules [2][1] - **XPO Module Introduction**: The introduction of the XPO module, featuring a rate of 12.8T and a single power consumption of up to 400W, necessitates liquid cooling for each optical module, reinforcing the trend towards full liquid cooling [2][1] - **New Technologies**: Technologies such as NPO, OCS, and CPO are actively advancing, with thin-film lithium phosphate materials gaining attention from optical module companies [2][1] - **Full Liquid Cooling Adoption**: The GTC conference confirmed that future products will adopt a 100% full liquid cooling solution, alleviating market concerns about some new products potentially not using liquid cooling [2][1] Key Developments in Chip and System Architecture - **Rubin System**: NVIDIA introduced the Rubin system, consisting of 7 chips and 5 architectures, set to be mass-produced in the second half of 2026, featuring HBM4 memory with a capacity of 288GB and a bandwidth 2.75 times that of HBM3e [3][4] - **Firman Architecture**: The next-generation GPU architecture "Firman" is designed for world models, utilizing TSMC's 1.6nm process, with a single GPU computing power of 50P and a 5-fold increase in inference performance compared to the previous generation [4][5] Market Expectations and Economic Concepts - **Token Factory Economics**: NVIDIA's "Token Factory Economics" concept emphasizes the importance of token throughput per watt as a core competitive metric, predicting AI chip demand to reach at least $1 trillion by 2027 [5][1] MSA Developments - **XPO, OpenCPX, and OCI**: These three MSAs aim to address core bottlenecks in optical interconnection for AI data centers, with XPO recognized for its innovative density and cooling capabilities, achieving 4 times the bandwidth density of mainstream OSFP optical modules [5][6] NPO and CPO Technologies - **NPO Technology**: Positioned as a mid-term solution for AI computing interconnection, NPO is expected to achieve scale before CPO, with significant reductions in power consumption and increased bandwidth density [7][1] - **CPO Technology**: CPO is gaining momentum, with NVIDIA planning to deploy it starting in 2026, and various companies showcasing CPO solutions at the OFC conference [8][9] OCS Technology - **OCS Commercialization**: OCS technology is moving towards large-scale commercialization, with Google and NVIDIA leading the way, promising significant reductions in latency and power consumption while enhancing bandwidth density [10][1] Hollow Fiber Technology - **Hollow Fiber Advancements**: Hollow fiber technology is transitioning to commercial use, with domestic manufacturers achieving global leadership in key metrics, offering significant bandwidth suitable for large-scale DCI interconnections [11][1]
GTC 2026 – 推理王国扩张 --- GTC 2026 – The Inference Kingdom Expands
2026-03-26 13:20
Summary of Nvidia's GTC 2026 Conference Call Company Overview - **Company**: Nvidia - **Event**: GTC 2026 Conference - **Date**: March 24, 2026 Key Announcements - Nvidia introduced three new systems: Groq LPX, Vera ETL256, and STX [5][6] - Updates were made to the Kyber rack architecture, including the introduction of the Rubin Ultra NVL576 and Feynman NVL1152 multi-rack systems [5][6] - The debut of CPO (Co-Packaged Optics) for scale-up networking was highlighted [5][6] - Jensen Huang's mention of InferenceX during the keynote was a significant highlight [5][6] Groq Acquisition - Nvidia "acquired" Groq for $20 billion to license their IP and hire most of their team, simplifying regulatory approval processes [10][11] - This transaction allows Nvidia immediate access to Groq's IP and personnel, facilitating rapid integration into Nvidia's systems [10][11] LPU Architecture - Groq's LPU architecture is designed to complement Nvidia's GPU, focusing on low latency and high bandwidth [12][13] - The LPU architecture includes various slices for different operations, such as VXM for vector operations and MEM for data loading [16][17] - The LPU's design emphasizes deterministic computation, allowing for aggressive instruction scheduling to hide latency [19] Performance and Market Position - The first generation LPU was built on a 14nm process, which was mature compared to competitors using more advanced nodes [20][21] - Groq's roadmap has stalled, with no LPU 2 shipped, widening the gap against competitors moving to 3nm processes [22][23] - The LPU 3 (LP30) is set to be productized by Nvidia, addressing previous design issues [30][31] Memory Hierarchy and Integration - The integration of SRAM in the memory hierarchy allows for low latency but at the cost of density and total throughput [27][28] - Nvidia aims to combine the strengths of LPU and GPU architectures to optimize performance in high-interactivity scenarios [45][46] Attention FFN Disaggregation (AFD) - AFD technique is introduced to improve decode phase latencies by leveraging the strengths of both GPUs and LPUs [45][46] - The decode phase in LLM inference is memory-bound, making LPU's high SRAM bandwidth advantageous [47][48] - Attention operations are stateful, while FFN operations are stateless, leading to their disaggregation for optimized performance [56][57] Future Developments - The next generation LP40 will be fabricated on TSMC N3P, incorporating more of Nvidia's IP and innovations like hybrid bonded DRAM [38][39] - Nvidia's roadmap includes significant advancements in memory capacity and bandwidth, with plans for future products to enhance performance [40] Conclusion - Nvidia's GTC 2026 showcased significant advancements in AI infrastructure, particularly through the integration of Groq's technology and the development of new systems aimed at enhancing performance in high-demand scenarios. The focus on low latency and high bandwidth solutions positions Nvidia favorably in the competitive landscape of AI hardware.
“词元”背后:新算力战争打响
财富FORTUNE· 2026-03-26 13:14
Core Insights - The article emphasizes that tokens are foundational to AI, marking a shift in productivity tools from mere software to entities capable of understanding and intervening in the physical world [1] - The rise of OpenClaw signifies a transformative moment in cloud computing, where new companies challenge traditional giants by focusing on efficiency and cost-effectiveness of tokens [4] Pricing Trends - Since March 2023, major cloud providers like Alibaba Cloud and Tencent Cloud have raised AI computing product prices by over 30%, with high-end GPU monthly rentals exceeding 50,000 yuan, indicating the end of the era of cheap computing [3] - Predictions suggest that global AI computing demand will grow by 58% year-on-year by 2026, with reasoning computing now accounting for over 70% of demand, and token consumption increasing by 2200% [3] OpenClaw and Token Dynamics - OpenClaw's rapid growth has positioned it as a potential new standard in AI tools, akin to Linux, and has catalyzed the emergence of token factories, shifting the focus from training to reasoning [6][7] - The introduction of OpenClaw has clarified token pricing, allowing it to be standardized and commercialized, moving away from the previous model where token value was highly variable [7] Cloud Computing Evolution - New cloud computing companies are emerging that focus solely on AI computing, optimizing for performance and cost efficiency, contrasting with traditional cloud giants that still carry the legacy of the internet era [4][12] - The transition to a "computing + skill" ecosystem is anticipated, with new cloud companies designed specifically for AI applications outperforming traditional ones in terms of efficiency [12][14] Competitive Landscape - Chinese new cloud companies are positioned to compete globally, leveraging a complete technology system, open-source contributions, and the ability to provide flexible computing solutions [18][19] - The competition between the US and China in the computing industry is expected to reach a dynamic balance, with both countries addressing their respective weaknesses [20]
NVIDIA Owns the Spotlight, But the Smart Money is Moving Downstream
247Wallst· 2026-03-26 12:50
Core Insights - Nvidia (NVDA) continues to lead the AI chip market under CEO Jensen Huang, but the stock has recently shown a lack of momentum despite strong quarterly results [2][4] - As 2026 progresses into a 'show me' phase for AI, major investors are reducing their positions in Nvidia and reallocating capital to adjacent opportunities [2][8] Company Performance - Nvidia remains at the forefront of the AI chip race, with significant investments in next-generation technologies, yet the stock has not gained traction following a strong quarter [4][6] - The recent trading activity indicates a mix of buying and selling among major investors, with some looking for less obvious AI winners downstream [5][14] Investment Trends - Investors are diversifying their AI chip investments beyond Nvidia, recognizing that it may not be the only significant player in the evolving AI landscape [7][14] - Companies like Coherent (COHR), Lumentum (LITE), and CoreWeave (CRWV) are highlighted as potential investment opportunities, particularly in the context of Nvidia's recent investments [10][11] Market Dynamics - The focus in 2026 is shifting towards monetization and prudent capital expenditure in AI infrastructure, which may help prevent a market bubble [8][9] - E-commerce and autonomous driving sectors are emerging as attractive areas for investment, with firms like Coupang (CPNG) and Pony AI (PONY) being noted for their potential [13][14]
英伟达AI“帝国”B面:20年收购史的“克制和清醒”
2 1 Shi Ji Jing Ji Bao Dao· 2026-03-26 12:49
Core Insights - Nvidia's strategy has evolved from being a GPU manufacturer to becoming a comprehensive AI infrastructure architect, focusing on a complete ecosystem around computing power, networking, and software platforms [2][12] - Recent investments, including $2 billion each in Lumentum and Coherent, highlight Nvidia's proactive positioning in critical segments of AI infrastructure [2] - The company's acquisition strategy has been characterized by a disciplined approach, targeting key technological nodes and industry transitions rather than merely expanding scale [2][12] Acquisition Strategy Evolution - Nvidia's early acquisitions were aimed at consolidating its GPU dominance, starting with the $70 million acquisition of 3dfx in 2000, which eliminated a major competitor and established its leadership in the GPU market [3] - Between 2004 and 2009, Nvidia expanded its GPU capabilities through various acquisitions, including PortalPlayer for mobile computing and Mental Images for ray tracing technology [3][4] - A shift occurred post-2010, where Nvidia's acquisition strategy became more aggressive and diversified, attempting to enter the mobile communication market with the $367 million acquisition of Icera, which ultimately failed [5][6] Data Center and Regulatory Challenges - The acquisition of Mellanox for $6.9 billion in 2019 marked a pivotal moment, transitioning Nvidia from a GPU manufacturer to a provider of complete data center solutions, significantly enhancing its networking capabilities [6][8] - The failed $40 billion acquisition of Arm in 2020 due to regulatory hurdles led Nvidia to adjust its strategy towards more flexible capability enhancement and ecosystem binding [7][8] - From 2019 to 2022, Nvidia solidified its data center capabilities while pivoting towards a full-stack AI infrastructure platform, making data center business a core growth engine [8][10] AI Ecosystem Focus - In recent years, Nvidia has accelerated its acquisition strategy, focusing on AI software and computing orchestration, with 83 investment actions involving 76 companies by December 2025 [10][12] - Key acquisitions include OmniML for model inference efficiency and Run:ai for AI workload scheduling, which enhance Nvidia's capabilities across the AI development lifecycle [10][11] - The company has adopted a "class acquisition" model, integrating technology and teams without traditional full acquisitions, effectively managing regulatory pressures while enhancing its technological edge [11][12] Future Outlook - Nvidia's future acquisitions will continue to focus on AI ecosystems, particularly in AI inference, computing orchestration, data security, and foundational software [13] - The company aims to optimize its "class acquisition" model to further solidify its leadership in AI computing power amidst regulatory and competitive challenges [13]
NVIDIA Accused of Hiding $1B Crypto Mining Revenue as 'Gaming' — Lawsuit Moves Forward After Supreme Court Snu
Yahoo Finance· 2026-03-26 12:45
Core Viewpoint - NVIDIA Corporation and its CEO, Jensen Huang, are facing a class-action lawsuit for allegedly concealing over $1 billion in GPU sales related to cryptocurrency mining during the 2017-2018 boom, raising concerns about the company's historical reliance on the crypto market [1][7]. Group 1: Lawsuit Details - The lawsuit claims NVIDIA misclassified significant revenue from crypto mining under its Gaming segment, misleading shareholders about the sustainability of its growth and downplaying exposure to the volatile crypto market [2]. - Plaintiffs, led by a Swedish investment firm, argue that NVIDIA executives knowingly understated the impact of crypto miners on gaming segment revenue [2]. - Internal documents and testimonies suggest NVIDIA earned between $1.1 billion and $1.35 billion more from crypto-related GPU sales than publicly disclosed, primarily through GeForce gaming cards sold to miners [3]. Group 2: Market Impact - The lawsuit alleges that the majority of crypto-related sales were through consumer GeForce cards, particularly in high-demand markets like China [4]. - Following a crash in crypto prices in late 2018, NVIDIA reduced its revenue guidance due to excess inventory and declining miner demand, resulting in a stock drop of over 28.5% in just two trading days [4]. - This decline led to a significant loss in market value and triggered the original lawsuit in 2018 [4]. Group 3: Legal Proceedings - Judge Haywood S. Gilliam Jr. certified the class after NVIDIA could not demonstrate that its statements had no effect on its stock price [5]. - An internal email from a senior vice president indicated that the company's valuation remained high due to public assurances regarding its business [5]. - A case management conference is scheduled for April 21, 2026, as the lawsuit progresses [5].