Workflow
AI Factories
icon
Search documents
How AI Factories Maximize Tokens, Power, and Profit With NVIDIA DSX
NVIDIA· 2026-03-18 15:48
The greatest infrastructure buildout in history [music] is underway. The world is racing to build chip system and AI [music] factories and every month of delay costs billions in lost revenues. AI factory revenues are equal to tokens per watt.So with power constraints, every [music] unused watt is revenue lost. NVIDIA DSX [music] is an Omniverse digital twin blueprint for designing and [music] operating AI factories for maximum token throughput, resilience, and energy efficiency. Developers [music] connect t ...
Nvidia's Jensen Huang Says AI Compute Could Near $1 Trillion by 2027
PYMNTS.com· 2026-03-17 01:23
Core Insights - The AI industry is at an "inference inflection point," where the demand for computing power is rapidly shifting from training AI models to running them in real-world applications [5][17] - Nvidia's CEO Jensen Huang highlighted that AI computing could approach a trillion dollars in data center infrastructure investments between now and 2027 [5] - The concept of "AI factories" is emerging, which are specialized data centers designed to generate AI outputs at scale, with intelligence tokens becoming the new currency [15][17] Inference and Token Economy - Inference is the process where trained AI models generate responses, and the demand for computing resources during inference can exceed that needed for training [6][7] - Tokens are the basic units of AI-generated text or data, and the efficiency of generating tokens at scale is becoming crucial for the long-term economics of AI [6][7][11] - Huang emphasized that "inference is your new workload, tokens are your new commodity," indicating a shift in how companies should optimize their architecture for future demands [11] AI Factories and Infrastructure Boom - Nvidia introduced its next-generation AI computing platform, Vera Rubin, which aims to deliver up to 10 times higher inference performance per watt and reduce token generation costs by approximately 90% [16] - The shift towards inference-driven workloads is transforming the technology industry's approach to computing infrastructure, moving from periodic model training to continuous token generation [17] - Huang stated that the future of computing will revolve around AI factories, fundamentally redefining the economics of computing [17]
Schneider Electric teams with NVIDIA to develop validated blueprints to design, simulate, build, operate and maintain gigawatt-scale AI Factories
Globenewswire· 2026-03-16 21:00
Core Insights - Schneider Electric, in collaboration with NVIDIA and AVEVA, announced advancements in AI data center infrastructure design, simulation, and operation during NVIDIA GTC 2026 [1][2] Group 1: NVIDIA Vera Rubin Reference Design - The new NVIDIA Vera Rubin reference design is validated for power and cooling of NVIDIA's NVL72 racks, addressing infrastructure requirements for rack-scale systems [3] - The design utilizes ETAP models for electrical system design and ITD CFD models for layout and airflow [3] Group 2: AVEVA Lifecycle Digital Twin Architecture - AVEVA and NVIDIA introduced a lifecycle digital twin architecture to enhance GPU efficiency and accelerate AI factory deployment [4] - The integration of AVEVA's software within the NVIDIA Omniverse DSX Blueprint is expected to optimize engineering processes and reduce time-to-token through simulations and collaborative design tools [4][5] Group 3: Agentic AI for Alarm Management - Schneider Electric tested the NVIDIA Nemotron model for a new agentic AI alarm management capability, which autonomously analyzes and diagnoses system alarms [7] - This technology aims to improve operational resilience and reduce unnecessary dispatches by providing faster issue resolution [7] Group 4: Historical Collaboration - The latest announcements build on a legacy of innovation between Schneider Electric and NVIDIA, focusing on validated blueprints for gigawatt-scale AI factories [8][10] Group 5: Enhanced Power Distribution - The new design allows for increased supply voltage of 480 VAC and supports higher TCS loop supply temperature of 45°C, enhancing efficiency [11] - It enables a new IT room architecture that clusters AI racks for optimized power delivery and performance [11]
NVIDIA GTC Keynote 2026
NVIDIA· 2026-03-12 22:17
Watch NVIDIA Founder and CEO Jensen Huang’s GTC keynote as he unveils the latest breakthroughs in AI and accelerated computing. See how agentic AI, AI factories, and physical AI are powering the next generation of intelligent systems. ...
Palo Alto Networks and Global Partners Announce Secure by Design AI Factories
Prnewswire· 2026-03-02 05:12
Core Insights - Palo Alto Networks has announced an expanded security ecosystem aimed at protecting AI Factories, which are essential for the new industrial backbone of high-performance AI [1] - The company has formed collaborations with Nokia, U Mobile, Aeris, and Celerway to enhance security for sovereign AI and autonomous edge operations without compromising performance [1] Group 1: Partnerships and Collaborations - The collaboration with Nokia focuses on integrating data center security with AI platforms to support the development of European 'Gigafactories', allowing customers to scale AI workloads while meeting data sovereignty requirements [1] - The partnership with Celerway extends enterprise security to remote teams, providing data-center-class protection for mission-critical 5G edge devices in challenging environments [1] - Aeris integration allows for unified visibility across global IoT fleets, enabling industries to scale AI and 5G initiatives while minimizing security risks associated with billions of devices [1] - The MoU with U Mobile aims to deliver a network-embedded Security-as-a-Service solution, enhancing cybersecurity for consumers and businesses through integrated Next-Generation Firewalls and AI-powered security [1] Group 2: Technological Advancements - Palo Alto Networks is leveraging AI-powered security services to create a secure digital infrastructure capable of managing multi-terabit throughput necessary for training AI models [1] - The company emphasizes the importance of a comprehensive architectural framework that extends security solutions from the network layer to workloads, ensuring the integrity of industrial digitization [1]
Nvidia Puts Another $2B Into CoreWeave, Offers New Chips | Bloomberg Tech 1/26/2026
Bloomberg Technology· 2026-01-26 19:28
>> "BLOOMBERG TECH" IS LIVE FROM COAST-TO-COAST WITH CAROLINE HYDE IN NEW YORK AND ED LUDLOW IN SAN FRANCISCO. ED: THIS IS "BLOOMBERG TECH. " COMING UP, NVIDIA INVESTED ADDITIONAL $2 BILLION IN COREWEAVE, EXPANDING THEIR PARTNERSHIP TO ACCELERATE THE BUILDOUT OF AI FACTORIES. CAROLINE: PLUS, CONTINUING THE ACQUISITION SPREE AS THE QUANTUM COMPANY BUYS A CHIPMAKER, SKY WANTED TECHNOLOGY, IN A DEAL WORTH $1.8% BILLION. ED: GIVING UP FOR BIG TECH EARNINGS, META, MICROSOFT, TESLA, APPLE ALSO TO REPORT EARNINGS ...
Optimizing AI Factories
NVIDIA· 2025-12-06 01:14
Data Center Construction - The industry needs to consider power train, thermal chain, and prefabrication to industrialize data center construction in unprecedented ways [1]
Amazon Closing The Gap In AI Race: Analysts
Benzinga· 2025-12-03 20:25
Core Viewpoint - Amazon.com Inc has received positive support from Wall Street due to its "agent-driven" AI strategy and advancements in custom chip technology showcased at the AWS re:Invent conference [1][2]. Group 1: AI Strategy and Innovations - Amazon's AWS is focusing on an agent-driven future, with CEO Matt Garman predicting the deployment of "billions" of autonomous agents across enterprises [2]. - The introduction of new frontier agents for security, DevOps, and continuity is a significant development in Amazon's AI capabilities [2]. - The concept of "AI Factories" allows customers to deploy dedicated AWS infrastructure, including Nvidia and Trainium chips, into their own data centers for enhanced performance [3]. Group 2: Revenue Growth Projections - Analysts expect AWS revenue growth to accelerate towards 25% by 2026, driven by increased capacity and demand for AI solutions [4]. - JP Morgan's analyst projects AWS revenue growth of 23% for both Q4 and 2026, indicating a potentially conservative estimate [5]. - Amazon's AWS is already surpassing a $130 billion run rate and is expected to see a 22% year-over-year growth next quarter as demand for AI services increases [10][11]. Group 3: Competitive Positioning - Amazon is narrowing the competitive gap in generative AI through advancements in its custom Trainium chips and partnerships with companies like Anthropic and OpenAI [5]. - The general availability of Trainium 3, which offers 4.4 times the compute performance of its predecessor, is a key factor for cost-effective AI deployment [6]. - The launch of the Nova 2 foundation models and AWS AI Factories is expected to enhance Amazon's ecosystem and accelerate AWS momentum [7][9]. Group 4: Analyst Ratings and Price Forecasts - Bank of America Securities raised its price forecast for Amazon from $272 to $303, maintaining a Buy rating [8]. - JP Morgan reiterated an Overweight rating with a price forecast of $305, while Wedbush set a price target of $340, reflecting strong confidence in Amazon's growth trajectory [8][9].
X @TechCrunch
TechCrunch· 2025-12-03 00:45
Industry Trend - Amazon is challenging competitors by offering on-premises Nvidia 'AI Factories' [1]
How DDN Supercharges GPU Productivity for Training, Inference & AI Factories | James Coomer
DDN· 2025-12-02 17:48
AI Infrastructure Challenges & Solutions - Data bottlenecks constrain GPU performance in AI training and inference, leading to wasted resources and reduced productivity [2][4][5][11] - DDN addresses these bottlenecks by optimizing data movement through fast storage systems and integration with AI frameworks and hardware like Nvidia [5][6] - Inference is becoming increasingly important, with spending expected to surpass training systems, posing challenges in model loading, RAG (Retrieval Augmented Generation), and KV cache management [7][8][9] - DDN Core combines Exascaler for training and Infinia for data management to provide a seamless AI experience [13][14] DDN's Value Proposition - DDN's solutions improve data center efficiency by increasing "answers per watt," delivering more compute with less energy consumption [12][13] - DDN handles KV cache, increasing the effective memory of GPU systems and improving productivity by up to 60% in large-scale GPU data centers [9][10] - DDN offers fast-track solutions for enterprises to adopt AI, whether on the cloud or on-premise, through partnerships like the one with Google Cloud [15][16][17] - DDN's platform supports various use cases, including HPC, AI training and inference, research data management, and secure data sharing [19][20] Strategic Considerations - DDN emphasizes the importance of considering data first when building AI at scale, advocating for data desiloing and secure access [28][29] - DDN supports sovereign AI, enabling nations to develop AI models relevant to their specific data, language, and culture while ensuring security and data sharing [20][21][22] - Partnerships are crucial for delivering efficient AI solutions tailored to customer preferences, whether cloud, on-premise, or hybrid [23][24] - AI factories, which integrate data preparation, training, simulation, and production, present complex data challenges where DDN excels [25][26][27]