NVIDIA
Search documents
NVIDIA Launches Family of Open Reasoning AI Models for Developers and Enterprises to Build Agentic AI Platforms
Globenewswire· 2025-03-18 19:10
Core Insights - NVIDIA has launched the Llama Nemotron family of models, which are designed to provide advanced AI reasoning capabilities for developers and enterprises [1][4] - The new models enhance multistep math, coding, reasoning, and complex decision-making through extensive post-training, improving accuracy by up to 20% and optimizing inference speed by 5x compared to other leading models [2][3] Model Features - The Llama Nemotron model family is available in three sizes: Nano, Super, and Ultra, each tailored for different deployment needs, with the Nano model optimized for PCs and edge devices, the Super model for single GPU throughput, and the Ultra model for multi-GPU servers [5] - The models are built on high-quality curated synthetic data and additional datasets co-created by NVIDIA, ensuring flexibility for enterprises to develop custom reasoning models [6] Industry Collaboration - Major industry players such as Microsoft, SAP, and Accenture are collaborating with NVIDIA to integrate Llama Nemotron models into their platforms, enhancing AI capabilities across various applications [4][7][8][10] - Microsoft is incorporating these models into Azure AI Foundry, while SAP is using them to improve its Business AI solutions and AI copilot, Joule [7][8] Deployment and Accessibility - The Llama Nemotron models and NIM microservices are available as hosted APIs, with free access for NVIDIA Developer Program members for development, testing, and research [12] - Enterprises can run these models in production using NVIDIA AI Enterprise on accelerated data center and cloud infrastructure, with additional tools and software to facilitate advanced reasoning in collaborative AI systems [16]
NVIDIA Announces Isaac GR00T N1 — the World's First Open Humanoid Robot Foundation Model — and Simulation Frameworks to Speed Robot Development
Newsfilter· 2025-03-18 19:08
Core Insights - NVIDIA has launched a portfolio of technologies aimed at enhancing humanoid robot development, featuring the NVIDIA Isaac GR00T N1, which is the first open and fully customizable foundation model for humanoid reasoning and skills [1][3][11] Group 1: Technology Overview - The GR00T N1 model includes a dual-system architecture inspired by human cognition, consisting of a fast-thinking action model (System 1) and a slow-thinking decision-making model (System 2) [4][5] - GR00T N1 is designed to generalize across common tasks and can perform multistep tasks, making it applicable in various use cases such as material handling and inspection [6] - NVIDIA has introduced the Isaac GR00T Blueprint for synthetic data generation, which allows developers to create large amounts of synthetic motion data from limited human demonstrations [16][17] Group 2: Collaborations and Partnerships - NVIDIA is collaborating with Google DeepMind and Disney Research to develop Newton, an open-source physics engine that enhances robots' ability to learn complex tasks [9][10] - Disney Research plans to utilize Newton to advance its robotic character platform, aiming to create more engaging and expressive robotic characters [13][14] Group 3: Performance and Data Generation - NVIDIA generated 780,000 synthetic trajectories in 11 hours, equating to 6,500 hours of human demonstration data, which improved GR00T N1's performance by 40% when combined with real data [17] - The GR00T N1 dataset is now available as part of a larger open-source physical AI dataset, providing valuable training data for developers [18][19] Group 4: Availability and Future Developments - The GR00T N1 training data and task evaluation scenarios are available for download, with the Newton physics engine expected to be released later this year [20]
NVIDIA Announces Isaac GR00T N1 — the World's First Open Humanoid Robot Foundation Model — and Simulation Frameworks to Speed Robot Development
GlobeNewswire News Room· 2025-03-18 19:08
Core Insights - NVIDIA has launched a portfolio of technologies aimed at enhancing humanoid robot development, including the NVIDIA Isaac GR00T N1, which is the first open and fully customizable foundation model for humanoid reasoning and skills [1][3][19] Group 1: Technology Overview - The GR00T N1 model features a dual-system architecture inspired by human cognition, consisting of a fast-thinking action model ("System 1") and a slow-thinking decision-making model ("System 2") [4][5] - GR00T N1 can generalize across common tasks and perform multistep tasks, applicable in areas such as material handling, packaging, and inspection [6] - NVIDIA has introduced the Isaac GR00T Blueprint for synthetic data generation, which allows developers to create large amounts of synthetic motion data from limited human demonstrations [15][16] Group 2: Collaborations and Partnerships - NVIDIA is collaborating with Google DeepMind and Disney Research to develop Newton, an open-source physics engine designed to enhance robot learning and task handling precision [9][10] - The collaboration aims to accelerate robotics machine learning workloads by over 70 times through the development of MuJoCo-Warp [11] - Disney Research plans to utilize Newton to advance its robotic character platform, enhancing the expressiveness of next-generation entertainment robots [12][13] Group 3: Performance and Data Generation - NVIDIA generated 780,000 synthetic trajectories in 11 hours, equating to 6,500 hours of human demonstration data, which improved GR00T N1's performance by 40% when combined with real data [16] - The GR00T N1 dataset is being released as part of a larger open-source physical AI dataset, now available on Hugging Face [17] Group 4: Availability and Future Developments - The GR00T N1 training data and task evaluation scenarios are available for download, along with the Isaac GR00T Blueprint for synthetic manipulation motion generation [20] - The Newton physics engine is expected to be available later in the year, further enhancing the capabilities of humanoid robots [21]
NVIDIA Blackwell RTX PRO Comes to Workstations and Servers for Designers, Developers, Data Scientists and Creatives to Build and Collaborate With Agentic AI
GlobeNewswire News Room· 2025-03-18 19:01
Core Insights - NVIDIA has launched the RTX PRO™ Blackwell series, a new generation of workstation and server GPUs aimed at enhancing workflows for AI, technical, creative, engineering, and design professionals through advanced computing technologies [1][3] Product Overview - The RTX PRO Blackwell series includes various GPU configurations: - Data center GPU: NVIDIA RTX PRO 6000 Blackwell Server Edition - Desktop GPUs: NVIDIA RTX PRO 6000, 5000, 4500, and 4000 Blackwell editions - Laptop GPUs: NVIDIA RTX PRO 5000, 4000, 3000, 2000, 1000, and 500 Blackwell editions [4][2] Performance Enhancements - The new GPUs feature significant performance improvements: - NVIDIA Streaming Multiprocessor offers up to 1.5x faster throughput - Fourth-Generation RT Cores provide up to 2x performance for photorealistic rendering - Fifth-Generation Tensor Cores deliver up to 4,000 AI trillion operations per second [5][6][7] Memory and Bandwidth - The GPUs support larger and faster GDDR7 memory, with up to 96GB for workstations and servers, enhancing the ability to handle complex datasets [5][6] - Fifth-Generation PCIe support doubles the bandwidth over the previous generation, improving data transfer speeds [5] Multi-Instance GPU Technology - The RTX PRO 6000 and 5000 series GPUs feature Multi-Instance GPU (MIG) technology, allowing secure partitioning of a single GPU into multiple instances for efficient resource allocation [6] Industry Applications - The RTX PRO Blackwell GPUs are designed for various industries, including healthcare, manufacturing, retail, and media, providing powerful performance for AI, scientific, and visual computing applications [8][9] Customer Feedback - Early evaluations indicate significant performance improvements, such as a 5x speed increase in rendering for Foster + Partners and up to 2x GPU processing time improvement for GE HealthCare [10][11] Availability - The RTX PRO 6000 Blackwell Server Edition will be available from major data center system partners and cloud service providers later this year [14][15] - The workstation and laptop editions will be available through global distribution partners starting in April and later this year, respectively [16][17]
NVIDIA Announces DGX Spark and DGX Station Personal AI Computers
Globenewswire· 2025-03-18 18:59
Core Insights - NVIDIA has launched DGX Spark and DGX Station, personal AI supercomputers powered by the NVIDIA Grace Blackwell platform, aimed at developers, researchers, and data scientists [1][2][3] - These systems allow users to prototype, fine-tune, and run large AI models locally or on NVIDIA DGX Cloud, enhancing accessibility to advanced AI capabilities [2][7] Product Features - DGX Spark is described as the world's smallest AI supercomputer, designed to empower millions of users with high performance for generative and physical AI applications [4] - The heart of DGX Spark is the NVIDIA GB10 Grace Blackwell Superchip, which delivers up to 1,000 trillion operations per second for AI compute tasks [5] - DGX Station features the NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip, providing 784GB of coherent memory space for large-scale AI workloads [8] Technical Specifications - The GB10 Superchip utilizes NVIDIA NVLink-C2C interconnect technology, offering 5x the bandwidth of fifth-generation PCIe, optimizing performance for memory-intensive workloads [6] - DGX Station includes the NVIDIA ConnectX-8 SuperNIC, supporting networking speeds of up to 800Gb/s, facilitating high-speed connectivity for larger workloads [9] Software and Integration - Users of DGX Spark can seamlessly transition their models to DGX Cloud or other infrastructures with minimal code changes, streamlining the development process [7] - The integration of NVIDIA CUDA-X AI platform and NVIDIA NIM microservices enhances the performance and deployment of AI applications [10] Availability - Reservations for DGX Spark systems are open, while DGX Station is expected to be available later this year through manufacturing partners [11]
NVIDIA Announces Spectrum-X Photonics, Co-Packaged Optics Networking Switches to Scale AI Factories to Millions of GPUs
Globenewswire· 2025-03-18 18:44
Core Insights - NVIDIA has introduced NVIDIA Spectrum-X and NVIDIA Quantum-X silicon photonics networking switches, which significantly enhance energy efficiency and operational costs for AI factories [1][2][3] - The new networking solutions integrate optical innovations, achieving 3.5x more power efficiency and 10x better network resiliency compared to traditional methods [2][4] Group 1: Product Features - NVIDIA Spectrum-X Ethernet platform offers 1.6x bandwidth density compared to traditional Ethernet, supporting multi-tenant, hyperscale AI factories [4] - The Spectrum-X switches can deliver a total bandwidth of 100Tb/s with configurations of 128 ports of 800Gb/s or 512 ports of 200Gb/s, and up to 400Tb/s with 512 ports of 800Gb/s or 2,048 ports of 200Gb/s [4] - NVIDIA Quantum-X Photonics switches provide 144 ports of 800Gb/s InfiniBand, featuring a liquid-cooled design for efficient cooling and offering 2x faster speeds and 5x higher scalability than previous generations [5] Group 2: Industry Collaboration - NVIDIA's silicon photonics ecosystem includes collaborations with industry leaders such as TSMC, Coherent, Corning Incorporated, and Foxconn, among others [6][7] - TSMC's silicon photonics solution enhances NVIDIA's capabilities to scale AI factories to millions of GPUs, leveraging advanced chip manufacturing and 3D chip stacking technologies [7] Group 3: Availability - NVIDIA Quantum-X Photonics InfiniBand switches are expected to be available later in 2025, while Spectrum-X Photonics Ethernet switches are anticipated in 2026 [8]
NVIDIA Blackwell Ultra AI Factory Platform Paves Way for Age of AI Reasoning
Globenewswire· 2025-03-18 18:34
Core Insights - NVIDIA has introduced the Blackwell Ultra AI factory platform, enhancing AI reasoning capabilities and enabling organizations to accelerate applications in AI reasoning, agentic AI, and physical AI [1][15] - The Blackwell Ultra platform is built on the Blackwell architecture and includes the GB300 NVL72 and HGX B300 NVL16 systems, significantly increasing AI performance and revenue opportunities for AI factories [2][3] Product Features - The GB300 NVL72 system delivers 1.5 times more AI performance compared to the previous GB200 NVL72, and increases revenue opportunities by 50 times for AI factories compared to those built with NVIDIA Hopper [2] - The HGX B300 NVL16 offers 11 times faster inference on large language models, 7 times more compute, and 4 times larger memory compared to the Hopper generation [5] System Architecture - The GB300 NVL72 connects 72 Blackwell Ultra GPUs and 36 Arm Neoverse-based Grace CPUs, designed for test-time scaling and improved AI model performance [3] - Blackwell Ultra systems integrate with NVIDIA Spectrum-X Ethernet and Quantum-X800 InfiniBand platforms, providing 800 Gb/s data throughput for each GPU, enhancing AI factory and cloud data center capabilities [6] Networking and Security - NVIDIA BlueField-3 DPUs in Blackwell Ultra systems enable multi-tenant networking, GPU compute elasticity, and real-time cybersecurity threat detection [7] Market Adoption - Major technology partners including Cisco, Dell Technologies, and Hewlett Packard Enterprise are expected to deliver servers based on Blackwell Ultra products starting in the second half of 2025 [8] - Leading cloud service providers such as Amazon Web Services, Google Cloud, and Microsoft Azure will offer Blackwell Ultra-powered instances [9] Software Innovations - The NVIDIA Dynamo open-source inference framework aims to scale reasoning AI services, improving throughput and reducing response times [10][11] - Blackwell systems are optimized for running new NVIDIA Llama Nemotron Reason models and the NVIDIA AI-Q Blueprint, supported by the NVIDIA AI Enterprise software platform [12] Ecosystem and Development - The Blackwell platform is supported by NVIDIA's ecosystem of development tools, including CUDA-X libraries, with over 6 million developers and 4,000+ applications [13]
NVIDIA Dynamo Open-Source Library Accelerates and Scales AI Reasoning Models
Globenewswire· 2025-03-18 18:17
Core Insights - NVIDIA has launched NVIDIA Dynamo, an open-source inference software aimed at enhancing AI reasoning models' performance and cost efficiency in AI factories [1][3][13] - The software is designed to maximize token revenue generation by orchestrating inference requests across a large fleet of GPUs, significantly improving throughput and reducing costs [2][3][4] Performance Enhancements - NVIDIA Dynamo doubles the performance and revenue of AI factories using the same number of GPUs when serving Llama models on the NVIDIA Hopper platform [4] - The software's intelligent inference optimizations can increase the number of tokens generated by over 30 times per GPU when running the DeepSeek-R1 model [4] Key Features - NVIDIA Dynamo includes several innovations such as a GPU Planner for dynamic GPU management, a Smart Router to minimize costly recomputations, a Low-Latency Communication Library for efficient data transfer, and a Memory Manager for cost-effective data handling [14][15] - The platform supports disaggregated serving, allowing different computational phases of large language models to be optimized independently across various GPUs [9][14] Industry Adoption - Major companies like Perplexity AI and Together AI are planning to leverage NVIDIA Dynamo for enhanced inference-serving efficiencies and to meet the compute demands of new AI reasoning models [8][10][11] - The software supports various frameworks including PyTorch and NVIDIA TensorRT, facilitating its adoption across enterprises, startups, and research institutions [6][14]
META Starts Testing Community Notes: Should You Taste the Stock?
ZACKS· 2025-03-18 18:15
Meta Platforms (META) is beginning to test its previously announced Community Notes from today. The program replaces its third-party fact-checking program in the United States to promote more free speech across its platforms, Facebook, Instagram and Threads. META’s decision to loosen control over content monitoring is expected to boost user engagement. In the fourth quarter of 2024, the daily active user base continued to grow across Facebook, Instagram and WhatsApp on a year-over-year basis, both globally ...
General Motors and NVIDIA Collaborate on AI for Next-Generation Vehicle Experience and Manufacturing
Globenewswire· 2025-03-18 17:48
Core Insights - General Motors (GM) and NVIDIA are extending their collaboration to develop next-generation vehicles, factories, and robots utilizing AI, simulation, and accelerated computing [1][2] - The partnership aims to create custom AI systems for optimizing GM's manufacturing processes and enhancing vehicle safety features [2][3] Group 1: Collaboration Details - GM will leverage NVIDIA's accelerated compute platforms, including NVIDIA Omniverse and NVIDIA DRIVE AGX, to train AI manufacturing models [2][4] - The collaboration will focus on creating digital twins of assembly lines for virtual testing and production simulations, aimed at reducing downtime and increasing efficiency [4][5] - GM's future vehicles will be built on NVIDIA DRIVE AGX, which can perform up to 1,000 trillion operations per second, facilitating the development of safe autonomous vehicles [5] Group 2: Leadership Statements - Mary Barra, GM's CEO, emphasized that AI optimizes manufacturing and enhances vehicle intelligence, merging technology with human ingenuity [3] - Jensen Huang, NVIDIA's CEO, highlighted the transformative impact of their partnership on transportation and manufacturing processes [3] Group 3: Event Information - NVIDIA will host a fireside chat during the GTC global AI conference to discuss the collaboration and its implications for automotive manufacturing and software development [6]