Summary of "NVIDIA SC25 Fireside Chat"

Summary of NVIDIA SC25 Fireside Chat

Overview

Ian Buck, NVIDIA’s Vice President and General Manager of Hyperscale and HPC, delivered a keynote reflecting on NVIDIA’s supercomputing journey since 2006. He highlighted the evolution and future directions of accelerated computing in simulation, AI, and quantum computing. The session included a panel discussion with leaders from major supercomputing centers and labs, focusing on technological advances, ecosystem development, and future challenges.


Key Technological Concepts and Product Features

  1. NVIDIA’s Supercomputing Evolution

    • CUDA introduced in 2006 at SC (Supercomputing Conference).
    • Innovations such as immersion cooling (2010), MVLink interconnects (2016), and the GraceHopper CPU-GPU system (2022).
    • Transition from PCIe GPU cards to side-mounted GPUs for better IO and multi-GPU connectivity.
    • Current dominance: 88 of the top 100 supercomputers accelerated with NVIDIA GPUs.
  2. Three Pillars of Modern Supercomputing

    • Simulation: Traditional scientific simulations of natural phenomena at scales/timeframes impossible for physical experiments.
    • AI: Explosive growth since 2016, applying AI to scientific discovery (e.g., climate modeling, CFD, seismology, drug discovery).
    • Quantum Computing: Early stage, with GPUs helping simulate quantum algorithms (e.g., Shor’s algorithm at 50 qubits) and quantum processor design.
  3. CUDA-X Ecosystem

    • Over 6,000 CUDA-accelerated applications across disciplines.
    • 600 million downloads, 7 million developers, 1,000+ SDKs and AI models.
    • Specialized libraries for different domains: KU Litho (circuit modeling), KUP (path planning), Physics Nemo (AI physics), Hollow Scan (radio astronomy), Alchemy (material science).
  4. Notable Application Highlights

    • Max Planck Institute’s Earth climate simulation at 1 km resolution simulating 146 days per day on 20,000+ Grace Hopper chips.
    • Luminary Cloud’s AI physics model for spacecraft nozzle design reducing CFD runtime from hours to seconds.
    • NASCAR CFD simulations accelerated from 11 hours to 48 minutes using NVIDIA GPUs.
    • Real-time interactive digital twin of fusion reactor (General Atomics) built in NVIDIA Omniverse, integrating sensor data, physics, engineering, and AI.
    • Real-time tsunami forecasting with AI-powered digital twins using seismic data.
  5. NVIDIA Warp

    • A Python package for physics simulations.
    • Supports automatic differentiation enabling integration with AI frameworks (PyTorch, JAX).
    • Allows physics simulations that can be optimized and solved interactively.
  6. Apollo AI Models

    • A new family of pre-trained AI models focused on physics-based simulations (CFD, manufacturing, structural analysis).
    • Demonstrated with aircraft wing design reducing simulation time from 8 hours to 5 seconds on RTX GPUs.
    • Open-source with workflows for fine-tuning and inference.
  7. Quantum Computing Integration

    • GPU supercomputers simulate quantum algorithms at unprecedented scale.
    • NVIDIA Q-Link: a reference architecture to process quantum processor signals.
    • Collaboration with 23+ supercomputing centers worldwide on hybrid quantum-GPU-CPU systems.
    • Full-chip simulation of quantum processors using thousands of GPUs.
  8. MVLink and Interconnect Innovations

    • MVLink: a high-throughput, low-latency shared memory fabric connecting GPUs, CPUs, and accelerators.
    • Enables 10x performance improvement in AI inference throughput compared to previous generations.
    • Upcoming systems (Reuben Ultra) will scale to 576 chips interconnected via MVLink.
    • Introduction of co-packaged optics to reduce power consumption and improve reliability in data center interconnects.
    • MVLink Fusion IP enables third-party CPUs/accelerators (e.g., ARM) to join the MVLink ecosystem.
  9. NVIDIA Vera CPU

    • Next-gen NVIDIA-designed ARM-based CPU core (Olympus).
    • Twice the performance of Grace CPU.
    • Supports 1.2–2 TB/s LPDDR5X memory bandwidth.
    • Integrated chip-to-chip MVLink interconnect at 1.5 TB/s.
    • Energy-efficient and optimized for HPC workloads like CFD and weather modeling.
    • Central to NSF Horizon supercomputer.
  10. New and Upcoming Supercomputers - Horizon (NSF’s largest academic supercomputer in the US) combining Grace Blackwell GPUs and Vera CPUs. - Solstice and Equinox at Argonne National Labs with 100,000 and 10,000 GPUs respectively. - Fugaku Next in Japan using NVIDIA GPUs and Fujitsu CPUs. - European AI factories and gigafactories with 100,000+ GPUs as part of public-private partnerships. - 80+ new supercomputers announced at SC25 with over 300,000 GPUs and multiple zettaflops of compute.


Analysis and Insights from Panel Discussion


Tutorials, Guides, and Resources Highlighted


Main Speakers / Sources


Conclusion

The NVIDIA SC25 Fireside Chat showcased the rapid evolution of supercomputing driven by NVIDIA’s innovations in GPU acceleration, AI integration, and emerging quantum computing. The discussion emphasized the importance of hybrid AI-HPC workflows, advanced interconnects like MVLink, and the growing software ecosystem that supports scientific discovery across simulation, AI, and quantum domains.

Panelists highlighted ongoing challenges in power, data management, and software maturity, while expressing optimism about the future of supercomputing as a critical national and global infrastructure for science and industry.

Category ?

Technology


Share this summary


Is the summary off?

If you think the summary is inaccurate, you can reprocess it with the latest model.

Video