ACES: Intro to the Grace Hopper Superchip

Overview

Instructor(s): James Chegwidden

Time: Tuesday, March 4, 2025 — 1:30PM-2:30PM CT

Location: online using Zoom

Prerequisite(s): Active ACCESS ID, general understanding of computer architecture

This short-course will discuss and examine the NVIDIA GH200 Grace Hopper Superchip. The Grace CPU architecture is designed to provide exceptional performance for AI, machine learning (ML), and scientific computing applications. Its key innovation lies in the integration of Grace processors with the Hopper GPU, forming a hybrid system optimized for massive parallelism and computational workloads.

In this short-course, we’ll explore the technical aspects of the Grace Chip, its advantages, challenges, and architectural design. The benchmarks conducted by HPRC researchers will be presented, highlighting the Grace Hopper system's performance across a variety of HPC workloads.

Course Materials

Presentation slides

  • ACES: Intro to the Grace Hopper Superchip (Spring 2025) PDF

Learning Objectives and Agenda

In this class, participants will learn:

  • The fundamental architecture and design principles of the NVIDIA GH200 Grace Hopper Superchip.
  • How the integration of Grace CPUs with Hopper GPUs creates a powerful hybrid system optimized for parallel processing.
  • Key performance metrics and real-world benchmarks of the Grace Hopper system.
  • The applications and advantages of using the Grace Hopper Superchip in high-performance computing (HPC) environments.
  • Potential challenges and considerations in deploying and utilizing the Grace Hopper system in production environments.

This course focuses, among others, on the following topics:

  • The NVIDIA GH200 Architecture
  • HPC Benchmarks and Performance
  • Energy Efficiency and Scalability
  • Challenges and Opportunities