AWS and NVIDIA Announce Strategic Collaboration to Offer New Supercomputing Infrastructure, Software and Services for Generative AI

Aws And Nvidia Announce Strategic Collaboration To Offer New Supercomputing Infrastructure, Software And Services For Generative Ai

Amazon Web Services (AWS) and NVIDIA are expanding their strategic collaboration to deliver advanced infrastructure, software, and services for generative artificial intelligence (AI) innovations. AWS will be the first cloud provider to bring NVIDIA GH200 Grace Hopper Superchips to the cloud, which will be available on Amazon Elastic Compute Cloud (EC2) instances.

NVIDIA and AWS will also collaborate to host NVIDIA DGX Cloud, an AI-training-as-a-service, on AWS. The companies are partnering on Project Ceiba to design the world’s fastest GPU-powered AI supercomputer. AWS will introduce three additional new Amazon EC2 instances powered by NVIDIA GPUs for a wide set of applications such as AI fine-tuning, inference, graphics, and video workloads.

AWS and NVIDIA Collaborate on AI Supercomputer

Amazon Web Services (AWS) and NVIDIA have announced an expansion of their strategic collaboration

to deliver advanced infrastructure, software, and services to power generative artificial intelligence (AI) innovations. The collaboration will combine the best of NVIDIA and AWS technologies, including NVIDIA’s newest multi-node systems featuring next-generation GPUs, CPUs, and AI software, and AWS Nitro System advanced virtualisation and security, Elastic Fabric Adapter (EFA) interconnect, and UltraCluster scalability. These technologies are ideal for training foundation models and building generative AI applications.

The expanded collaboration builds on a longstanding relationship that has fuelled the generative AI era by offering early machine learning (ML) pioneers the compute performance required to advance the state-of-the-art in these technologies. As part of the expanded collaboration to supercharge generative AI across all industries, AWS will be the first cloud provider to bring NVIDIA® GH200 Grace Hopper Superchips with new multi-node NVLink™ technology to the cloud.

New Amazon EC2 Instances Powered by NVIDIA

AWS will introduce three additional new Amazon EC2 instances: P5e instances, powered by NVIDIA H200 Tensor Core GPUs, for large-scale and cutting-edge generative AI and HPC workloads, and G6 and G6e instances, powered by NVIDIA L4 GPUs and NVIDIA L40S GPUs, respectively, for a wide set of applications such as AI fine-tuning, inference, graphics and video workloads. G6e instances are particularly suitable for developing 3D workflows, digital twins and other applications using NVIDIA Omniverse™, a platform for connecting and building generative AI-enabled 3D applications.

AWS will be the first cloud provider to offer NVIDIA GH200 Grace Hopper Superchips with multi-node NVLink technology. Each GH200 Superchip combines an Arm-based Grace CPU with an NVIDIA Hopper™ architecture GPU on the same module. A single Amazon EC2 instance with GH200 NVL32 can provide up to 20 TB of shared memory to power terabyte-scale workloads.

AWS to Host NVIDIA DGX Cloud

AWS will collaborate with NVIDIA to host NVIDIA DGX™ Cloud—NVIDIA’s AI-training-as-a-service—on AWS. It will be the first DGX Cloud featuring GH200 NVL32, providing developers the largest shared memory in a single instance. DGX Cloud on AWS will accelerate training of cutting-edge generative AI and large language models that can reach beyond 1 trillion parameters.

Project Ceiba: The World’s Fastest GPU-Powered AI Supercomputer

NVIDIA and AWS are partnering on Project Ceiba to design the world’s fastest GPU-powered AI supercomputer—an at-scale system with GH200 NVL32 and Amazon EFA interconnect hosted by AWS for NVIDIA’s own research and development team. This first-of-its-kind supercomputer—featuring 16,384 NVIDIA GH200 Superchips and capable of processing 65 exaflops of AI—will be used by NVIDIA to propel its next wave of generative AI innovation.

NVIDIA Software on AWS to Boost Generative AI Development

NVIDIA announced software on AWS to boost generative AI development. NVIDIA NeMo™ Retriever microservice offers new tools to create highly accurate chatbots and summarisation tools using accelerated semantic retrieval. NVIDIA BioNeMo™, available on Amazon SageMaker now and coming to AWS on NVIDIA DGX Cloud, enables pharmaceutical companies to speed drug discovery by simplifying and accelerating the training of models using their own data.

“AWS and NVIDIA have collaborated for more than 13 years, beginning with the world’s first GPU cloud instance. Today, we offer the widest range of NVIDIA GPU solutions for workloads including graphics, gaming, high performance computing, machine learning, and now, generative AI,” said Adam Selipsky, CEO at AWS. “We continue to innovate with NVIDIA to make AWS the best place to run GPUs, combining next-gen NVIDIA Grace Hopper Superchips with AWS’s EFA powerful networking, EC2 UltraClusters’ hyper-scale clustering, and Nitro’s advanced virtualization capabilities.”

“Generative AI is transforming cloud workloads and putting accelerated computing at the foundation of diverse content generation,” said Jensen Huang, founder and CEO of NVIDIA. “Driven by a common mission to deliver cost-effective state-of-the-art generative AI to every customer, NVIDIA and AWS are collaborating across the entire computing stack, spanning AI infrastructure, acceleration libraries, foundation models, to generative AI services.”

Summary

“Amazon Web Services (AWS) and NVIDIA are expanding their collaboration to deliver advanced infrastructure, software and services for generative artificial intelligence (AI) innovations. This includes the introduction of the world’s fastest GPU-powered AI supercomputer, Project Ceiba, and the first cloud AI supercomputer, NVIDIA DGX Cloud, both designed to accelerate the development and training of large-scale AI and machine learning models.”

  • Amazon Web Services (AWS) and NVIDIA are expanding their strategic collaboration to advance generative artificial intelligence (AI) innovations.
  • AWS will be the first cloud provider to offer NVIDIA GH200 Grace Hopper Superchips, which will be available on Amazon Elastic Compute Cloud (EC2) instances.
  • NVIDIA and AWS will host NVIDIA DGX Cloud, an AI-training-as-a-service, on AWS. This will be the first DGX Cloud featuring GH200 NVL32, which will accelerate training of generative AI and large language models.
  • The companies are partnering on Project Ceiba to design the world’s fastest GPU-powered AI supercomputer, which will be used by NVIDIA for its own research and development.
  • AWS will introduce three new Amazon EC2 instances powered by NVIDIA GPUs for a wide range of applications, including AI fine-tuning, graphics, and video workloads.
  • AWS and NVIDIA have been collaborating for over 13 years, with AWS offering a wide range of NVIDIA GPU solutions for various workloads.
  • AWS will be the first to host NVIDIA DGX Cloud powered by Grace Hopper, and NVIDIA will use the Project Ceiba supercomputer for research and development in various fields.
  • NVIDIA software on AWS will boost generative AI development, with tools for creating chatbots and accelerating drug discovery.