AWS re:Invent 2023 - Customer Keynote Fireside Chat with NVIDIA

AWS re:Invent 2023 - Customer Keynote Fireside Chat with NVIDIA

April 16, 2024
Share
Author: Big Y

AWS and NVIDIA Partnership: Bringing Advanced Infrastructure for Generative AI Workloads with GPUs

In recent years, the demand for generative AI workloads has increased significantly. To meet this demand, AWS and NVIDIA have been working together to bring the most advanced infrastructure for generative AI workloads with GPUs. In this article, we will discuss the partnership between AWS and NVIDIA, the latest innovations, and how they are going to deliver the most advanced infrastructure for generative AI workloads with GPUs.

Table of Contents

1. Introduction

2. AWS and NVIDIA Partnership

3. Expanding the Partnership with More Innovations

4. The Latest GPUs: L4, L40S, and H200

5. The Power of ARM Processors

6. NVIDIA DGX Cloud on AWS

7. Project Ceiba: The Largest AI Factory NVIDIA Has Ever Built

8. Off-the-Shelf Models and Custom AIs

9. Benefits of the Partnership

10. Conclusion

Introduction

AWS and NVIDIA have been working together for a long time to bring the most advanced infrastructure for generative AI workloads with GPUs. The collaboration between the two companies has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing.

AWS and NVIDIA Partnership

AWS and NVIDIA have been collaborating for a long time to bring the most advanced infrastructure for generative AI workloads with GPUs. The partnership has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing.

Expanding the Partnership with More Innovations

AWS and NVIDIA are expanding their partnership with more innovations that are going to deliver the most advanced infrastructure for generative AI workloads with GPUs. The collaboration between the two companies has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing.

The Latest GPUs: L4, L40S, and H200

AWS and NVIDIA are deploying a whole new family of GPUs: L4, L40S, and the brand new H200. The H200 is a combination of the brand new Tensor RT LLM optimizing compilers for generative AI and improves the throughput of inference, large language model inference, by a factor of four, reducing the cost in just one year by a factor of four.

The Power of ARM Processors

AWS and NVIDIA are passionate about ARM processors. ARM processors are incredibly low energy and cost-effective. The Grace Hopper, which is GH200, connects two revolutionary processors together in a unique way. It connects them together using a chip-to-chip interconnect called NVLink at one terabyte per second, and it's connected in a coherent way, so that the GPU could access all of the CPU's memory, the CPU can access all of the GPU's memory, and so the two of these processors could work hand-in-hand in a really fast way.

NVIDIA DGX Cloud on AWS

AWS and NVIDIA are partnering to bring the NVIDIA DGX Cloud to AWS. DGX Cloud is NVIDIA's AI factory. This is how their researchers advance AI. They use AI to do neural graphics. The way they do computer graphics today is impossible without AI. They use their AI factories to advance their digital biology models, their large language models, use it for robotics, for self-driving cars. They use it to simulate Earth Two, to a digital twin of the Earth to predict weather and climate.

Project Ceiba: The Largest AI Factory NVIDIA Has Ever Built

AWS and NVIDIA are building the largest AI factory NVIDIA has ever built. They call it Project Ceiba. Ceiba is going to be 16,384 GPUs connected into one giant AI supercomputer. This is utterly incredible. They will be able to reduce the training time of the largest language models, the next generation MOEs, these large, extremely large mixture of experts models, and be able to train it in just half of the time.

Off-the-Shelf Models and Custom AIs

AWS and NVIDIA are bringing off-the-shelf models and custom AIs to customers. It's great to be able to use off-the-shelf models. And there's going to be a whole bunch of off the shelf models in public clouds, it's going to be available in software platform companies, SaaS companies, for example, Salesforce that you mentioned earlier, they'll have all kinds of off-the-shelf co-pilots and generative AI models. But a lot of companies need to build their own proprietary models. They set up the AI factory so that they could partner with them to help them create their custom AIs, and now it'll run all on AWS.

Benefits of the Partnership

The partnership between AWS and NVIDIA has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing. The collaboration between the two companies has resulted in the deployment of a whole new family of GPUs: L4, L40S, and the brand new H200. The NVIDIA DGX Cloud is coming to AWS, and they are building the largest AI factory NVIDIA has ever built.

Conclusion

AWS and NVIDIA are working together to bring the most advanced infrastructure for generative AI workloads with GPUs. The partnership between the two companies has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing. The collaboration between the two companies has resulted in the deployment of a whole new family of GPUs: L4, L40S, and the brand new H200. The NVIDIA DGX Cloud is coming to AWS, and they are building the largest AI factory NVIDIA has ever built.

Highlights

- AWS and NVIDIA are working together to bring the most advanced infrastructure for generative AI workloads with GPUs.

- The partnership between the two companies has resulted in the deployment of two million GPUs in AWS in the last several years alone.

- AWS and NVIDIA are deploying a whole new family of GPUs: L4, L40S, and the brand new H200.

- The NVIDIA DGX Cloud is coming to AWS, and they are building the largest AI factory NVIDIA has ever built.

FAQ

Q: What is the partnership between AWS and NVIDIA?

A: AWS and NVIDIA are working together to bring the most advanced infrastructure for generative AI workloads with GPUs.

Q: What are the latest GPUs deployed by AWS and NVIDIA?

A: AWS and NVIDIA are deploying a whole new family of GPUs: L4, L40S, and the brand new H200.

Q: What is the NVIDIA DGX Cloud?

A: The NVIDIA DGX Cloud is NVIDIA's AI factory. This is how their researchers advance AI.

Q: What is Project Ceiba?

A: Project Ceiba is the largest AI factory NVIDIA has ever built. It will be 16,384 GPUs connected into one giant AI supercomputer.

Q: What are the benefits of the partnership between AWS and NVIDIA?

A: The partnership between AWS and NVIDIA has resulted in the deployment of two million GPUs in AWS in the last several years alone. This has made AWS the world's first cloud to recognize the importance of GPU-accelerated computing.

- End -
VOC AI Inc. 8 The Green,Ste A, in the City of Dover County of Kent, Delaware Zip Code: 19901 Copyright © 2024 VOC AI Inc.All Rights Reserved. Terms & Conditions Privacy Policy
This website uses cookies
VOC AI uses cookies to ensure the website works properly, to store some information about your preferences, devices, and past actions. This data is aggregated or statistical, which means that we will not be able to identify you individually. You can find more details about the cookies we use and how to withdraw consent in our Privacy Policy.
We use Google Analytics to improve user experience on our website. By continuing to use our site, you consent to the use of cookies and data collection by Google Analytics.
Are you happy to accept these cookies?
Accept all cookies
Reject all cookies