menu
techminis

A naukri.com initiative

google-web-stories
Home

>

Cloud News

>

Nvidia and...
source image

Siliconangle

2w

read

204

img
dot

Image Credit: Siliconangle

Nvidia and AWS team up to accelerate AI deployments in the cloud

  • Nvidia and Amazon Web Services (AWS) have announced new developments in their artificial intelligence (AI) and quantum computing partnership.
  • Nvidia’s NIM microservices are now available on various AWS AI services, enabling faster inference with lower latency for AI developers. The NIM microservices are available as prebuilt containers and they come with a choice of inference engines.
  • Developers are also getting access to a new infrastructure offering, namely the Nvidia DGX Cloud, which is now available through AWS Marketplace Private Offers, giving customers access to a fully-managed, high-performance compute platform for training, customising and deploying AI models.
  • Nvidia is making a number of new AI Blueprints available for instant deployment on AWS. These provide ready-to-deploy AI agents for various tasks, such as video search and text summarisation, that can easily be integrated into existing developer workflows.
  • The two companies are looking to accelerate AI-powered simulations for use in robotics. Nvidia’s Omniverse platform will provide developers with a reference application to create virtual environments and digital twins, using highly realistic physics to simulate AI-powered robots in any environment.
  • Nvidia’s BioNeMo NIM microservices and AI Blueprints for advancing drug discovery are now available with AWS HealthOmics, giving researchers the chance to experiment with more AI models.
  • Nvidia’s CUDA-Q platform, which is used to develop “hybrid quantum/classical computing applications” that span traditional and quantum computers, is being integrated with AWS’ Braket service to accelerate quantum computing development.
  • The Nvidia DGX Cloud platform currently provides access to Nvidia’s most powerful GPUs, the Nvidia H100 and H200, and will soon be expanded to include the next-generation Blackwell GPUs, slated to launch in the new year. AWS said the Blackwell chips will be available as part of the GB200 NVL supercomputing system.
  • Users will be able to deploy the NIM microservices across multiple AWS services, including Amazon Elastic Compute Cloud, Amazon SageMaker and the Amazon Elastic Kubernetes Service.
  • Customers will also get direct access to the company’s experts, who will be on hand to provide the technical expertise needed to scale their AI deployments.

Read Full Article

like

12 Likes

For uninterrupted reading, download the app