AI

NVIDIA and Google infrastructure cuts AI inference costs

Google Cloud and NVIDIA logos as, at the Google Cloud Next conference, the companies outlined their hardware roadmap designed to address the cost of AI inference at scale.

At the Google Cloud Next conference, Google and NVIDIA outlined their hardware roadmap designed to address the cost of AI inference at scale.

The companies detailed the new A5X bare-metal instances, which run on NVIDIA Vera Rubin NVL72 rack-scale systems. Through hardware and software codesign,…

Read Full Article at Source