In a significant advancement for the machine learning (ML) community, Google Cloud and NVIDIA have deepened their partnership, focusing on empowering developers to efficiently construct, escalate, and administer generative AI applications. Announced at GTC on March 18, 2024, this collaboration is set to unlock new potentials in AI development.
As part of this collaboration, Google Cloud has embraced the NVIDIA Grace Blackwell AI computing platform and the NVIDIA DGX Cloud service, marking a milestone in their joint efforts to advance AI technology. The NVIDIA H100-powered DGX Cloud platform is also now universally accessible on Google Cloud, signifying a leap in the availability of advanced AI computing resources.
Building upon their recent cooperation on the Gemma family of open models, Google will incorporate NVIDIA NIM inference microservices, offering developers a versatile platform for training and deploying AI models using their preferred tools and frameworks. Additionally, the partnership now includes support for JAX on NVIDIA GPUs and Vertex AI instances powered by NVIDIA H100 and L4 Tensor Core GPUs.
“The strength of our long-lasting partnership with NVIDIA begins at the hardware level and extends across our portfolio – from state-of-the-art GPU accelerators, to the software ecosystem, to our managed Vertex AI platform,” stated Thomas Kurian, CEO of Google Cloud. “Together with NVIDIA, our team is committed to providing a highly accessible, open and comprehensive AI platform for ML developers.”
Jensen Huang, founder and CEO of NVIDIA, highlighted the demand for generative AI solutions, stating, “Enterprises are looking for solutions that empower them to take full advantage of generative AI in weeks and months instead of years. With expanded infrastructure offerings and new integrations with NVIDIA’s full-stack AI, Google Cloud continues to provide customers with an open, flexible platform to easily scale generative AI applications.”
This enhanced partnership between NVIDIA and Google Cloud underscores their shared commitment to offering the AI community leading-edge capabilities throughout the AI stack. Highlights of the partnership expansion include the adoption of NVIDIA Grace Blackwell for real-time inference on large language models (LLMs), the introduction of Grace Blackwell-powered DGX Cloud on Google Cloud, and the integration of NVIDIA NIM inference microservices into Google Kubernetes Engine (GKE).
The collaboration further extends to the support of the NVIDIA NeMo framework across Google Cloud platforms, facilitating the deployment of generative AI models. This initiative underscores both companies’ dedication to advancing data science and analytics, with Vertex AI and Dataflow expanding support for NVIDIA GPUs.
Google Cloud and NVIDIA’s holistic partnership paves the way for AI researchers, scientists, and developers to harness the full potential of the largest and most sophisticated AI models with their preferred tools and frameworks, optimized and available on Google Cloud. This collaborative effort promises to drive innovation and accelerate the adoption of generative AI applications across industries.