Accelerate innovation with Google Cloud and NVIDIA
NVIDIA and Google Cloud deliver accelerator-optimized solutions that address your most demanding workloads, including machine learning, high performance computing, data analytics, graphics, and gaming workloads.
Google Cloud at NVIDIA GTC 2026-Elite sponsor
Join us at GTC 2026 to see how Google Cloud and NVIDIA are architecting the future of AI scale, security, and agentic workflows. Connect with our industry and product leaders during our exclusive sessions and happy hour to explore the breakthroughs shaping the next frontier of innovation.
Marquee and fireside chats
Infrastructure and scalability
AI agents and applied research
Security and confidential computing
Hands-on labs and tutorials
Networking and ecosystem
“Attending GTC with my team was invaluable, the ability to meet with so many customers alongside our product and engineering partners—it's set us up for success in 2025”
Lauren Kapnick, Head of Sales Hedge Funds FSI, Google Cloud
High Performing GPUs on Google Cloud
Accelerate machine learning, scientific computing, and generative AI with high-performance GPUs on Google Cloud.
Key Benefits:
Key Features
NVIDIA technologies on Google Cloud
Google Kubernetes Engine (GKE)
Leverage GKE's scalability, NVIDIA Multi-Instance GPU (MIG) support, and GPU time-sharing for efficient generative AI training, inference, and other compute-intensive workloads. Optimize resource utilization and minimize operational costs.
Vertex AI
Combine NVIDIA accelerated computing with Vertex AI, a unified MLOps platform. Utilize NVIDIA GPUs and AI software (such as, Triton™ Inference Server) within Vertex AI Training, Prediction, Pipelines, and Notebooks to accelerate generative AI development and deployment without infrastructure complexities.
Cloud Run
Deploy generative AI faster with NVIDIA NIM on Cloud Run, a fully managed serverless platform. Cloud Run's GPU support allows NIM to optimize performance and accelerate gen AI model deployment in a serverless environment.
Dynamic Workload Scheduler
Access NVIDIA GPU capacity on Google Cloud for short-duration AI workloads (training, fine-tuning, experimentation). Flexible scheduling and atomic provisioning enhance resource utilization and optimize costs across services like GKE, Vertex AI, and Batch.
Google Distributed Cloud
The NVIDIA Blackwell platform on Google Distributed Cloud enables secure, on-premises deployment of advanced agentic AI (including Google Gemini models). This offers breakthrough AI performance and scalability for sensitive, regulated workloads, ensuring data privacy, sovereignty, and compliance.
Technical resources for deploying NVIDIA technologies on Google Cloud
Google Cloud basics
Tutorials
Google Cloud at NVIDIA GTC 2026-Elite sponsor
Join us at GTC 2026 to see how Google Cloud and NVIDIA are architecting the future of AI scale, security, and agentic workflows. Connect with our industry and product leaders during our exclusive sessions and happy hour to explore the breakthroughs shaping the next frontier of innovation.
Marquee and fireside chats
Infrastructure and scalability
AI agents and applied research
Security and confidential computing
Hands-on labs and tutorials
Networking and ecosystem
“Attending GTC with my team was invaluable, the ability to meet with so many customers alongside our product and engineering partners—it's set us up for success in 2025”
Lauren Kapnick, Head of Sales Hedge Funds FSI, Google Cloud
High Performing GPUs on Google Cloud
Accelerate machine learning, scientific computing, and generative AI with high-performance GPUs on Google Cloud.
Key Benefits:
Key Features
NVIDIA technologies on Google Cloud
Google Kubernetes Engine (GKE)
Leverage GKE's scalability, NVIDIA Multi-Instance GPU (MIG) support, and GPU time-sharing for efficient generative AI training, inference, and other compute-intensive workloads. Optimize resource utilization and minimize operational costs.
Vertex AI
Combine NVIDIA accelerated computing with Vertex AI, a unified MLOps platform. Utilize NVIDIA GPUs and AI software (such as, Triton™ Inference Server) within Vertex AI Training, Prediction, Pipelines, and Notebooks to accelerate generative AI development and deployment without infrastructure complexities.
Cloud Run
Deploy generative AI faster with NVIDIA NIM on Cloud Run, a fully managed serverless platform. Cloud Run's GPU support allows NIM to optimize performance and accelerate gen AI model deployment in a serverless environment.
Dynamic Workload Scheduler
Access NVIDIA GPU capacity on Google Cloud for short-duration AI workloads (training, fine-tuning, experimentation). Flexible scheduling and atomic provisioning enhance resource utilization and optimize costs across services like GKE, Vertex AI, and Batch.
Google Distributed Cloud
The NVIDIA Blackwell platform on Google Distributed Cloud enables secure, on-premises deployment of advanced agentic AI (including Google Gemini models). This offers breakthrough AI performance and scalability for sensitive, regulated workloads, ensuring data privacy, sovereignty, and compliance.
Technical resources for deploying NVIDIA technologies on Google Cloud
Google Cloud basics
Tutorials





Compared with another inference platform, running on GKE with NVIDIA NIM and GPUs delivered 6.1x acceleration in average answer/response generation speed for the Amazfit AI agent
Jia Li Co-Founder, Chief AI Officer, LiveX AI