NVIDIA, Google Cloud Enter Into Collaboration to Accelerate GenAI Development

NVIDIA and Google Cloud announced a new collaboration to help startups, worldwide, to accelerate the creation of GenAI applications and services. This announcement, which was issued at Google Cloud Next ‘24 in Las Vegas, brings together NVIDIA’s Inception program for startups and the Google for Startups cloud program to widen access to cloud credits, go-to-market support and technical expertise.

Qualified members of NVIDIA Inception, a global program that supports over 18,000
startups, will gain an accelerated path to using Google Cloud infrastructure with access to
Google Cloud credits, offering up to $350,000 for focusing on AI.

Google for Startups cloud program members can join NVIDIA Inception to gain access to
technological expertise, NVIDIA Deep Learning Institute course credits and NVIDIA hardware and software. Eligible members can also participate in NVIDIA Inception Capital Connect, a platform that gives startups exposure to venture capital firms that are interested in the space.

Emerging software makers also gain access to fast-tracked onboarding to Google Cloud Marketplace, co-marketing and product acceleration support.

Teams from both companies worked closely together to accelerate the performance of Gemma — built from the same research and technology used to create Google DeepMind’s Gemini model — with NVIDIA TensorRT-LLM, an open-source library for optimizing LLM inference when running on NVIDIA GPUs. NVIDIA NIM microservices, which are part of the NVIDIA AI Enterprise software platform, together with the Google Kubernetes Engine provide a streamlined path for developing AI-powered apps and deploying optimized AI models into production. Built on inference engines including the NVIDIA Triton inference server and TensorRT-LLM, NIM supports leading AI models and delivers seamless, scalable AI inferencing to accelerate generative AI deployment in enterprises.

The Gemma family of models including Gemma 7B, RecurrentGemma and CodeGemma, all
available from the NVIDIA API catalog for users to try from a browser, prototype with the API endpoints and self-host with NIM.

Google Cloud has made it easier to deploy the NVIDIA NeMo framework across its platform
via the GKE and Google Cloud HPC Toolkit. This enables developers to automate and scale the training and serving of generative AI models, allowing them to rapidly deploy turnkey
environments through customizable blueprints that jump-start the development process.

NVIDIA NeMo, part of NVIDIA AI Enterprise, is also available in Google Cloud Marketplace,
providing customers another way to easily access NeMo and other frameworks to
accelerate AI development.

Click here and here, respectively, for information on NVIDIA’s and Google Cloud’s partner programs.