- Enhanced vLLM integration on Google Cloud TPUs and availability of Red Hat AI Inference Server on Google Cloud enable more cost-effective AI inference and hardware choice for gen AI at scale
- Google Cloud joins Red Hat’s new llm-d open source project as founding contributor
- Red Hat to become early tester for Google’s open Gemma 3 model, providing Day 0 support
Red Hat, the world's leading provider of open source solutions, and Google Cloud today announced an expanded collaboration to advance AI for enterprise applications by uniting Red Hat’s open source technologies with Google Cloud’s purpose-built infrastructure and Google’s family of open models, Gemma.
Together, the companies will advance enterprise-grade use cases for scaling AI by:
- Launching the llm-d open source project with Google as a founding contributor
- Enabling support for vLLM on Google Cloud TPUs and GPU-based virtual machines (VMs) to enhance AI inference
- Delivering Day 0 support for vLLM on Gemma 3 model distributions
- Supporting Red Hat AI Inference Server on Google Cloud
- Propelling agentic AI with Red Hat as a community contributor for Google’s Agent2Agent (A2A) protocol
Bolstering AI inference with vLLM
Demonstrating its commitment to Day 0 readiness, Red Hat is now an early tester for Google’s family of open models, Gemma, starting with Gemma 3, delivering immediate support for vLLM. vLLM is an open source inference server that speeds the output of generative AI (gen AI) applications. As the leading commercial contributor to vLLM, Red Hat is driving a more cost-efficient and responsive platform for gen AI applications.
Additionally, Google Cloud TPUs, the high-performance AI accelerators powering Google’s AI portfolio, are now fully enabled on vLLM. This integration empowers developers to maximize resources while achieving the performance and efficiency crucial for fast and accurate inference.
Recognizing the shift from AI research to real-world deployment, many organizations face the complexities of a diverse AI ecosystem and the need to shift to more distributed compute strategies. To address this, Red Hat has launched the llm-d open source project, with Google as a founding contributor. Building on the momentum of the vLLM community, this initiative pioneers a new era of gen AI inference. The goal is to enable greater scalability across heterogeneous resources, optimize costs and enhance workload efficiency – all while fostering continued innovation.
Driving enterprise AI with community-powered innovation
Bringing the latest upstream community advancements to the enterprise, Red Hat AI Inference Server is now available on Google Cloud. As Red Hat’s enterprise distribution of vLLM, Red Hat AI Inference Server helps enterprises optimize model inference across their entire hybrid cloud environment. By leveraging the robust and trusted infrastructure of Google Cloud, enterprises can deploy production-ready gen AI models that are both highly responsive and cost-efficient at scale.
Underscoring their joint commitment to open AI, Red Hat is also now contributing to Google’s Agent2Agent (A2A) protocol – an application-level protocol facilitating more seamless communication between end-users or agents across diverse platforms and cloud environments. By actively participating in the A2A ecosystem, Red Hat aims to help users unlock new avenues for rapid innovation, ensuring AI workflows remain dynamic and highly effective through the power of agentic AI.
Red Hat Summit
Join the Red Hat Summit keynotes to hear the latest from Red Hat executives, customers and partners:
- Modernized infrastructure meets enterprise-ready AI — Tuesday, May 20, 8-10 a.m. EDT (YouTube)
- Hybrid cloud evolves to deliver enterprise innovation — Wednesday, May 21, 8-9:30 a.m. EDT (YouTube)
Supporting Quotes
Brian Stevens, senior vice president and Chief Technology Officer – AI, Red Hat
"With this extended collaboration, Red Hat and Google Cloud are committed to driving groundbreaking AI innovations with our combined expertise and platforms. Bringing the power of vLLM and Red Hat open source technologies to Google Cloud and Google’s Gemma equips developers with the resources they need to build more accurate, high-performing AI solutions, powered by optimized inference capabilities.”
Mark Lohmeyer, vice president and general manager, AI and Computing Infrastructure, Google Cloud
“The deepening of our collaboration with Red Hat is driven by our shared commitment to foster open innovation and bring the full potential of AI to our customers. As we enter a new age of AI inference, together we are paving the way for organizations to more effectively scale AI inference and enable agentic AI with the necessary cost-efficiency and high performance.”
Additional Resources
- Learn more about Red Hat Summit
- See all of Red Hat’s announcements this week in the Red Hat Summit newsroom
- Follow @RedHatSummit or #RHSummit on X for event-specific updates
Connect with Red Hat
- Learn more about Red Hat
- Get more news in the Red Hat newsroom
- Read the Red Hat blog
- Follow Red Hat on X
- Follow Red Hat on Instagram
- Follow Red Hat on LinkedIn
- Watch Red Hat videos on YouTube
About Red Hat
Red Hat is the open hybrid cloud technology leader, delivering a trusted, consistent and comprehensive foundation for transformative IT innovation and AI applications. Its portfolio of cloud, developer, AI, Linux, automation and application platform technologies enables any application, anywhere—from the datacenter to the edge. As the world's leading provider of enterprise open source software solutions, Red Hat invests in open ecosystems and communities to solve tomorrow's IT challenges. Collaborating with partners and customers, Red Hat helps them build, connect, automate, secure and manage their IT environments, supported by consulting services and award-winning training and certification offerings.
About Google Cloud
Google Cloud is the new way to the cloud, providing AI, infrastructure, developer, data, security, and collaboration tools built for today and tomorrow. Google Cloud offers a powerful, fully integrated and optimized AI stack with its own planet-scale infrastructure, custom-built chips, generative AI models and development platform, as well as AI-powered applications, to help organizations transform. Customers in more than 200 countries and territories turn to Google Cloud as their trusted technology partner.
Forward-Looking Statements
Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company’s current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.
Red Hat and the Red Hat logo are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the U.S. and other countries.
View source version on businesswire.com: https://www.businesswire.com/news/home/20250520022385/en/
Contacts
Media Contact:
Jessie Beach, Red Hat
+1 (919) 602-2836
jbeach@redhat.com
Google Cloud
press@google.com