UPDATED 15:43 EDT / MAY 21 2025

Chris Wright, chief technology officer and senior vice president of global engineering at Red Hat, talks with theCUBE about strategic updates during Red Hat Summit – 2025. AI

Red Hat sharpens its AI game with vLLM and llm-d

Red Hat Inc. used its annual conference this week to spotlight a series of strategic updates, including vLLM and llm-d — two tools aimed at streamlining AI production across enterprise environments. The company’s approach supports scalable, cost-effective workloads while helping customers evolve their AI strategies without overhauling infrastructure.

Chris Wright, chief technology officer and senior vice president of global engineering at Red Hat, talks with theCUBE about strategic updates during Red Hat Summit - 2025.

Red Hat’s Chris Wright talks with theCUBE about the new open-source projects Red Hat is driving for enterprise AI.

“We can feel there’s a ton of potential in all this opportunity that sits in the space of AI and how you use AI to either improve your business or improve your operations,” said Chris Wright (pictured), chief technology officer and senior vice president of global engineering at Red Hat. “How do we sort of gracefully bring those worlds together? In technology transitions, it’s never healthy to throw everything away and start over. In general, we have to have some kind of evolutionary path. So, we’re focused on that.”

Wright spoke with theCUBE’s Rebecca Knight and Rob Strechay at Red Hat Summit, during an exclusive broadcast on theCUBE, SiliconANGLE Media’s livestreaming studio. They discussed new open-source projects that Red Hat is driving for enterprise AI. (* Disclosure below.)

Open-source strategic updates

Part of that evolutionary path will involve leveraging existing open-source projects to streamline the use of AI models, according to Wright. One such project for Red Hat is vLLM, a library of open-source code that performs the functions of an inference server.

“vLLM is important from our point of view in the context of operating a model to generate results, call it inference, and that is what’s required to put models into production,” he said. “A lot of focus has been on training, building the largest possible frontier models, multi-trillion parameters. We’re really focused in on vLLM as this way to bring that work into a production environment.”

The Red Hat Summit also provided an opportunity for the IBM Corp. subsidiary to launch llm-d, an open-source collaboration with partners such as Google Cloud, Nvidia Corp., Cisco Systems, Inc., Hugging Face Inc., Intel Corp. and Advanced Micro Devices, Inc. As one of the company’s most notable strategic updates, llm-d is powered by a native Kubernetes architecture and designed to unlock production at scale for AI inferencing, according to Wright.

“llm-d takes the inference engine of the [virtual large language model], focused essentially on a single server type environment, and distributes it across your infrastructure,” he explained. “These are really two critical projects that are bringing the models that we’ve built into production environments and then scaling that efficiently.”

Here’s the complete video interview, part of SiliconANGLE’s and theCUBE’s coverage of Red Hat Summit:

(* Disclosure: Red Hat Inc. sponsored this segment of theCUBE. Neither Red Hat nor other sponsors have editorial control over content on theCUBE or SiliconANGLE.)

Photo: SiliconANGLE

A message from John Furrier, co-founder of SiliconANGLE:

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About SiliconANGLE Media
SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.