Amita Sharma
RHOAI Engineering Manager
Sessions
As a cornerstone of Production AI, Kubeflow continues to redefine how enterprises orchestrate machine learning workflows. However, the project's true strength lies in its ecosystem. In this lightning talk, we’ll share Red Hat’s vision for Kubeflow community engagement in 2026. Join Amita Sharma (Kubeflow Trainer lead) and the team as we discuss our roadmap for fostering contributor growth, enhancing upstream collaboration, and what you can expect from the Kubeflow presence at this year's community booth. Whether you're a seasoned contributor or an ML enthusiast, learn how you can help shape the future of open-source AI.
GPUs are the backbone of modern AI and cloud workloads. But in reality, many GPUs sit idle most of the time. Even in well-run data centers, a large part of GPU capacity goes unused, which increases costs and slows teams down.
In this talk, we’ll break down why GPU utilization is so low and what you can do about it.
We’ll start with the basics, how GPUs are used today and where things go wrong. You’ll learn about common problems like uneven workloads, inefficient scheduling, limited visibility into GPU usage, and mismatches between hardware and software.
Next, we’ll walk through practical solutions. This includes GPU sharing, right-sizing workloads, better scheduling, and using the right monitoring tools. The focus will be on approaches you can actually apply in real systems.
We’ll also share real-world lessons from building a GPU-as-a-Service (GPUaaS) platform, covering features like model checkpointing, job preemption and resume, and queue-based scheduling with open-source tools such as Kueue to improve GPU efficiency.
By the end of the session, you’ll have a clear understanding of how to use GPUs more efficiently in AI, ML, and cloud environments, without unnecessary complexity.