Ramakrishna Yekulla
Red Hat
Job title –Principal Architect
Sessions
In this workshop, application developers will learn how to ideate, prototype, build, and refine AI applications directly within their local development environment. Using Podman AI Lab, participants will explore how to run and connect AI models seamlessly, enabling hands-on experimentation. Topics include selecting the right large language model (LLM), crafting and testing prompts, working with custom data, and benchmarking performance. Additionally, we’ll dive into Instruct Lab to demonstrate how it can be used to fine-tune LLMs for specific use cases. This session is ideal for developers eager to harness the power of AI in their applications.
Deploying and managing large language models (LLMs) in production often presents significant challenges.By self-hosting LLMs on Kubernetes, organizations gain enhanced data privacy, flexibility in model training, and potential cost savings. This talk aims to enabling beginners by demystifying the process of self-hosting LLMs within the robust Kubernetes ecosystem.We will place a special emphasis on harnessing the capabilities of the Podman Desktop AI Lab extension to accelerate and simplify the development, deployment, and management of LLM workloads on Kubernetes.
Key topics will include:
- Strategically selecting and containerizing suitable open-source LLM models for optimal performance
- Crafting Kubernetes deployment manifests tailored for LLM workloads
- Provisioning and managing Kubernetes resources to meet the computational demands of LLMs
- Deep dive into leveraging the Podman Desktop AI Lab extension for streamlined LLM workflows on Kubernetes