DevConf.IN 2025

Shardul Inamdar

Shardul is an AI Engineer at Finfactor, developing AI-driven solutions for fintech. Previously, he worked at Red Hat, driving innovation within the internal ecosystem. His interests include LLMs, Java, DevOps, and Flutter. A passionate educator, he has taught over 500 students across various colleges. Beyond technology, Shardul is an advocate of mindfulness and meditation, emphasizing balance and clarity in both work and life.


Company or affiliation

Finfactor

Job title

AI Engineer


Sessions

02-28
10:40
35min
Demystifying Self-Hosted LLMs : A Beginner's Guide to Self-Hosting on Kubernetes with Podman Desktop
Praveen Kumar, Ramakrishna Yekulla, Shardul Inamdar

Deploying and managing large language models (LLMs) in production often presents significant challenges.By self-hosting LLMs on Kubernetes, organizations gain enhanced data privacy, flexibility in model training, and potential cost savings. This talk aims to enabling beginners by demystifying the process of self-hosting LLMs within the robust Kubernetes ecosystem.We will place a special emphasis on harnessing the capabilities of the Podman Desktop AI Lab extension to accelerate and simplify the development, deployment, and management of LLM workloads on Kubernetes.
Key topics will include:
- Strategically selecting and containerizing suitable open-source LLM models for optimal performance
- Crafting Kubernetes deployment manifests tailored for LLM workloads
- Provisioning and managing Kubernetes resources to meet the computational demands of LLMs
- Deep dive into leveraging the Podman Desktop AI Lab extension for streamlined LLM workflows on Kubernetes

AI, Data Science, and Emerging Tech
Raigad Room | Chanakya Building / School of Business (capacity 200)
02-28
13:00
80min
Making LLM fine-tuning accessible with InstructLab
Shardul Inamdar, Dasharath Masirkar, Presha Pathak, Ravindra Patil

The rise of large language models (LLMs) has opened up exciting possibilities for developers looking to build intelligent applications. However, the process of adapting these models to specific use cases can be difficult, requiring deep expertise and substantial resources. In this talk, we'll introduce you to InstructLab, an open-source project that aims to make LLM tuning accessible to developers and engineers of all skill levels, on consumer-grade hardware.

We'll explore how InstructLab's innovative approach combines collaborative knowledge curation, efficient data generation, and instruction training to enable developers to refine foundation models for specific use cases. In this workshop, you’ll be provided a RHEL VM and learn how to enhance an LLM with new knowledge and capabilities for targeted applications, without needing data science expertise. Join us to explore how LLM tuning can be more accessible and democratized, empowering developers to build on the power of AI in their projects.

Pre-requisites for Workshop.

  1. Have a laptop running Fedora, Mac or any linux variant.
    If running windows, you can run a Linux virtual machine inside Hypervisor
    [Please have any one hypervisor installed and linux system installed as Virtual Machine]

  2. Download the 3 models locally from GoogleDrive below :
    https://drive.google.com/drive/folders/1NZOk46RU4l4XrYkXI3KkXXk07PHaYtPm?usp=drive_link

We will not have wifi facilities at venue, hence please download these models beforehand.

AI, Data Science, and Emerging Tech
Workshops | School of Design Floor-7/8 (capacity 50)