DevConf.CZ 2025

Daniel Hiller

Daniel Hiller is a software engineer with more than 20 years of work experience. He strives to create software that is useful for people so they can do stuff that matters.
He’s currently part of the KubeVirt community. KubeVirt is a Kubernetes add-on to run virtual machines on Kubernetes clusters. There he maintains, improves and automates CI and testing processes for the KubeVirt organization using Prow on Kubernetes and Golang.


Company or affiliation

Red Hat

Job title

Senior Software Engineer


Session

06-14
12:30
35min
Zap the Flakes! Leveraging AI to Combat Flaky Tests with CANNIER
Daniel Hiller

Flakes aka tests that don’t behave deterministically, i.e. they fail sometimes and pass sometimes, are an ever recurring problem in software development. This is especially the sad reality when running end-to-end tests where a lot of components are involved. There are various reasons why a test can be flaky, however the impact can be as fatal as CI being loaded beyond capacity causing overly long feedback cycles or even users losing trust in CI itself.

Ideally CI maintainers want potential flakes to be flagged at the earliest stage of the development lifecycle, so that they do not even enter the end-to-end test suite. They want a gate that acts as a safeguard for developers, pointing out to them what kind of stability issues a test has. This reduces CI user frustration and improves trust in CI. At the same time it cuts down on unnecessary waste of CI system resources.

This talk will explore the CANNIER approach, which aims to reduce the time cost of rerunning tests. The speaker and the audience will take a closer look at the feature set used to create training data from existing test code of the KubeVirt project, enabling CI systems to predict probable flakiness of a certain test with an AI model.

The speaker will cover how a subset of the CANNIER feature set is implemented, how it is used to generate training data for an AI model, and how the model is used to analyze flakiness probability, generating actionable feedback for test authors.

Attendees will gain insight on how the CANNIER approach can help them improve test reliability, ultimately enhancing overall software quality and team productivity.

Artificial Intelligence and Data Science
D105 (capacity 300)