Developers often run tests to check that their latest changes to a code repository did not break any previously working functionality. Ideally, any new test failures would indicate regressions caused by the latest changes. However, some test failures may not be due to the latest changes but due to non-determinism in the tests, popularly called flaky tests. The typical way to detect flaky tests is to rerun failing tests repeatedly. Unfortunately, rerunning failing tests can be costly and can slow down the development cycle.
We present the first extensive evaluation of rerunning failing tests and propose a new technique, called DeFlaker, that detects if a test failure is due to a flaky test without rerunning and with very low runtime overhead. DeFlaker monitors the coverage of latest code changes and marks as flaky any newly failing test that did not execute any of the changes. We deployed DeFlaker live, in the build process of 96 Java projects on TravisCI, and found 87 previously unknown flaky tests in 10 of these projects. We also ran experiments on project histories, where DeFlaker detected 1,874 flaky tests from 4,846 failures, with a low false alarm rate (1.5%). DeFlaker had a higher recall (95.5% vs. 23%) of confirmed flaky tests than Maven’s default flaky test detector.
Thu 31 May Times are displayed in time zone: (GMT+02:00) Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
|11:00 - 11:20|
Jonathan BellGeorge Mason University, Owolabi LegunsenUniversity of Illinois at Urbana-Champaign, Michael HiltonCarnegie Mellon University, USA, Lamyaa Eloussi, Tifany Yung, Darko MarinovUniversity of Illinois at Urbana-ChampaignPre-print Media Attached
|11:20 - 11:40|
|11:40 - 12:00|
Michaël MarcozziImperial College London, Sebastien Bardin, Nikolai Kosmatov, Mike PapadakisUniversity of Luxembourg, Virgile Prevosto, Loïc CorrensonLink to publication DOI File Attached
|12:00 - 12:20|
|12:20 - 12:30|