* ICSE 2018 *
Sun 27 May - Sun 3 June 2018 Gothenburg, Sweden
Thu 31 May 2018 11:00 - 11:20 at Congress Hall - Test Improvement Chair(s): Yves Le Traon

Developers often run tests to check that their latest changes to a code repository did not break any previously working functionality. Ideally, any new test failures would indicate regressions caused by the latest changes. However, some test failures may not be due to the latest changes but due to non-determinism in the tests, popularly called flaky tests. The typical way to detect flaky tests is to rerun failing tests repeatedly. Unfortunately, rerunning failing tests can be costly and can slow down the development cycle.

We present the first extensive evaluation of rerunning failing tests and propose a new technique, called DeFlaker, that detects if a test failure is due to a flaky test without rerunning and with very low runtime overhead. DeFlaker monitors the coverage of latest code changes and marks as flaky any newly failing test that did not execute any of the changes. We deployed DeFlaker live, in the build process of 96 Java projects on TravisCI, and found 87 previously unknown flaky tests in 10 of these projects. We also ran experiments on project histories, where DeFlaker detected 1,874 flaky tests from 4,846 failures, with a low false alarm rate (1.5%). DeFlaker had a higher recall (95.5% vs. 23%) of confirmed flaky tests than Maven’s default flaky test detector.

Thu 31 May
Times are displayed in time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change

11:00 - 12:30: Test ImprovementTechnical Papers at Congress Hall
Chair(s): Yves Le TraonUniversity of Luxembourg
11:00 - 11:20
DeFlaker: Automatically Detecting Flaky Tests
Technical Papers
Jonathan BellGeorge Mason University, Owolabi LegunsenUniversity of Illinois at Urbana-Champaign, Michael HiltonCarnegie Mellon University, USA, Lamyaa Eloussi, Tifany Yung, Darko MarinovUniversity of Illinois at Urbana-Champaign
Pre-print Media Attached
11:20 - 11:40
DetReduce: Minimizing Android GUI Test Suites for Regression Testing
Technical Papers
Wontae Choi, Koushik SenUniversity of California, Berkeley, George NeculaUniversity of California, Berkeley, Wenyu WangUniversity of Illinois at Urbana-Champaign
11:40 - 12:00
Time to Clean your Test Objectives
Technical Papers
Link to publication DOI File Attached
12:00 - 12:20
Prioritizing Browser Environments for Web Application Test Execution
Technical Papers
Junghyun Kwon, In-Young KoKorea Advanced Institute of Science and Technology, Gregg RothermelUniversity of Nebraska - Lincoln
12:20 - 12:30
Q&A in groups
Technical Papers