Investigating the Validity of Ground Truth in Code Reviewer Recommendation Studies

Autor: K. Ayberk Tecimer, Emre Doğan, H. Altay Güvenir, Eray Tüzün
Přispěvatelé: Doğan, Emre, Tüzün, Eray, Tecimer, Kazım Ayberk, Güvenir, Halil Altay
Rok vydání: 2019
Předmět:
Zdroj: ESEM
International Symposium on Empirical Software Engineering and Measurement
DOI: 10.1109/esem.2019.8870190
Popis: Date of Conference: 19-20 September 2019 Conference Name: 13th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement, ESEM 2019 Background: Selecting the ideal code reviewer in modern code review is a crucial first step to perform effective code reviews. There are several algorithms proposed in the literature for recommending the ideal code reviewer for a given pull request. The success of these code reviewer recommendation algorithms is measured by comparing the recommended reviewers with the ground truth that is the assigned reviewers selected in real life. However, in practice, the assigned reviewer may not be the ideal reviewer for a given pull request.Aims: In this study, we investigate the validity of ground truth data in code reviewer recommendation studies.Method: By conducting an informal literature review, we compared the reviewer selection heuristics in real life and the algorithms used in recommendation models. We further support our claims by using empirical data from code reviewer recommendation studies.Results: By literature review, and accompanying empirical data, we show that ground truth data used in code reviewer recommendation studies is potentially problematic. This reduces the validity of the code reviewer datasets and the reviewer recommendation studies. Conclusion: We demonstrated the cases where the ground truth in code reviewer recommendation studies are invalid and discussed the potential solutions to address this issue.
Databáze: OpenAIRE