Depending on the amount of data to process, file generation may take longer.

If it takes too long to generate, you can limit the data by, for example, reducing the range of years.

Article

Download BibTeX

Title

ACoRA – A Platform for Automating Code Review Tasks

Authors

[ 1 ] Instytut Informatyki, Wydział Informatyki i Telekomunikacji, Politechnika Poznańska | [ P ] employee

Scientific discipline (Law 2.0)

[2.3] Information and communication technology

Year of publication

2025

Published in

E-Informatica Software Engineering Journal

Journal year: 2025 | Journal volume: vol. 19 | Journal number: no. 1

Article type

scientific article

Publication language

english

Keywords
EN
  • Code Reviews
  • Continous Integration
  • BERT
  • Machine Learning
Abstract

EN Background: Modern Code Reviews (MCR) are frequently adopted when assuring code and design quality in continuous integration and deployment projects. Although tiresome, they serve a secondary purpose of learning about the software product. Aim: Our objective is to design and evaluate a support tool to help software developers focus on the most important code fragments to review and provide them with suggestions on what should be reviewed in this code. Method: We used design science research to develop and evaluate a tool for automating code reviews by providing recommendations for code reviewers. The tool is based on Transformer-based machine learning models for natural language processing, applied to both programming language code (patch content) and the review comments. We evaluate both the ability of the language model to match similar lines and the ability to correctly indicate the nature of the potential problems encoded in a set of categories. We evaluated the tool on two open-source projects and one industry project. Results: The proposed tool was able to correctly annotate (only true positives) 35%–41% and partially correctly annotate 76%–84% of code fragments to be reviewed with labels corresponding to different aspects of code the reviewer should focus on. Conclusion: By comparing our study to similar solutions, we conclude that indicating lines to be reviewed and suggesting the nature of the potential problems in the code allows us to achieve higher accuracy than suggesting entire changes in the code considered in other studies. Also, we have found that the differences depend more on the consistency of commenting rather than on the ability of the model to find similar lines.

Date of online publication

03.10.2024

Pages (from - to)

250102-1 - 250102-36

DOI

10.37190/e-Inf250102

URL

https://www.e-informatyka.pl/EISEJ/papers/2025/1/2/

Comments

Article Number: 250102

License type

CC BY (attribution alone)

Open Access Mode

open journal

Open Access Text Version

final published version

Date of Open Access to the publication

at the time of publication

Ministry points / journal

40

Impact Factor

1,2 [List 2023]

This website uses cookies to remember the authenticated session of the user. For more information, read about Cookies and Privacy Policy.