Thank you for joining this interview! The purpose of this interview is to understand how reviewing bots impact code review behaviors. This page shows interview questions with our recent results. We would like to discuss two themes as below:
Q1-1: *What kinds of jobs in the Jenkins were used by Nova project?*
- pep8, tox-python, docs, functional job
- Tempest_dsvm_cells
- Tempest_dsvm_full_devstack_plugin_ceph
- Tempest−dsvm−neutron−multinode−ful
- Tempest−dsvm−multinode−live−migration
- Grenade_dsvm_neutron_multinode
- Tempest−dsvm−ipa−wholedisk−bios−agent
Q1-2: *Which jobs are necessary or optional to verify source code?*
-> We would like to know what jobs are needed to study
Q2: *How have the jobs been used in code reviews over time?*
- Coverage of a job equals how much of reviews run the job per a hundred reviews in order.
- Tox-python job (2nd plot) has reached almost 100% in early period, then has decreased to 85% over time.
Q3: *Are the Jenkins's jobs expected to detect minor issues?*
- Typo
- Documentation rule
- Code convention etc.
Q4: *How do those jobs influence code reviews?*
- Quality (content, quality etc.)?
- Cost (time, #revision etc.)?
- Motivation (participation etc.)?
Q1: *What kinds of features show those perceived benefits?*
- Quality -> bug information?
- Cost -> review time or discussion length?
- Motivation -> The number of review participants?
Q2: *What are the challenges or your expectations of reviewing bots of the future?*
Toshiki Hirao
Nara Institute of Science and Technology, Nara, Japan
Shane McIntosh
McGill University, Québec, Canada