@InProceedings{meyers-EtAl:2018:Short,
  author    = {Meyers, Benjamin S  and  Munaiah, Nuthan  and  Prud'hommeaux, Emily  and  Meneely, Andrew  and  Wolff, Josephine  and  Ovesdotter Alm, Cecilia  and  Murukannaiah, Pradeep},
  title     = {A dataset for identifying actionable feedback in collaborative software development},
  booktitle = {Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},
  month     = {July},
  year      = {2018},
  address   = {Melbourne, Australia},
  publisher = {Association for Computational Linguistics},
  pages     = {126--131},
  abstract  = {Software developers and testers have long struggled with how to elicit proactive responses from their coworkers when reviewing code for security vulnerabilities and errors. For a code review to be successful, it must not only identify potential problems but also elicit an active response from the colleague responsible for modifying the code. To understand the factors that contribute to this outcome, we analyze a novel dataset of more than one million code reviews for the Google Chromium project, from which we extract linguistic features of feedback that elicited responsive actions from coworkers. Using a manually-labeled subset of reviewer comments, we trained a highly accurate classifier to identify acted-upon comments (AUC = 0.85). Our results demonstrate the utility of our dataset, the feasibility of using NLP for this new task, and the potential of NLP to improve our understanding of how communications between colleagues can be authored to elicit positive, proactive responses.},
  url       = {http://www.aclweb.org/anthology/P18-2021}
}

