<?xml version="1.0" encoding="UTF-8" ?>
<volume id="W17">
  <paper id="1600">
    <title>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</title>
    <editor>Dirk Hovy</editor>
    <editor>Shannon Spruit</editor>
    <editor>Margaret Mitchell</editor>
    <editor>Emily M. Bender</editor>
    <editor>Michael Strube</editor>
    <editor>Hanna Wallach</editor>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <url>http://www.aclweb.org/anthology/W17-16</url>
    <bibtype>book</bibtype>
    <bibkey>EthNLP:2017</bibkey>
  </paper>

  <paper id="1601">
    <title>Gender as a Variable in Natural-Language Processing: Ethical Considerations</title>
    <author><first>Brian</first><last>Larson</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>1&#8211;11</pages>
    <url>http://www.aclweb.org/anthology/W17-1601</url>
    <abstract>Researchers and practitioners in natural-language processing (NLP) and related
	fields should attend to ethical principles in study design, ascription of
	categories/variables to study participants, and reporting of findings or
	results. This paper discusses theoretical and ethical frameworks for using
	gender as a variable in NLP studies and proposes four guidelines for
	researchers and practitioners. The principles outlined here should guide
	practitioners, researchers, and peer reviewers, and they may be applicable to
	other social categories, such as race, applied to human beings connected to NLP
	research.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>larson:2017:EthNLP</bibkey>
  </paper>

  <paper id="1602">
    <title>These are not the Stereotypes You are Looking For: Bias and Fairness in Authorial Gender Attribution</title>
    <author><first>Corina</first><last>Koolen</last></author>
    <author><first>Andreas</first><last>van Cranenburgh</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>12&#8211;22</pages>
    <url>http://www.aclweb.org/anthology/W17-1602</url>
    <abstract>Stylometric and text categorization results show that author gender can be
	discerned in texts with relatively high accuracy. However, it is difficult to
	explain what gives rise to these results and there are many possible
	confounding factors, such as the domain, genre, and target audience of a text.
	More fundamentally, such classification efforts risk invoking stereotyping and
	essentialism. We explore this issue in two datasets of Dutch literary novels,
	using commonly used descriptive (LIWC, topic modeling) and predictive (machine
	learning) methods. Our results show the importance of controlling for variables
	in the corpus and we argue for taking care not to overgeneralize from the
	results.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>koolen-vancranenburgh:2017:EthNLP</bibkey>
  </paper>

  <paper id="1603">
    <title>A Quantitative Study of Data in the NLP community</title>
    <author><first>Margot</first><last>Mieskes</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>23&#8211;29</pages>
    <url>http://www.aclweb.org/anthology/W17-1603</url>
    <abstract>We present results on a quantitative analysis of publications in the NLP domain
	on collecting, publishing and availability of research data. We find that a
	wide range of publications rely on data crawled from the web, but few give
	details on how potentially sensitive data was treated. Additionally, we find
	that while links to repositories of data are given, they often do not work even
	a short time after publication. We put together several suggestions on how to
	improve this situation based on publications from the NLP domain, but also
	other research areas.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>mieskes:2017:EthNLP</bibkey>
  </paper>

  <paper id="1604">
    <title>Ethical by Design: Ethics Best Practices for Natural Language Processing</title>
    <author><first>Jochen L.</first><last>Leidner</last></author>
    <author><first>Vassilis</first><last>Plachouras</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>30&#8211;40</pages>
    <url>http://www.aclweb.org/anthology/W17-1604</url>
    <abstract>Natural language processing (NLP) systems analyze and/or generate human
	language, typically on users’ behalf. One natural and necessary question that
	needs to be addressed in this context, both in research projects and in
	production settings, is the question how ethical the work is, both regarding
	the process and its outcome.
	    Towards this end, we articulate a set of issues, propose a set of best
	practices, notably a process featuring an ethics review board, and sketch and
	how they could be meaningfully applied. Our main argument is that ethical
	outcomes ought to be achieved by design, i.e. by following a process aligned
	by ethical values. We also offer some response options for those facing ethics
	issues.
	    While a number of previous works exist that discuss ethical issues, in
	particular around big data and machine learning, to the authors’ knowledge
	this is the first account of NLP and ethics from the perspective of a
	principled process.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>leidner-plachouras:2017:EthNLP</bibkey>
  </paper>

  <paper id="1605">
    <title>Building Better Open-Source Tools to Support Fairness in Automated Scoring</title>
    <author><first>Nitin</first><last>Madnani</last></author>
    <author><first>Anastassia</first><last>Loukina</last></author>
    <author><first>Alina</first><last>von Davier</last></author>
    <author><first>Jill</first><last>Burstein</last></author>
    <author><first>Aoife</first><last>Cahill</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>41&#8211;52</pages>
    <url>http://www.aclweb.org/anthology/W17-1605</url>
    <abstract>Automated scoring of written and spoken responses is an NLP application that
	can significantly impact lives especially when deployed as part of high-stakes
	tests such as the GREėxtregistered~ and the TOEFLėxtregistered~. Ethical considerations require that
	automated scoring algorithms treat all test- takers fairly. The educational
	measurement community has done significant research on fairness in assessments
	and automated scoring systems must incorporate their recommendations. The best
	way to do that is by making available automated, non-proprietary tools to NLP
	researchers that directly incorporate these recommendations and generate the
	analyses needed to help identify and resolve biases in their scoring systems.
	In this paper, we attempt to provide such a solution.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>madnani-EtAl:2017:EthNLP</bibkey>
  </paper>

  <paper id="1606">
    <title>Gender and Dialect Bias in YouTube's Automatic Captions</title>
    <author><first>Rachael</first><last>Tatman</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>53&#8211;59</pages>
    <url>http://www.aclweb.org/anthology/W17-1606</url>
    <abstract>This project evaluates the accuracy of YouTube's automatically-generated
	captions across two genders and five dialect groups. Speakers' dialect and
	gender was controlled for by using videos uploaded as part of the &#x201c;accent tag
	challenge", where speakers explicitly identify their language background. The
	results show robust differences in accuracy across both gender and dialect,
	with lower accuracy for 1) women and 2) speakers from Scotland. This finding
	builds on earlier research finding that speaker's sociolinguistic identity may
	negatively impact their ability to use automatic speech recognition, and
	demonstrates the need for sociolinguistically-stratified validation of systems.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>tatman:2017:EthNLP</bibkey>
  </paper>

  <paper id="1607">
    <title>Integrating the Management of Personal Data Protection and Open Science with Research Ethics</title>
    <author><first>Dave</first><last>Lewis</last></author>
    <author><first>Joss</first><last>Moorkens</last></author>
    <author><first>Kaniz</first><last>Fatema</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>60&#8211;65</pages>
    <url>http://www.aclweb.org/anthology/W17-1607</url>
    <abstract>We examine the impact of the EU General
	Data Protection Regulation and the push
	from research funders to provide open access
	research data on the current practices
	in Language Technology Research.
	We analyse the challenges that arise and
	the opportunities to address many of them
	through the use of existing open data practices.
	We discuss the impact of this also on
	current practice in research ethics.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>lewis-moorkens-fatema:2017:EthNLP</bibkey>
  </paper>

  <paper id="1608">
    <title>Ethical Considerations in NLP Shared Tasks</title>
    <author><first>Carla</first><last>Parra Escart&#237;n</last></author>
    <author><first>Wessel</first><last>Reijers</last></author>
    <author><first>Teresa</first><last>Lynn</last></author>
    <author><first>Joss</first><last>Moorkens</last></author>
    <author><first>Andy</first><last>Way</last></author>
    <author><first>Chao-Hong</first><last>Liu</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>66&#8211;73</pages>
    <url>http://www.aclweb.org/anthology/W17-1608</url>
    <abstract>Shared tasks are increasingly common in our field, and new challenges are
	suggested at almost every conference and  workshop. However, as this has become
	an established way of pushing research forward, it is important to discuss how
	we researchers organise and participate in shared tasks, and make that
	information available to the  community to allow further research improvements.
	In this paper, we present a number of ethical issues along with other areas of
	concern that are related to the competitive nature of shared tasks. As such
	issues could potentially impact on research ethics in the Natural Language
	Processing community, we also propose the development of a framework for the
	organisation of and participation in shared tasks that can help mitigate
	against these issues arising.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>parraescartin-EtAl:2017:EthNLP</bibkey>
  </paper>

  <paper id="1609">
    <title>Social Bias in Elicited Natural Language Inferences</title>
    <author><first>Rachel</first><last>Rudinger</last></author>
    <author><first>Chandler</first><last>May</last></author>
    <author><first>Benjamin</first><last>Van Durme</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>74&#8211;79</pages>
    <url>http://www.aclweb.org/anthology/W17-1609</url>
    <abstract>We analyze the Stanford Natural Language Inference (SNLI) corpus in an
	investigation of bias and stereotyping in NLP data. The SNLI human-elicitation
	protocol makes it prone to amplifying bias and stereotypical associations,
	which we demonstrate statistically (using pointwise mutual information) and
	with qualitative examples.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>rudinger-may-vandurme:2017:EthNLP</bibkey>
  </paper>

  <paper id="1610">
    <title>A Short Review of Ethical Challenges in Clinical Natural Language Processing</title>
    <author><first>Simon</first><last>Suster</last></author>
    <author><first>Stephan</first><last>Tulkens</last></author>
    <author><first>Walter</first><last>Daelemans</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>80&#8211;87</pages>
    <url>http://www.aclweb.org/anthology/W17-1610</url>
    <abstract>Clinical NLP has an immense potential in contributing to how clinical practice
	will be revolutionized by the advent of large scale processing of clinical
	records. However, this potential has remained largely untapped due to slow
	progress primarily caused by strict data access policies for researchers. In
	this paper, we discuss the concern for privacy and the measures it entails. We
	also suggest sources of less sensitive data. Finally, we draw attention to
	biases that can compromise the validity of empirical research and lead to
	socially harmful applications.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>suster-tulkens-daelemans:2017:EthNLP</bibkey>
  </paper>

  <paper id="1611">
    <title>Goal-Oriented Design for Ethical Machine Learning and NLP</title>
    <author><first>Tyler</first><last>Schnoebelen</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>88&#8211;93</pages>
    <url>http://www.aclweb.org/anthology/W17-1611</url>
    <abstract>The argument made in this paper is that to act ethically in machine learning
	and NLP requires focusing on goals. NLP projects are often classificatory
	systems that deal with human subjects, which means that goals from people
	affected by the systems should be included. The paper takes as its core example
	a model that detects criminality, showing the problems of training data,
	categories, and outcomes. The paper is oriented to the kinds of critiques on
	power and the reproduction of inequality that are found in social theory, but
	it also includes concrete suggestions on how to put goal-oriented design into
	practice.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>schnoebelen:2017:EthNLP</bibkey>
  </paper>

  <paper id="1612">
    <title>Ethical Research Protocols for Social Media Health Research</title>
    <author><first>Adrian</first><last>Benton</last></author>
    <author><first>Glen</first><last>Coppersmith</last></author>
    <author><first>Mark</first><last>Dredze</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>94&#8211;102</pages>
    <url>http://www.aclweb.org/anthology/W17-1612</url>
    <abstract>Social media have transformed data-driven research in political science, the
	social sciences, health, and medicine. Since health research often touches on
	sensitive topics that relate to ethics of treatment and patient privacy,
	similar ethical considerations should be acknowledged when using social media
	data in health research.  While much has been said regarding the ethical
	considerations of social media research, health research leads to an additional
	set of concerns.  We provide practical suggestions in the form of guidelines
	for researchers working with social media data in health research.  These
	guidelines can inform an IRB proposal for researchers new to social media
	health research.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>benton-coppersmith-dredze:2017:EthNLP</bibkey>
  </paper>

  <paper id="1613">
    <title>Say the Right Thing Right: Ethics Issues in Natural Language Generation Systems</title>
    <author><first>Charese</first><last>Smiley</last></author>
    <author><first>Frank</first><last>Schilder</last></author>
    <author><first>Vassilis</first><last>Plachouras</last></author>
    <author><first>Jochen L.</first><last>Leidner</last></author>
    <booktitle>Proceedings of the First ACL Workshop on Ethics in Natural Language Processing</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>103&#8211;108</pages>
    <url>http://www.aclweb.org/anthology/W17-1613</url>
    <abstract>We discuss the ethical implications of Natural Language Generation systems.
	We use one particular system as a case study to identify and classify issues,
	and we provide an ethics checklist, in the hope that future system designers
	may benefit from conducting their own ethics reviews based on our checklist.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>smiley-EtAl:2017:EthNLP</bibkey>
  </paper>

</volume>

