Contrastive Data and Learning for Natural Language Processing

Rui Zhang, Yangfeng Ji, Yue Zhang, Rebecca J. Passonneau


Abstract
Current NLP models heavily rely on effective representation learning algorithms. Contrastive learning is one such technique to learn an embedding space such that similar data sample pairs have close representations while dissimilar samples stay far apart from each other. It can be used in supervised or unsupervised settings using different loss functions to produce task-specific or general-purpose representations. While it has originally enabled the success for vision tasks, recent years have seen a growing number of publications in contrastive NLP. This first line of works not only delivers promising performance improvements in various NLP tasks, but also provides desired characteristics such as task-agnostic sentence representation, faithful text generation, data-efficient learning in zero-shot and few-shot settings, interpretability and explainability. In this tutorial, we aim to provide a gentle introduction to the fundamentals of contrastive learning approaches and the theory behind them. We then survey the benefits and the best practices of contrastive learning for various downstream NLP applications including Text Classification, Question Answering, Summarization, Text Generation, Interpretability and Explainability, Commonsense Knowledge and Reasoning, Vision-and-Language.This tutorial intends to help researchers in the NLP and computational linguistics community to understand this emerging topic and promote future research directions of using contrastive learning for NLP applications.
Anthology ID:
2022.naacl-tutorials.6
Volume:
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Tutorial Abstracts
Month:
July
Year:
2022
Address:
Seattle, United States
Editors:
Miguel Ballesteros, Yulia Tsvetkov, Cecilia O. Alm
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
39–47
Language:
URL:
https://aclanthology.org/2022.naacl-tutorials.6
DOI:
10.18653/v1/2022.naacl-tutorials.6
Bibkey:
Cite (ACL):
Rui Zhang, Yangfeng Ji, Yue Zhang, and Rebecca J. Passonneau. 2022. Contrastive Data and Learning for Natural Language Processing. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Tutorial Abstracts, pages 39–47, Seattle, United States. Association for Computational Linguistics.
Cite (Informal):
Contrastive Data and Learning for Natural Language Processing (Zhang et al., NAACL 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.naacl-tutorials.6.pdf