Jooyoung Choi
2022
ANNA: Enhanced Language Representation for Question Answering
Changwook Jun
|
Hansol Jang
|
Myoseop Sim
|
Hyun Kim
|
Jooyoung Choi
|
Kyungkoo Min
|
Kyunghoon Bae
Proceedings of the 7th Workshop on Representation Learning for NLP
Pre-trained language models have brought significant improvements in performance in a variety of natural language processing tasks. Most existing models performing state-of-the-art results have shown their approaches in the separate perspectives of data processing, pre-training tasks, neural network modeling, or fine-tuning. In this paper, we demonstrate how the approaches affect performance individually, and that the language model performs the best results on a specific question answering task when those approaches are jointly considered in pre-training models. In particular, we propose an extended pre-training task, and a new neighbor-aware mechanism that attends neighboring tokens more to capture the richness of context for pre-training language modeling. Our best model achieves new state-of-the-art results of 95.7% F1 and 90.6% EM on SQuAD 1.1 and also outperforms existing pre-trained language models such as RoBERTa, ALBERT, ELECTRA, and XLNet on the SQuAD 2.0 benchmark.
Korean-Specific Dataset for Table Question Answering
Changwook Jun
|
Jooyoung Choi
|
Myoseop Sim
|
Hyun Kim
|
Hansol Jang
|
Kyungkoo Min
Proceedings of the Thirteenth Language Resources and Evaluation Conference
Existing question answering systems mainly focus on dealing with text data. However, much of the data produced daily is stored in the form of tables that can be found in documents and relational databases, or on the web. To solve the task of question answering over tables, there exist many datasets for table question answering written in English, but few Korean datasets. In this paper, we demonstrate how we construct Korean-specific datasets for table question answering: Korean tabular dataset is a collection of 1.4M tables with corresponding descriptions for unsupervised pre-training language models. Korean table question answering corpus consists of 70k pairs of questions and answers created by crowd-sourced workers. Subsequently, we then build a pre-trained language model based on Transformer and fine-tune the model for table question answering with these datasets. We then report the evaluation results of our model. We make our datasets publicly available via our GitHub repository and hope that those datasets will help further studies for question answering over tables, and for the transformation of table formats.
Search
Co-authors
- Changwook Jun 2
- Hansol Jang 2
- Myoseop Sim 2
- Hyun Kim 2
- Kyungkoo Min 2
- show all...