Songhao Piao
2019
Inspecting Unification of Encoding and Matching with Transformer: A Case Study of Machine Reading Comprehension
Hangbo Bao
|
Li Dong
|
Furu Wei
|
Wenhui Wang
|
Nan Yang
|
Lei Cui
|
Songhao Piao
|
Ming Zhou
Proceedings of the 2nd Workshop on Machine Reading for Question Answering
Most machine reading comprehension (MRC) models separately handle encoding and matching with different network architectures. In contrast, pretrained language models with Transformer layers, such as GPT (Radford et al., 2018) and BERT (Devlin et al., 2018), have achieved competitive performance on MRC. A research question that naturally arises is: apart from the benefits of pre-training, how many performance gain comes from the unified network architecture. In this work, we evaluate and analyze unifying encoding and matching components with Transformer for the MRC task. Experimental results on SQuAD show that the unified model outperforms previous networks that separately treat encoding and matching. We also introduce a metric to inspect whether a Transformer layer tends to perform encoding or matching. The analysis results show that the unified model learns different modeling strategies compared with previous manually-designed models.
Search
Co-authors
- Hangbo Bao 1
- Li Dong 1
- Furu Wei 1
- Wenhui Wang 1
- Nan Yang 1
- show all...
Venues
- ws1