%0 Conference Proceedings %T DUT-NLP at MEDIQA 2019: An Adversarial Multi-Task Network to Jointly Model Recognizing Question Entailment and Question Answering %A Zhou, Huiwei %A Li, Xuefei %A Yao, Weihong %A Lang, Chengkun %A Ning, Shixian %Y Demner-Fushman, Dina %Y Cohen, Kevin Bretonnel %Y Ananiadou, Sophia %Y Tsujii, Junichi %S Proceedings of the 18th BioNLP Workshop and Shared Task %D 2019 %8 August %I Association for Computational Linguistics %C Florence, Italy %F zhou-etal-2019-dut %X In this paper, we propose a novel model called Adversarial Multi-Task Network (AMTN) for jointly modeling Recognizing Question Entailment (RQE) and medical Question Answering (QA) tasks. AMTN utilizes a pre-trained BioBERT model and an Interactive Transformer to learn the shared semantic representations across different task through parameter sharing mechanism. Meanwhile, an adversarial training strategy is introduced to separate the private features of each task from the shared representations. Experiments on BioNLP 2019 RQE and QA Shared Task datasets show that our model benefits from the shared representations of both tasks provided by multi-task learning and adversarial training, and obtains significant improvements upon the single-task models. %R 10.18653/v1/W19-5046 %U https://aclanthology.org/W19-5046 %U https://doi.org/10.18653/v1/W19-5046 %P 437-445