Donghyun Lee
Also published as: DongHyun Lee
2023
Quantifying Information of Tokens for Simple and Flexible Simultaneous Machine Translation
DongHyun Lee
|
Minkyung Park
|
Byung-Jun Lee
Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL)
Simultaneous Translation (ST) involves translating with only partial source inputs instead of the entire source inputs, a process that can potentially result in translation quality degradation. Previous approaches to balancing translation quality and latency have demonstrated that it is more efficient and effective to leverage an offline model with a reasonable policy. However, using an offline model also leads to a distribution shift since it is not trained with partial source inputs, and it can be improved by training an additional module that informs us when to translate. In this paper, we propose an Information Quantifier (IQ) that models source and target information to determine whether the offline model has sufficient information for translation, trained with oracle action sequences generated from the offline model. IQ, by quantifying information, helps in formulating a suitable policy for Simultaneous Translation that better generalizes and also allows us to control the trade-off between quality and latency naturally. Experiments on various language pairs show that our proposed model outperforms baselines.
2019
VAE-PGN based Abstractive Model in Multi-stage Architecture for Text Summarization
Hyungtak Choi
|
Lohith Ravuru
|
Tomasz Dryjański
|
Sunghan Rye
|
Donghyun Lee
|
Hojung Lee
|
Inchul Hwang
Proceedings of the 12th International Conference on Natural Language Generation
This paper describes our submission to the TL;DR challenge. Neural abstractive summarization models have been successful in generating fluent and consistent summaries with advancements like the copy (Pointer-generator) and coverage mechanisms. However, these models suffer from their extractive nature as they learn to copy words from the source text. In this paper, we propose a novel abstractive model based on Variational Autoencoder (VAE) to address this issue. We also propose a Unified Summarization Framework for the generation of summaries. Our model eliminates non-critical information at a sentence-level with an extractive summarization module and generates the summary word by word using an abstractive summarization module. To implement our framework, we combine submodules with state-of-the-art techniques including Pointer-Generator Network (PGN) and BERT while also using our new VAE-PGN abstractive model. We evaluate our model on the benchmark Reddit corpus as part of the TL;DR challenge and show that our model outperforms the baseline in ROUGE score while generating diverse summaries.
Search
Co-authors
- Minkyung Park 1
- Byung-Jun Lee 1
- Hyungtak Choi 1
- Lohith Ravuru 1
- Tomasz Dryjański 1
- show all...