Thanh Nguyen


2021

pdf bib
FaBULOUS: Fact-checking Based on Understanding of Language Over Unstructured and Structured information
Mostafa Bouziane | Hugo Perrin | Amine Sadeq | Thanh Nguyen | Aurélien Cluzeau | Julien Mardas
Proceedings of the Fourth Workshop on Fact Extraction and VERification (FEVER)

As part of the FEVEROUS shared task, we developed a robust and finely tuned architecture to handle the joint retrieval and entailment on text data as well as structured data like tables. We proposed two training schemes to tackle the hurdles inherent to multi-hop multi-modal datasets. The first one allows having a robust retrieval of full evidence sets, while the second one enables entailment to take full advantage of noisy evidence inputs. In addition, our work has revealed important insights and potential avenue of research for future improvement on this kind of dataset. In preliminary evaluation on the FEVEROUS shared task test set, our system achieves 0.271 FEVEROUS score, with 0.4258 evidence recall and 0.5607 entailment accuracy.

2020

pdf bib
Learning Robust Models for e-Commerce Product Search
Thanh Nguyen | Nikhil Rao | Karthik Subbian
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics

Showing items that do not match search query intent degrades customer experience in e-commerce. These mismatches result from counterfactual biases of the ranking algorithms toward noisy behavioral signals such as clicks and purchases in the search logs. Mitigating the problem requires a large labeled dataset, which is expensive and time-consuming to obtain. In this paper, we develop a deep, end-to-end model that learns to effectively classify mismatches and to generate hard mismatched examples to improve the classifier. We train the model end-to-end by introducing a latent variable into the cross-entropy loss that alternates between using the real and generated samples. This not only makes the classifier more robust but also boosts the overall ranking performance. Our model achieves a relative gain compared to baselines by over 26% in F-score, and over 17% in Area Under PR curve. On live search traffic, our model gains significant improvement in multiple countries.