Is the ranking of PubMed similar articles good enough? An evaluation of text similarity methods for three datasets

Mariana Neves, Ines Schadock, Beryl Eusemann, Gilbert Schnfelder, Bettina Bert, Daniel Butzke


Abstract
The use of seed articles in information retrieval provides many advantages, such as a longercontext and more details about the topic being searched for. Given a seed article (i.e., a PMID), PubMed provides a pre-compiled list of similar articles to support the user in finding equivalent papers in the biomedical literature. We aimed at performing a quantitative evaluation of the PubMed Similar Articles based on three existing biomedical text similarity datasets, namely, RELISH, TREC-COVID, and SMAFIRA-c. Further, we carried out a survey and an evaluation of various text similarity methods on these three datasets. Our experiments considered the original title and abstract from PubMed as well as automatically detected sections and manually annotated relevant sentences. We provide an overview about which methods better performfor each dataset and compare them to the ranking in PubMed similar articles. While resultsvaried considerably among the datasets, we were able to obtain a better performance thanPubMed for all of them. Datasets and source codes are available at: https://github.com/mariananeves/reranking
Anthology ID:
2023.bionlp-1.11
Volume:
The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Dina Demner-fushman, Sophia Ananiadou, Kevin Cohen
Venue:
BioNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
133–144
Language:
URL:
https://aclanthology.org/2023.bionlp-1.11
DOI:
10.18653/v1/2023.bionlp-1.11
Bibkey:
Cite (ACL):
Mariana Neves, Ines Schadock, Beryl Eusemann, Gilbert Schnfelder, Bettina Bert, and Daniel Butzke. 2023. Is the ranking of PubMed similar articles good enough? An evaluation of text similarity methods for three datasets. In The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks, pages 133–144, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Is the ranking of PubMed similar articles good enough? An evaluation of text similarity methods for three datasets (Neves et al., BioNLP 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.bionlp-1.11.pdf
Video:
 https://aclanthology.org/2023.bionlp-1.11.mp4