Jinghao Yuan
2023
HIT-MI&T Lab’s Submission to Eval4NLP 2023 Shared Task
Rui Zhang
|
Fuhai Song
|
Hui Huang
|
Jinghao Yuan
|
Muyun Yang
|
Tiejun Zhao
Proceedings of the 4th Workshop on Evaluation and Comparison of NLP Systems
Recently, Large Language Models (LLMs) have boosted the research in natural language processing and shown impressive capabilities across numerous domains, including machine translation evaluation. This paper presents our methods developed for the machine translation evaluation sub-task of the Eval4NLP 2023 Shared Task. Based on the provided LLMs, we propose a generation-based method as well as a probability-based method to perform evaluation, explore different strategies when selecting the demonstrations for in-context learning, and try different ensemble methods to further improve the evaluation accuracy. The experiment results on the development set and test set demonstrate the effectiveness of our proposed method.
Search