2025
pdf
bib
abs
Deploying Multi-task Online Server with Large Language Model
Yincen Qu
|
Hengyue Liu
|
Kun Wang
|
Xiangying Dai
|
Xiaoou Lu
|
Hui Zhou
|
Chao Ma
Proceedings of the 31st International Conference on Computational Linguistics: Industry Track
In the industry, numerous tasks are deployed online. Traditional approaches often tackle each task separately by its own network, which leads to excessive costs for developing and scaling models, especially in the context of large language models. Although multi-task methods can save costs through parameter sharing, they often struggle to outperform single-task methods in real-world applications. To tackle these challenges, we present a three-stage multi-task learning framework for large language models. It involves task filtering, followed by fine-tuning on high-resource tasks, and finally fine-tuning on all tasks. We conducted comprehensive experiments in single-task and multi-task settings. Our approach, exemplified on different benchmarks, demonstrates that it is able to achieve performance comparable to the single-task method while reducing up to 90.9% of its overhead.
2018
pdf
bib
abs
Joint Neural Entity Disambiguation with Output Space Search
Hamed Shahbazi
|
Xiaoli Fern
|
Reza Ghaeini
|
Chao Ma
|
Rasha Mohammad Obeidat
|
Prasad Tadepalli
Proceedings of the 27th International Conference on Computational Linguistics
In this paper, we present a novel model for entity disambiguation that combines both local contextual information and global evidences through Limited Discrepancy Search (LDS). Given an input document, we start from a complete solution constructed by a local model and conduct a search in the space of possible corrections to improve the local solution from a global view point. Our search utilizes a heuristic function to focus more on the least confident local decisions and a pruning function to score the global solutions based on their local fitness and the global coherences among the predicted entities. Experimental results on CoNLL 2003 and TAC 2010 benchmarks verify the effectiveness of our model.
2016
pdf
bib
Improving Users’ Demographic Prediction via the Videos They Talk about
Yuan Wang
|
Yang Xiao
|
Chao Ma
|
Zhen Xiao
Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing
2014
pdf
bib
Prune-and-Score: Learning for Greedy Coreference Resolution
Chao Ma
|
Janardhan Rao Doppa
|
J. Walker Orr
|
Prashanth Mannem
|
Xiaoli Fern
|
Tom Dietterich
|
Prasad Tadepalli
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)