Enhancing LLM Capabilities Beyond Scaling Up

Wenpeng Yin, Muhao Chen, Rui Zhang, Ben Zhou, Fei Wang, Dan Roth


Abstract
General-purpose large language models (LLMs) are progressively expanding both in scale and access to unpublic training data. This has led to notable progress in a variety of AI problems. Nevertheless, two questions exist: i) Is scaling up the sole avenue of extending the capabilities of LLMs? ii) Instead of developing general-purpose LLMs, how to endow LLMs with specific knowledge? This tutorial targets researchers and practitioners who are interested in capability extension of LLMs that go beyond scaling up. To this end, we will discuss several lines of research that follow that direction, including (i) the adaptation of LLMs to assimilate new information in situations where conflicts arise, (ii) the adaptation of LLMs to address target problems with inherent constraints, (iii) the customization of LLMs to align with user-specific instructions and preference, (iv) the defense against potential attacks and threads by malicious users, and (v) the collaboration with external models directly or through APIs. At last, we will conclude the tutorial by outlining directions for further investigation.
Anthology ID:
2024.emnlp-tutorials.1
Volume:
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1–10
Language:
URL:
https://aclanthology.org/2024.emnlp-tutorials.1
DOI:
Bibkey:
Cite (ACL):
Wenpeng Yin, Muhao Chen, Rui Zhang, Ben Zhou, Fei Wang, and Dan Roth. 2024. Enhancing LLM Capabilities Beyond Scaling Up. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts, pages 1–10, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Enhancing LLM Capabilities Beyond Scaling Up (Yin et al., EMNLP 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.emnlp-tutorials.1.pdf