A + B: A General Generator-Reader Framework for Optimizing LLMs to Unleash Synergy Potential

Wei Tang, Yixin Cao, Jiahao Ying, Bo Wang, Yuyue Zhao, Yong Liao, Peng Zhou


Abstract
Retrieval-Augmented Generation (RAG) is an effective solution to supplement necessary knowledge to large language models (LLMs). Targeting its bottleneck of retriever performance, “generate-then-read” pipeline is proposed to replace the retrieval stage with generation from the LLM itself. Although promising, this research direction is underexplored and still cannot work in the scenario when source knowledge is given. In this paper, we formalize a general “A + B” framework with varying combinations of foundation models and types for systematic investigation. We explore the efficacy of the base and chat versions of LLMs and found their different functionalities suitable for generator A and reader B, respectively. Their combinations consistently outperform single models, especially in complex scenarios. Furthermore, we extend the application of the “A + B” framework to scenarios involving source documents through continuous learning, enabling the direct integration of external knowledge into LLMs. This approach not only facilitates effective acquisition of new knowledge but also addresses the challenges of safety and helpfulness post-adaptation. The paper underscores the versatility of the “A + B” framework, demonstrating its potential to enhance the practical application of LLMs across various domains.
Anthology ID:
2024.findings-acl.219
Volume:
Findings of the Association for Computational Linguistics ACL 2024
Month:
August
Year:
2024
Address:
Bangkok, Thailand and virtual meeting
Editors:
Lun-Wei Ku, Andre Martins, Vivek Srikumar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3670–3685
Language:
URL:
https://aclanthology.org/2024.findings-acl.219
DOI:
Bibkey:
Cite (ACL):
Wei Tang, Yixin Cao, Jiahao Ying, Bo Wang, Yuyue Zhao, Yong Liao, and Peng Zhou. 2024. A + B: A General Generator-Reader Framework for Optimizing LLMs to Unleash Synergy Potential. In Findings of the Association for Computational Linguistics ACL 2024, pages 3670–3685, Bangkok, Thailand and virtual meeting. Association for Computational Linguistics.
Cite (Informal):
A + B: A General Generator-Reader Framework for Optimizing LLMs to Unleash Synergy Potential (Tang et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-acl.219.pdf