Chieh-Yen Lin
2024
I Need Help! Evaluating LLM’s Ability to Ask for Users’ Support: A Case Study on Text-to-SQL Generation
Cheng-Kuang Wu
|
Zhi Rui Tam
|
Chao-Chung Wu
|
Chieh-Yen Lin
|
Hung-yi Lee
|
Yun-Nung Chen
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
This study explores the proactive ability of LLMs to seek user support. We propose metrics to evaluate the trade-off between performance improvements and user burden, and investigate whether LLMs can determine when to request help under varying information availability. Our experiments show that without external feedback, many LLMs struggle to recognize their need for user support. The findings highlight the importance of external signals and provide insights for future research on improving support-seeking strategies. Source code: https://github.com/appier-research/i-need-help
Let Me Speak Freely? A Study On The Impact Of Format Restrictions On Large Language Model Performance.
Zhi Rui Tam
|
Cheng-Kuang Wu
|
Yi-Lin Tsai
|
Chieh-Yen Lin
|
Hung-yi Lee
|
Yun-Nung Chen
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track
Structured generation, the process of producing content in standardized formats like JSON and XML, is widely utilized in real-world applications to extract key output information from large language models (LLMs).This study investigates whether such constraints on generation space impact LLMs’ abilities, including reasoning and domain knowledge comprehension. Specifically, we evaluate LLMs’ performance when restricted to adhere to structured formats versus generating free-form responses across various common tasks. Surprisingly, we observe a significant decline in LLMs’ reasoning abilities under format restrictions. Furthermore, we find that stricter format constraints generally lead to greater performance degradation in reasoning tasks.
Search
Co-authors
- Cheng-Kuang Wu 2
- Zhi Rui Tam 2
- Hung-Yi Lee 2
- Yun-Nung Chen 2
- Chao-Chung Wu 1
- show all...