Nicholas Roy


2024

pdf bib
PRompt Optimization in Multi-Step Tasks (PROMST): Integrating Human Feedback and Heuristic-based Sampling
Yongchao Chen | Jacob Arkin | Yilun Hao | Yang Zhang | Nicholas Roy | Chuchu Fan
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing

Prompt optimization aims to find the best prompt to a large language model (LLM) for a given task. LLMs have been successfully used to help find and improve prompt candidates for single-step tasks. However, realistic tasks for agents are multi-step and introduce new challenges: (1) Prompt content is likely to be more extensive and complex, making it more difficult for LLMs to analyze errors, (2) the impact of an individual step is difficult to evaluate, and (3) different people may have varied preferences about task execution. While humans struggle to optimize prompts, they are good at providing feedback about LLM outputs; we therefore introduce a new LLM-driven discrete prompt optimization framework PROMST that incorporates human-designed feedback rules to automatically offer direct suggestions for improvement. We also use an extra learned heuristic model that predicts prompt performance to efficiently sample from prompt candidates. This approach significantly outperforms both human-engineered prompts and several other prompt optimization methods across 11 representative multi-step tasks (an average 10.6%-29.3% improvement to current best methods on five LLMs respectively). We believe our work can serve as a benchmark for automatic prompt optimization for LLM-driven multi-step tasks.

2019

pdf bib
Leveraging Past References for Robust Language Grounding
Subhro Roy | Michael Noseworthy | Rohan Paul | Daehyung Park | Nicholas Roy
Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)

Grounding referring expressions to objects in an environment has traditionally been considered a one-off, ahistorical task. However, in realistic applications of grounding, multiple users will repeatedly refer to the same set of objects. As a result, past referring expressions for objects can provide strong signals for grounding subsequent referring expressions. We therefore reframe the grounding problem from the perspective of coreference detection and propose a neural network that detects when two expressions are referring to the same object. The network combines information from vision and past referring expressions to resolve which object is being referred to. Our experiments show that detecting referring expression coreference is an effective way to ground objects described by subtle visual properties, which standard visual grounding models have difficulty capturing. We also show the ability to detect object coreference allows the grounding model to perform well even when it encounters object categories not seen in the training data.

2013

pdf bib
Probabilistic Dialogue Modeling for Speech-Enabled Assistive Technology
William Li | Jim Glass | Nicholas Roy | Seth Teller
Proceedings of the Fourth Workshop on Speech and Language Processing for Assistive Technologies

2012

pdf bib
Toward Learning Perceptually Grounded Word Meanings from Unaligned Parallel Data
Stefanie Tellex | Pratiksha Thaker | Josh Joseph | Nicholas Roy
Proceedings of the Second Workshop on Semantic Interpretation in an Actionable Context

2000

pdf bib
Spoken Dialogue Management Using Probabilistic Reasoning
Nicholas Roy | Joelle Pineau | Sebastian Thrun
Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics