Yun Wang
2024
Neural Search Space in Gboard Decoder
Yanxiang Zhang
|
Yuanbo Zhang
|
Haicheng Sun
|
Yun Wang
|
Gary Sivek
|
Shumin Zhai
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track
Gboard Decoder produces suggestions by looking for paths that best match input touch points on the context aware search space, which is backed by the language Finite State Transducers (FST). The language FST is currently an N-gram language model (LM). However, N-gram LMs, limited in context length, are known to have sparsity problem under device model size constraint. In this paper, we propose Neural Search Space which substitutes the N-gram LM with a Neural Network LM (NN-LM) and dynamically constructs the search space during decoding. Specifically, we integrate the long range context awareness of NN-LM into the search space by converting its outputs given context, into the language FST at runtime. This involves language FST structure redesign, pruning strategies tuning, and data structure optimizations. Online experiments demonstrate improved quality results, reducing Words Modified Ratio by [0.26%, 1.19%] on various locales with acceptable latency increases. This work opens new avenues for further improving keyboard decoding quality by enhancing neural LM more directly.
2023
LLM4Vis: Explainable Visualization Recommendation using ChatGPT
Lei Wang
|
Songheng Zhang
|
Yun Wang
|
Ee-Peng Lim
|
Yong Wang
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track
Data visualization is a powerful tool for exploring and communicating insights in various domains. To automate visualization choice for datasets, a task known as visualization recommendation has been proposed. Various machine-learning-based approaches have been developed for this purpose, but they often require a large corpus of dataset-visualization pairs for training and lack natural explanations for their results. To address this research gap, we propose LLM4Vis, a novel ChatGPT-based prompting approach to perform visualization recommendation and return human-like explanations using very few demonstration examples. Our approach involves feature description, demonstration example selection, explanation generation, demonstration example construction, and inference steps. To obtain demonstration examples with high-quality explanations, we propose a new explanation generation bootstrapping to iteratively refine generated explanations by considering the previous generation and template-based hint. Evaluations on the VizML dataset show that LLM4Vis outperforms or performs similarly to supervised learning models like Random Forest, Decision Tree, and MLP, in both few-shot and zero-shot settings. The qualitative evaluation also shows the effectiveness of explanations generated by LLM4Vis.
Search
Co-authors
- Yanxiang Zhang 1
- Yuanbo Zhang 1
- Haicheng Sun 1
- Gary Sivek 1
- Shumin Zhai 1
- show all...