Abhanshu Sharma
2024
Chart-based Reasoning: Transferring Capabilities from LLMs to VLMs
Victor Carbune
|
Hassan Mansoor
|
Fangyu Liu
|
Rahul Aralikatte
|
Gilles Baechler
|
Jindong Chen
|
Abhanshu Sharma
Findings of the Association for Computational Linguistics: NAACL 2024
Vision-language models (VLMs) are achieving increasingly strong performance on multimodal tasks. However, reasoning capabilities remain limited particularly for smaller VLMs, while those of large-language models (LLMs) have seen numerous improvements. We pro-pose a technique to transfer capabilities from LLMs to VLMs. On the recently introduced ChartQA, our method obtains state-of-the-artperformance when applied on the PaLI3-5B VLM by Chen et al. (2023c), while also enabling much better performance on PlotQA and FigureQA.We first improve the chart representation by continuing the pre-training stage using an improved version of the chart-to-table translation task by Liu et al. (2023a). We then propose constructing a 20x larger dataset than the original training set. To improve general reasoning capabilities and improve numerical operations, we synthesize reasoning traces using the table representation of charts. Lastly, our model is fine-tuned using the multitask loss introduced by Hsieh et al. (2023).Our variant ChartPaLI-5B outperforms even 10x larger models such as PaLIX-55B without using an upstream OCR system, while keeping inference time constant compared to the PaLI3-5B baseline. When rationales are further refined with a simple program-of-thought prompt (Chen et al., 2023a), our model outperforms the recently introduced Gemini Ultra and GPT-4V.
2022
Towards Better Semantic Understanding of Mobile Interfaces
Srinivas Sunkara
|
Maria Wang
|
Lijuan Liu
|
Gilles Baechler
|
Yu-Chung Hsiao
|
Jindong Chen
|
Abhanshu Sharma
|
James W. W. Stout
Proceedings of the 29th International Conference on Computational Linguistics
Improving the accessibility and automation capabilities of mobile devices can have a significant positive impact on the daily lives of countless users. To stimulate research in this direction, we release a human-annotated dataset with approximately 500k unique annotations aimed at increasing the understanding of the functionality of UI elements. This dataset augments images and view hierarchies from RICO, a large dataset of mobile UIs, with annotations for icons based on their shapes and semantics, and associations between different elements and their corresponding text labels, resulting in a significant increase in the number of UI elements and the categories assigned to them. We also release models using image-only and multimodal inputs; we experiment with various architectures and study the benefits of using multimodal inputs on the new dataset. Our models demonstrate strong performance on an evaluation set of unseen apps, indicating their generalizability to newer screens. These models, combined with the new dataset, can enable innovative functionalities like referring to UI elements by their labels, improved coverage and better semantics for icons etc., which would go a long way in making UIs more usable for everyone.
Search
Co-authors
- Gilles Baechler 2
- Jindong Chen 2
- Victor Cărbune 1
- Hassan Mansoor 1
- Fangyu Liu 1
- show all...