@inproceedings{baker-etal-2021-spoon,
title = "The Spoon Is in the Sink: Assisting Visually Impaired People in the Kitchen",
author = "Baker, Katie and
Parekh, Amit and
Fabre, Adrien and
Addlesee, Angus and
Kruiper, Ruben and
Lemon, Oliver",
editor = "Howes, Christine and
Dobnik, Simon and
Breitholtz, Ellen and
Chatzikyriakidis, Stergios",
booktitle = "Proceedings of the Reasoning and Interaction Conference (ReInAct 2021)",
month = oct,
year = "2021",
address = "Gothenburg, Sweden",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.reinact-1.5",
pages = "32--39",
abstract = "Visual Question Answering (VQA) systems are increasingly adept at a variety of tasks, and this technology can be used to assist blind and partially sighted people. To do this, the system{'}s responses must not only be accurate, but usable. It is also vital for assistive technologies to be designed with a focus on: (1) privacy, as the camera may capture a user{'}s mail, medication bottles, or other sensitive information; (2) transparency, so that the system{'}s behaviour can be explained and trusted by users; and (3) controllability, to tailor the system for a particular domain or user group. We have therefore extended a conversational VQA framework, called Aye-saac, with these objectives in mind. Specifically, we gave Aye-saac the ability to answer visual questions in the kitchen, a particularly challenging area for visually impaired people. Our system can now answer questions about quantity, positioning, and system confidence in regards to 299 kitchen objects. Questions about the spatial relations between these objects are particularly helpful to visually impaired people, and our system output more usable answers than other state of the art end-to-end VQA systems.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="baker-etal-2021-spoon">
<titleInfo>
<title>The Spoon Is in the Sink: Assisting Visually Impaired People in the Kitchen</title>
</titleInfo>
<name type="personal">
<namePart type="given">Katie</namePart>
<namePart type="family">Baker</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Amit</namePart>
<namePart type="family">Parekh</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Adrien</namePart>
<namePart type="family">Fabre</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Angus</namePart>
<namePart type="family">Addlesee</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ruben</namePart>
<namePart type="family">Kruiper</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Oliver</namePart>
<namePart type="family">Lemon</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-10</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Reasoning and Interaction Conference (ReInAct 2021)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Christine</namePart>
<namePart type="family">Howes</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Simon</namePart>
<namePart type="family">Dobnik</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ellen</namePart>
<namePart type="family">Breitholtz</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Stergios</namePart>
<namePart type="family">Chatzikyriakidis</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Gothenburg, Sweden</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Visual Question Answering (VQA) systems are increasingly adept at a variety of tasks, and this technology can be used to assist blind and partially sighted people. To do this, the system’s responses must not only be accurate, but usable. It is also vital for assistive technologies to be designed with a focus on: (1) privacy, as the camera may capture a user’s mail, medication bottles, or other sensitive information; (2) transparency, so that the system’s behaviour can be explained and trusted by users; and (3) controllability, to tailor the system for a particular domain or user group. We have therefore extended a conversational VQA framework, called Aye-saac, with these objectives in mind. Specifically, we gave Aye-saac the ability to answer visual questions in the kitchen, a particularly challenging area for visually impaired people. Our system can now answer questions about quantity, positioning, and system confidence in regards to 299 kitchen objects. Questions about the spatial relations between these objects are particularly helpful to visually impaired people, and our system output more usable answers than other state of the art end-to-end VQA systems.</abstract>
<identifier type="citekey">baker-etal-2021-spoon</identifier>
<location>
<url>https://aclanthology.org/2021.reinact-1.5</url>
</location>
<part>
<date>2021-10</date>
<extent unit="page">
<start>32</start>
<end>39</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T The Spoon Is in the Sink: Assisting Visually Impaired People in the Kitchen
%A Baker, Katie
%A Parekh, Amit
%A Fabre, Adrien
%A Addlesee, Angus
%A Kruiper, Ruben
%A Lemon, Oliver
%Y Howes, Christine
%Y Dobnik, Simon
%Y Breitholtz, Ellen
%Y Chatzikyriakidis, Stergios
%S Proceedings of the Reasoning and Interaction Conference (ReInAct 2021)
%D 2021
%8 October
%I Association for Computational Linguistics
%C Gothenburg, Sweden
%F baker-etal-2021-spoon
%X Visual Question Answering (VQA) systems are increasingly adept at a variety of tasks, and this technology can be used to assist blind and partially sighted people. To do this, the system’s responses must not only be accurate, but usable. It is also vital for assistive technologies to be designed with a focus on: (1) privacy, as the camera may capture a user’s mail, medication bottles, or other sensitive information; (2) transparency, so that the system’s behaviour can be explained and trusted by users; and (3) controllability, to tailor the system for a particular domain or user group. We have therefore extended a conversational VQA framework, called Aye-saac, with these objectives in mind. Specifically, we gave Aye-saac the ability to answer visual questions in the kitchen, a particularly challenging area for visually impaired people. Our system can now answer questions about quantity, positioning, and system confidence in regards to 299 kitchen objects. Questions about the spatial relations between these objects are particularly helpful to visually impaired people, and our system output more usable answers than other state of the art end-to-end VQA systems.
%U https://aclanthology.org/2021.reinact-1.5
%P 32-39
Markdown (Informal)
[The Spoon Is in the Sink: Assisting Visually Impaired People in the Kitchen](https://aclanthology.org/2021.reinact-1.5) (Baker et al., ReInAct 2021)
ACL