Boring Problems Are Sometimes the Most Interesting

Richard Sproat


Abstract
In a recent position paper, Turing Award Winners Yoshua Bengio, Geoffrey Hinton, and Yann LeCun make the case that symbolic methods are not needed in AI and that, while there are still many issues to be resolved, AI will be solved using purely neural methods. In this piece I issue a challenge: Demonstrate that a purely neural approach to the problem of text normalization is possible. Various groups have tried, but so far nobody has eliminated the problem of unrecoverable errors, errors where, due to insufficient training data or faulty generalization, the system substitutes some other reading for the correct one. Solutions have been proposed that involve a marriage of traditional finite-state methods with neural models, but thus far nobody has shown that the problem can be solved using neural methods alone. Though text normalization is hardly an “exciting” problem, I argue that until one can solve “boring” problems like that using purely AI methods, one cannot claim that AI is a success.
Anthology ID:
2022.cl-2.8
Volume:
Computational Linguistics, Volume 48, Issue 2 - June 2022
Month:
June
Year:
2022
Address:
Cambridge, MA
Venue:
CL
SIG:
Publisher:
MIT Press
Note:
Pages:
483–490
Language:
URL:
https://aclanthology.org/2022.cl-2.8
DOI:
10.1162/coli_a_00439
Bibkey:
Cite (ACL):
Richard Sproat. 2022. Boring Problems Are Sometimes the Most Interesting. Computational Linguistics, 48(2):483–490.
Cite (Informal):
Boring Problems Are Sometimes the Most Interesting (Sproat, CL 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.cl-2.8.pdf