@article{sproat-2022-boring,
title = "Boring Problems Are Sometimes the Most Interesting",
author = "Sproat, Richard",
journal = "Computational Linguistics",
volume = "48",
number = "2",
month = jun,
year = "2022",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/2022.cl-2.8/",
doi = "10.1162/coli_a_00439",
pages = "483--490",
abstract = "In a recent position paper, Turing Award Winners Yoshua Bengio, Geoffrey Hinton, and Yann LeCun make the case that symbolic methods are not needed in AI and that, while there are still many issues to be resolved, AI will be solved using purely neural methods. In this piece I issue a challenge: Demonstrate that a purely neural approach to the problem of text normalization is possible. Various groups have tried, but so far nobody has eliminated the problem of unrecoverable errors, errors where, due to insufficient training data or faulty generalization, the system substitutes some other reading for the correct one. Solutions have been proposed that involve a marriage of traditional finite-state methods with neural models, but thus far nobody has shown that the problem can be solved using neural methods alone. Though text normalization is hardly an {\textquotedblleft}exciting{\textquotedblright} problem, I argue that until one can solve {\textquotedblleft}boring{\textquotedblright} problems like that using purely AI methods, one cannot claim that AI is a success."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="sproat-2022-boring">
<titleInfo>
<title>Boring Problems Are Sometimes the Most Interesting</title>
</titleInfo>
<name type="personal">
<namePart type="given">Richard</namePart>
<namePart type="family">Sproat</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2022-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<genre authority="bibutilsgt">journal article</genre>
<relatedItem type="host">
<titleInfo>
<title>Computational Linguistics</title>
</titleInfo>
<originInfo>
<issuance>continuing</issuance>
<publisher>MIT Press</publisher>
<place>
<placeTerm type="text">Cambridge, MA</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">periodical</genre>
<genre authority="bibutilsgt">academic journal</genre>
</relatedItem>
<abstract>In a recent position paper, Turing Award Winners Yoshua Bengio, Geoffrey Hinton, and Yann LeCun make the case that symbolic methods are not needed in AI and that, while there are still many issues to be resolved, AI will be solved using purely neural methods. In this piece I issue a challenge: Demonstrate that a purely neural approach to the problem of text normalization is possible. Various groups have tried, but so far nobody has eliminated the problem of unrecoverable errors, errors where, due to insufficient training data or faulty generalization, the system substitutes some other reading for the correct one. Solutions have been proposed that involve a marriage of traditional finite-state methods with neural models, but thus far nobody has shown that the problem can be solved using neural methods alone. Though text normalization is hardly an “exciting” problem, I argue that until one can solve “boring” problems like that using purely AI methods, one cannot claim that AI is a success.</abstract>
<identifier type="citekey">sproat-2022-boring</identifier>
<identifier type="doi">10.1162/coli_a_00439</identifier>
<location>
<url>https://aclanthology.org/2022.cl-2.8/</url>
</location>
<part>
<date>2022-06</date>
<detail type="volume"><number>48</number></detail>
<detail type="issue"><number>2</number></detail>
<extent unit="page">
<start>483</start>
<end>490</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Journal Article
%T Boring Problems Are Sometimes the Most Interesting
%A Sproat, Richard
%J Computational Linguistics
%D 2022
%8 June
%V 48
%N 2
%I MIT Press
%C Cambridge, MA
%F sproat-2022-boring
%X In a recent position paper, Turing Award Winners Yoshua Bengio, Geoffrey Hinton, and Yann LeCun make the case that symbolic methods are not needed in AI and that, while there are still many issues to be resolved, AI will be solved using purely neural methods. In this piece I issue a challenge: Demonstrate that a purely neural approach to the problem of text normalization is possible. Various groups have tried, but so far nobody has eliminated the problem of unrecoverable errors, errors where, due to insufficient training data or faulty generalization, the system substitutes some other reading for the correct one. Solutions have been proposed that involve a marriage of traditional finite-state methods with neural models, but thus far nobody has shown that the problem can be solved using neural methods alone. Though text normalization is hardly an “exciting” problem, I argue that until one can solve “boring” problems like that using purely AI methods, one cannot claim that AI is a success.
%R 10.1162/coli_a_00439
%U https://aclanthology.org/2022.cl-2.8/
%U https://doi.org/10.1162/coli_a_00439
%P 483-490
Markdown (Informal)
[Boring Problems Are Sometimes the Most Interesting](https://aclanthology.org/2022.cl-2.8/) (Sproat, CL 2022)
ACL