Nav: Home

Speech recognition technology is not a solution for poor readers

May 13, 2019

Even today about one in five humans is considered to be low literate or illiterate; they cannot read or write simple statements about everyday life. Low literacy can be due to no or little reading practice or reading impairments such as dyslexia. For developing countries with low literacy rates, voice recognition has been hailed as a solution by companies such as Google, calling it 'the next big leap in technology'. But is speech technology really the solution for low literacy?

Falk Huettig and Martin Pickering argue that it is not. In an opinion article in Trends in Cognitive Sciences, the psycholinguists suggest that relying on speech technology might be counterproductive, as literacy has crucial benefits even beyond reading. "It is very relevant and timely to look at the advantages of reading on speech, especially as people tend to read less and in different ways than they used to", says Falk Huettig. "Contemporary social media writing and reading habits for example are quite different from traditional print media. Information that people used to get from written sources such as novels, newspapers, public notices, or even recipe books they get more and more from YouTube videos, podcasts, or audiobooks".

This is not necessarily a bad thing, as some of the general benefits of reading can also be obtained from listening to audiobooks. As audio books also provide 'book language', listening to them will confer some similar advantages--such as a larger vocabulary, increased knowledge of the world and a larger short-term ('working') memory, which is important to keep track of information and multiple entities over several sentences, paragraphs, or often even pages.

But according to Huettig and Pickering, reading itself--the actual physical act of reading--is crucially important for developing the skill of predicting upcoming words, which transfers from reading to understanding spoken language. Reading trains the language prediction system although even very young children--who cannot read yet--can predict where a sentence is going. When 2-year-olds hear "the boy eats a big cake", they already look at something edible (i.e. a cake) after hearing "eats", but before hearing "cake". Predicting upcoming information is useful, as it reduces processing load and frees up limited brain resources. And crucially, skilled readers get much better at predicting.

Children who are among the most avid readers encounter over 4 million words a year, while children who rarely read encounter only about 50,000 words. As a result, good readers get a deeper understanding of the meaning of words and build large networks of words with strong associations between them--which helps them to predict upcoming words. As poor readers have smaller vocabularies and weaker representations of words in their mind (i.e. the recollection of the sound and meaning of a word), the predictive relationships between words are also weaker (e.g. the prediction that 'read the ... ' is often followed by 'book').

The literate mind

As reading is self-paced, there is a strong incentive to predict upcoming words, as this speeds up reading, which is typically much faster than listening. Skilled readers tend to take in whole words at one glance (gazing with their eyes at multiple letters at the same time) and time their eye movements to optimise the reading process. Printed texts (even given the occasional changes in fonts and word capitalizations) are much more regular than conversational speech, which is full of disfluencies, incomplete word pronunciations, and speech errors. This regularity of written texts helps readers to form the predictive relationships between words that then, by extension, can also be used to better predict words when listening to speech.

It is hard to imagine for someone who learned to read a long time ago but even the concept of a 'word' is an invention of the literate mind; it is very hard to grasp if you are an illiterate who only ever hears a stream of speech sounds. For example, when illiterates or children who haven't learned to read yet are asked to repeat the last word of a spoken sentence, they tend to repeat the whole sentence. In contrast, words clearly stand out in written language, typically being separated by white space. Written forms make words more salient and precise: readers become more aware that words are stable units in language. Storing the written form of words in memory also helps to make spoken word forms more salient, to be accessed faster when predicting upcoming speech. And, again, it is prediction of upcoming language that makes language understanding become really fast and proficient.

"Our arguments provide one more reason why more efforts should be undertaken to teach the hundreds of millions of illiterates in developing countries and functional illiterates across the world how to read (or to read better) and why a focus on artificial intelligence voice recognition and voice assistants to overcome literacy-related problems has its dangers", the authors argue.

"Writing is an ancient human technology that we shouldn't give up easily. Teaching how to read and how to read better remains very important even in a modern technological world", concludes Huettig.
-end-
Falk Huettig & Martin J. Pickering (2019). Literacy advantages beyond reading: Prediction of spoken language. Trends in Cognitive Sciences.

Questions? Contact:

Falk Huettig
Phone: +31 24 3521374
Email: falk.huettig@mpi.nl

Marjolein Scherphuis (press officer)
Phone: +31 24 3521947
Email: marjolein.scherphuis@mpi.nl

Max Planck Institute for Psycholinguistics

Related Language Articles:

Human language most likely evolved gradually
One of the most controversial hypotheses for the origin of human language faculty is the evolutionary conjecture that language arose instantaneously in humans through a single gene mutation.
'She' goes missing from presidential language
MIT researchers have found that although a significant percentage of the American public believed the winner of the November 2016 presidential election would be a woman, people rarely used the pronoun 'she' when referring to the next president before the election.
How does language emerge?
How did the almost 6000 languages of the world come into being?
New research quantifies how much speakers' first language affects learning a new language
Linguistic research suggests that accents are strongly shaped by the speaker's first language they learned growing up.
Why the language-ready brain is so complex
In a review article published in Science, Peter Hagoort, professor of Cognitive Neuroscience at Radboud University and director of the Max Planck Institute for Psycholinguistics, argues for a new model of language, involving the interaction of multiple brain networks.
Do as i say: Translating language into movement
Researchers at Carnegie Mellon University have developed a computer model that can translate text describing physical movements directly into simple computer-generated animations, a first step toward someday generating movies directly from scripts.
Learning language
When it comes to learning a language, the left side of the brain has traditionally been considered the hub of language processing.
Learning a second alphabet for a first language
A part of the brain that maps letters to sounds can acquire a second, visually distinct alphabet for the same language, according to a study of English speakers published in eNeuro.
Sign language reveals the hidden logical structure, and limitations, of spoken language
Sign languages can help reveal hidden aspects of the logical structure of spoken language, but they also highlight its limitations because speech lacks the rich iconic resources that sign language uses on top of its sophisticated grammar.
Lying in a foreign language is easier
It is not easy to tell when someone is lying.
More Language News and Language Current Events

Trending Science News

Current Coronavirus (COVID-19) News

Top Science Podcasts

We have hand picked the top science podcasts of 2020.
Now Playing: TED Radio Hour

Listen Again: Reinvention
Change is hard, but it's also an opportunity to discover and reimagine what you thought you knew. From our economy, to music, to even ourselves–this hour TED speakers explore the power of reinvention. Guests include OK Go lead singer Damian Kulash Jr., former college gymnastics coach Valorie Kondos Field, Stockton Mayor Michael Tubbs, and entrepreneur Nick Hanauer.
Now Playing: Science for the People

#562 Superbug to Bedside
By now we're all good and scared about antibiotic resistance, one of the many things coming to get us all. But there's good news, sort of. News antibiotics are coming out! How do they get tested? What does that kind of a trial look like and how does it happen? Host Bethany Brookeshire talks with Matt McCarthy, author of "Superbugs: The Race to Stop an Epidemic", about the ins and outs of testing a new antibiotic in the hospital.
Now Playing: Radiolab

Dispatch 6: Strange Times
Covid has disrupted the most basic routines of our days and nights. But in the middle of a conversation about how to fight the virus, we find a place impervious to the stalled plans and frenetic demands of the outside world. It's a very different kind of front line, where urgent work means moving slow, and time is marked out in tiny pre-planned steps. Then, on a walk through the woods, we consider how the tempo of our lives affects our minds and discover how the beats of biology shape our bodies. This episode was produced with help from Molly Webster and Tracie Hunte. Support Radiolab today at Radiolab.org/donate.