A computational approach to understanding how infants perceive language

January 29, 2021

Languages differ in the sounds they use. The Japanese language, for example, does not distinguish between "r" and "l" sounds as in "rock" versus "lock." Remarkably, infants become attuned to the sounds of their native language before they learn to speak. One-year-old babies, for example, less readily distinguish between "rock" and "lock" when living in an environment where Japanese, rather than English, is spoken.

Influential scientific accounts of this early phonetic learning phenomenon initially proposed that infants group sounds into native vowel- and consonant-like phonetic categories through a statistical clustering mechanism known as "distributional learning."

The idea that infants learn consonant- and vowel-like phonetic categories has been challenged, however, by a new study published this week in the Proceedings of the National Academy of Sciences.

In the study, a multi-institutional team of cognitive scientists and computational linguists have introduced a quantitative modeling framework that is based on a large-scale simulation of the language learning process in infants. Using computationally efficient machine learning techniques, this approach allows learning mechanisms to be systematically linked to testable predictions regarding infants' attunement to their native language.

"Hypotheses about what is being learned by infants have traditionally driven researchers' attempts to understand this surprising phenomenon," says Thomas Schatz, a postdoctoral associate in the University of Maryland of Maryland Institute for Advanced Computer Studies (UMIACS) who was lead author of the study. "We propose to start from hypotheses about how infants might learn."

In addition to Schatz, the study's authors include Naomi Feldman, an associate professor of linguistics at the University of Maryland with an appointment in UMIACS; Sharon Goldwater, a professor in the Institute for Language, Cognition and Computation at the University of Edinburgh's School of Informatics; Xuân-Nga Cao, a research engineer at Ecole Normale Supérieure (ENS) in Paris and co-founder of the Langinnov and Gazouyi startups; and Emmanuel Dupoux,a professor who directs the Cognitive Machine Learning team at ENS.

For their study, the researchers simulated the learning process in infants by training a computationally efficient clustering algorithm on realistic speech input. The algorithm was fed spectrogram-like auditory features sampled at regular time intervals that were obtained from naturalistic speech recordings in a target language. In this study, American English and Japanese were the two languages used.

This yielded a candidate model for the early phonetic knowledge of, say, a Japanese infant, the researchers say. Next, they asked two questions of the trained models. Could they explain the observed differences in how Japanese- and English-learning infants discriminate speech sounds? And, did the models learn vowel- and consonant-like phonetic categories?

The dominant scientific accounts of early phonetic learning would have expected the answers to these questions to match (either both should be "yes" or both should be "no"). The researchers found that the answer to the first question was positive: Their models did account for infants' observed behavior, in particular for the Japanese infants' difficulty with distinguishing words like "rock" and "lock." The answer to the second question, however, was negative: The models were found to have learned speech units too brief and acoustically variable to correspond to vowel- and consonant-like phonetic categories.

These results suggest a striking reinterpretation of the existing literature on early phonetic learning. Difficulties in scaling up distributional learning of phonetic categories to realistic learning conditions may be better interpreted as questioning the idea that what infants learn are phonetic categories, rather than the idea that how infants learn is through pure distributional learning (the traditional interpretation).

Cognitive science has not traditionally made use of such large-scale modeling, says Schatz, but recent advances in computing power, large datasets, and machine-learning algorithms make this approach more feasible than ever before.

Schatz and Feldman are part of the Computational Linguistics and Information Procession (CLIP) Laboratory in UMIACS, where Feldman is the current director. The robust computing resources in the CLIP lab and the Cognitive Machine Learning lab in Paris were instrumental to the research project, Feldman says.

In conclusion, the researchers believe their computationally-based modeling approach--together with ongoing efforts in the field to collect empirical data on a large scale, such as large-scale recordings of infants' learning environments at home and large-scale assessment of infants' learning outcomes--opens the path toward a much deeper understanding of early language acquisition.
The research paper, "Early phonetic learning without phonetic categories - Insights from large-scale simulations on realistic input," was published on January 28, 2021, in the Proceedings of the National Academy of Sciences.

University of Maryland

Related Language Articles from Brightsurf:

Learning the language of sugars
We're told not to eat too much sugar, but in reality, all of our cells are covered in sugar molecules called glycans.

How effective are language learning apps?
Researchers from Michigan State University recently conducted a study focusing on Babbel, a popular subscription-based language learning app and e-learning platform, to see if it really worked at teaching a new language.

Chinese to rise as a global language
With the continuing rise of China as a global economic and trading power, there is no barrier to prevent Chinese from becoming a global language like English, according to Flinders University academic Dr Jeffrey Gil.

'She' goes missing from presidential language
MIT researchers have found that although a significant percentage of the American public believed the winner of the November 2016 presidential election would be a woman, people rarely used the pronoun 'she' when referring to the next president before the election.

How does language emerge?
How did the almost 6000 languages of the world come into being?

New research quantifies how much speakers' first language affects learning a new language
Linguistic research suggests that accents are strongly shaped by the speaker's first language they learned growing up.

Why the language-ready brain is so complex
In a review article published in Science, Peter Hagoort, professor of Cognitive Neuroscience at Radboud University and director of the Max Planck Institute for Psycholinguistics, argues for a new model of language, involving the interaction of multiple brain networks.

Do as i say: Translating language into movement
Researchers at Carnegie Mellon University have developed a computer model that can translate text describing physical movements directly into simple computer-generated animations, a first step toward someday generating movies directly from scripts.

Learning language
When it comes to learning a language, the left side of the brain has traditionally been considered the hub of language processing.

Learning a second alphabet for a first language
A part of the brain that maps letters to sounds can acquire a second, visually distinct alphabet for the same language, according to a study of English speakers published in eNeuro.

Read More: Language News and Language Current Events
Brightsurf.com is a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for sites to earn advertising fees by advertising and linking to Amazon.com.