Similar brain glitch found in slips of signing, speaking

May 04, 2020

When we speak, we give little thought to how the words form in our brain before we say them. It's similar for deaf people using sign language.

Speaking and signing come naturally, except when we stumble over words, or swap one word for another when we speak or sign too quickly, are tired or preoccupied.

Fluency and the occasional disfluency both happen because of how we choose what to say or sign, when a neural mechanism takes place in our brains as we make decisions and monitor how we communicate.

It's this mechanism that fascinates San Diego State University researchers Stephanie Ries and Karen Emmorey in the School of Speech, Language and Hearing Sciences. Their analysis could help inform rehabilitation therapy for those relearning how to speak or sign after a stroke.

Using electroencephalogram (EEG) recordings, they studied how hearing and deaf signers process the act of signing and found the same monitoring mechanism took place in the brains of both groups. Among deaf signers, it was more prevalent with those for whom American Sign Language (ASL) is their first language.

"When we are doing an action, whether it's speaking, signing, pressing buttons or typing, we see the same mechanism," Ries said. "Any time we are making a decision to do something, this neural mechanism comes into play."

Their study, published by MIT Press in the Journal of Cognitive Neuroscience on April 30, may advance our understanding of how deaf individuals recover their ability to sign after a traumatic brain injury or stroke, when they suffer aphasia: the inability to understand others or express themselves due to brain damage.

"When stroke victims are more aware of their speech errors and have a better functioning speech monitoring mechanism, they have a better chance of recovering than those who don't have that awareness," Ries said. "This study helped us extend that understanding to signing ability for deaf people."

Melding speech with sign language expertise

The work also represents a long-held dream to combine the skills and training of two researchers with niche expertise in complementary fields - speech monitoring and sign monitoring.

Ries is an assistant professor specializing in the neuroscience of speech and language disorders who first met Emmorey at a workshop on language production in 2007 when Ries was a Ph.D. student in Marseille. Emmorey, a distinguished professor, sign language expert and director of the Laboratory for Language and Cognitive Neuroscience at SDSU, presented a study about sign monitoring which sparked an abiding interest in Ries, who wanted to work with Emmorey. When they crossed paths at another conference five years ago, Emmorey urged her to apply for the assistant professorship at SDSU, and they eventually began working together.

"I've always been interested in what inner signing would be like, and if it's similar to inner speech," said Emmorey, the study's senior author. "It's an internal process. When you speak, you can hear yourself. But if you're signing, are you seeing yourself like in a mirror, or is it a mental image of you signing, or a motor representation so you can feel how you sign?"

These were the underlying aspects of signing no one quite understood, and it has long been Emmorey's goal to tease them apart so we truly understand what sign language processing is like. Knowing this will help sign language educators figure out the best learning strategy for signers, much like the techniques used to teach hearing people foreign languages.

Since Ries was already working on speech monitoring with hearing people in France, when she joined SDSU, the two researchers combined their expertise to study sign monitoring in hearing and deaf people.

Monitoring for self-editing

They used the EEG data recorded with 21 hearing signers and 26 deaf signers in the Neurocognition Lab of Philip Holcomb and Katherine Midgley, colleagues in the psychology department. The participants were shown pictures to identify by signing, while wearing an EEG cap with 32-channel tin electrodes to monitor the mechanism behind signing.

"We wanted to study sign monitoring in-depth to understand the underlying mechanism and whether it's universal," Ries said. "Before people start to sign, you see this component rising, and we observed it happen with hearing signers as well, except it wasn't as clear."

This difference was possibly because deaf signers were more proficient in ASL than hearing signers. It's important to note that both deaf and hearing signers are bilingual in English and ASL, except ASL is more dominant for deaf signers.

"When we're speaking we catch ourselves when we are about to make an error. That's thanks to this monitoring process which is located in the medial frontal cortex of the brain," Ries said. "It peaks 40 milliseconds after you begin speaking, so it's extremely fast. We make an error because we may not have selected the right word when semantically related words are competing in your brain."

Words that share similar meanings such as 'oven' and 'fridge' or names may be switched in the brain (e.g., swapping your children's names). Other times, syllables get transposed.

Such errors can happen in signing too, when signs for different words are mixed up or an incorrect handshape is swapped for the desired handshape, which indicates signers are actually assembling phonological units during language production, similar to assembling the phonemes in a spoken word.

"Learning how sign production is represented in the brain will help us understand sign language disorders, and if a signer needs epileptic surgery we will know which part of the brain processes sign," Emmorey said.
The study's co-authors include Linda Nadalet and Soren Mickelson, who were master's students in speech language pathology, and Megan Mott, who was a master's student in psychology.

Funding came from a grant from the SDSU Center for Cognitive and Clinical Neuroscience, designed to encourage interdisciplinary collaborations across campus. Emmorey and Ries are also funded by grants from the National Institute for Deafness and other Communication Disorders within the National Institutes for Health.

San Diego State University

Related Stroke Articles from Brightsurf:

Stroke alarm clock may streamline and accelerate time-sensitive acute stroke care
An interactive, digital alarm clock may speed emergency stroke care, starting at hospital arrival and through each step of the time-sensitive treatment process.

Stroke patients with COVID-19 have increased inflammation, stroke severity and death
Stroke patients who also have COVID-19 showed increased systemic inflammation, a more serious stroke severity and a much higher rate of death, compared to stroke patients who did not have COVID-19, according a retrospective, observational, cross-sectional study of 60 ischemic stroke patients admitted to UAB Hospital between late March and early May 2020.

'Time is vision' after a stroke
University of Rochester researchers studied stroke patients who experienced vision loss and found that the patients retained some visual abilities immediately after the stroke but these abilities diminished gradually and eventually disappeared permanently after approximately six months.

More stroke awareness, better eating habits may help reduce stroke risk for young adult African-Americans
Young African-Americans are experiencing higher rates of stroke because of health conditions such as high blood pressure, diabetes and obesity, yet their perception of their stroke risk is low.

How to help patients recover after a stroke
The existing approach to brain stimulation for rehabilitation after a stroke does not take into account the diversity of lesions and the individual characteristics of patients' brains.

Kids with headache after stroke might be at risk for another stroke
A new study has found a high incidence of headaches in pediatric stroke survivors and identified a possible association between post-stroke headache and stroke recurrence.

High stroke impact in low- and middle-income countries examined at 11th World Stroke Congress
Less wealthy countries struggle to meet greater need with far fewer resources.

Marijuana use might lead to higher risk of stroke, World Stroke Congress to be told
A five-year study of hospital statistics from the United States shows that the incidence of stroke has risen steadily among marijuana users even though the overall rate of stroke remained constant over the same period.

We need to talk about sexuality after stroke
Stroke survivors and their partners are not adequately supported to deal with changes to their relationships, self-identity, gender roles and intimacy following stroke, according to new research from the University of Sydney.

Standardized stroke protocol can ensure ELVO stroke patients are treated within 60 minutes
A new study shows that developing a standardized stroke protocol of having neurointerventional teams meet suspected emergent large vessel occlusion (ELVO) stroke patients upon their arrival at the hospital achieves a median door-to-recanalization time of less than 60 minutes.

Read More: Stroke News and Stroke Current Events is a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for sites to earn advertising fees by advertising and linking to