Nav: Home

Algorithms reveal changes in stereotypes, according to new Stanford research

April 18, 2018

Artificial intelligence systems and machine-learning algorithms have come under fire recently because they can pick up and reinforce existing biases in our society, depending on what data they are programmed with.

But an interdisciplinary group of Stanford scholars turned this problem on its head in a new Proceedings of the National Academy of Sciences paper published April 3.

The researchers used word embeddings - an algorithmic technique that can map relationships and associations between words - to measure changes in gender and ethnic stereotypes over the past century in the United States. They analyzed large databases of American books, newspapers and other texts and looked at how those linguistic changes correlated with actual U.S. Census demographic data and major social shifts such as the women's movement in the 1960s and the increase in Asian immigration, according to the research.

"Word embeddings can be used as a microscope to study historical changes in stereotypes in our society," said James Zou, an assistant professor of biomedical data science. "Our prior research has shown that embeddings effectively capture existing stereotypes and that those biases can be systematically removed. But we think that, instead of removing those stereotypes, we can also use embeddings as a historical lens for quantitative, linguistic and sociological analyses of biases."

Zou co-authored the paper with history Professor Londa Schiebinger, linguistics and computer science Professor Dan Jurafsky and electrical engineering graduate student Nikhil Garg, who was the lead author.

"This type of research opens all kinds of doors to us," Schiebinger said. "It provides a new level of evidence that allow humanities scholars to go after questions about the evolution of stereotypes and biases at a scale that has never been done before."

The geometry of words

A word embedding is an algorithm that is used, or trained, on a collection of text. The algorithm then assigns a geometrical vector to every word, representing each word as a point in space. The technique uses location in this space to capture associations between words in the source text.

"Embeddings are a powerful linguistic tool for measuring subtle aspects of word meaning, such as bias," Jurafsky said.

Take the word "honorable." Using the embedding tool, previous research found that the adjective has a closer relationship to the word "man" than the word "woman."

In its new research, the Stanford team used embeddings to identify specific occupations and adjectives that were biased toward women and particular ethnic groups by decade from 1900 to the present. The researchers trained those embeddings on newspaper databases and also used embeddings previously trained by Stanford computer science graduate student Will Hamilton on other large text datasets, such as the Google Books corpus of American books, which contains over 130 billion words published during the 20th and 21st centuries.

The researchers compared the biases found by those embeddings to demographical changes in the U.S. Census data between 1900 and the present.

Shifts in stereotypes

The research findings showed quantifiable shifts in gender portrayals and biases toward Asians and other ethnic groups during the 20th century.

One of the key findings to emerge was how biases toward women changed for the better - in some ways - over time.

For example, adjectives such as "intelligent," "logical" and "thoughtful" were associated more with men in the first half of the 20th century. But since the 1960s, the same words have increasingly been associated with women with every following decade, correlating with the women's movement in the 1960s, although a gap still remains.

The research also showed a dramatic change in stereotypes toward Asians and Asian Americans.

For example, in the 1910s, words like "barbaric," "monstrous" and "cruel" were the adjectives most associated with Asian last names. By the 1990s, those adjectives were replaced by words like "inhibited," "passive" and "sensitive." This linguistic change correlates with a sharp increase in Asian immigration to the United States in the 1960s and 1980s and a change in cultural stereotypes, the researchers said.

"The starkness of the change in stereotypes stood out to me," Garg said. "When you study history, you learn about propaganda campaigns and these outdated views of foreign groups. But how much the literature produced at the time reflected those stereotypes was hard to appreciate."

Overall, the researchers demonstrated that changes in the word embeddings tracked closely with demographic shifts measured by the U.S. Census.

Fruitful collaboration

The new research illuminates the value of interdisciplinary teamwork between humanities and the sciences, researchers said.

Schiebinger said she reached out to Zou, who joined Stanford in 2016, after she read his previous work on de-biasing machine-learning algorithms.

"This led to a very interesting and fruitful collaboration," Schiebinger said, adding that members of the group are working on further research together.

"It underscores the importance of humanists and computer scientists working together. There is a power to these new machine-learning methods in humanities research that is just being understood," she said.

Zou is also a member of Stanford Bio-X and a Chan Zuckerberg Biohub investigator.
-end-


Stanford University

Related Stereotypes Articles:

Effects of gender bias, stereotypes in surgical training
This randomized clinical trial investigated the association between pro-male gender bias and negative stereotypes against women during surgical residency on surgical skills and proactive career development of residents in general surgery training programs.
Do girls read better than boys? If so, gender stereotypes may be to blame
A new longitudinal study of fifth and sixth graders in Germany examined the relation between classmates' gender stereotypes and individual students' reading outcomes to shed light on how these stereotypes contribute to the gender gap in reading.
Bad behavior between moms driven by stereotypes, judgment
Mothers are often their own toughest critics, but new Iowa State University research shows they judge other mothers just as harshly.
Even scientists have gender stereotypes ... which can hamper the career of women researchers
However convinced we may be that science is not just for men, the concept of science remains much more strongly associated with masculinity than with femininity in people's minds.
Study: Some stereotypes seem to be universally applied to biracial groups in the US
A new Northwestern University study has found evidence that there are some stereotypes that seem to be universally applied to biracial groups in the U.S.
In fiction young people choose traditional love and gender stereotypes
Fictional television series can have an influence on the construction of young people's identities and values.
Stereotypes of romantic love may justify gender-based violence
The media have become key agents of socialization in the construction of teenagers' and young people's identities.
Facebook memes during 2016 presidential election differ from gender stereotypes
Facebook political memes of Donald Trump in the 2016 election were more likely to focus on his hairstyle and facial expressions, while those of Hillary Clinton were more likely to center on the email scandal and her relationships -- a contrast to historical gender stereotypes in politics, a Baylor University study has found.
Us vs. them: Understanding the neurobiology of stereotypes
In a review published in the journal Trends in Cognitive Science, Alvaro Pascual-Leone, MD, PhD, and colleagues describe how non-invasive brain stimulation -- a technique he and others have pioneered to unlock the secrets of the brain -- could shed light on the neurobiology underlying implicit bias.
Algorithms reveal changes in stereotypes, according to new Stanford research
New Stanford research shows that, over the past century, linguistic changes in gender and ethnic stereotypes correlated with major social movements and demographic changes in the US Census data.
More Stereotypes News and Stereotypes Current Events

Trending Science News

Current Coronavirus (COVID-19) News

Top Science Podcasts

We have hand picked the top science podcasts of 2020.
Now Playing: TED Radio Hour

Processing The Pandemic
Between the pandemic and America's reckoning with racism and police brutality, many of us are anxious, angry, and depressed. This hour, TED Fellow and writer Laurel Braitman helps us process it all.
Now Playing: Science for the People

#568 Poker Face Psychology
Anyone who's seen pop culture depictions of poker might think statistics and math is the only way to get ahead. But no, there's psychology too. Author Maria Konnikova took her Ph.D. in psychology to the poker table, and turned out to be good. So good, she went pro in poker, and learned all about her own biases on the way. We're talking about her new book "The Biggest Bluff: How I Learned to Pay Attention, Master Myself, and Win".
Now Playing: Radiolab

Invisible Allies
As scientists have been scrambling to find new and better ways to treat covid-19, they've come across some unexpected allies. Invisible and primordial, these protectors have been with us all along. And they just might help us to better weather this viral storm. To kick things off, we travel through time from a homeless shelter to a military hospital, pondering the pandemic-fighting power of the sun. And then, we dive deep into the periodic table to look at how a simple element might actually be a microbe's biggest foe. This episode was reported by Simon Adler and Molly Webster, and produced by Annie McEwen and Pat Walters. Support Radiolab today at Radiolab.org/donate.