• Home
  • Latest
  • Fortune 500
  • Finance
  • Tech
  • Leadership
  • Lifestyle
  • Rankings
  • Multimedia
TechAI

Google’s Hate Speech Detection A.I. Has a Racial Bias Problem

By
Jonathan Vanian
Jonathan Vanian
Down Arrow Button Icon
By
Jonathan Vanian
Jonathan Vanian
Down Arrow Button Icon
August 16, 2019, 1:29 PM ET

A Google-created tool that uses artificial intelligence to police hate speech in online comments on sites like the New York Times has become racially biased, according to a new study.

The tool, developed by Google and a subsidiary of its parent company, often classified comments written in the African-American vernacular as toxic, researchers from the University of Washington, Carnegie Mellon, and the Allen Institute for Artificial Intelligence said in a paper presented in early August at the Association for Computational Linguistics conference in Florence, Italy.

The findings underscore machine learning’s susceptibility to bias and its unintended consequences on groups that are underrepresented in the data used to train machine-learning systems. A.I. researchers who are studying the problem worry that the technology─which learns to recognize patterns within vast quantities of data─could perpetrate inequality.

Google has dealt with machine learning bias problems in that past, like in 2015 when one of its tools used to search photos tagged African Americans as gorillas. Another tool released in 2018 that matched people’s selfies with popular artworks, inadvertently correlated the faces of African Americans with artwork depicting slaves, “perhaps because of an overreliance on Western art,” journalist Vauhini Vara wrote in Fortune.

The study also highlights the limitations of machine learning as a tool to filter and screen content. Despite its ability to automate tasks, thus reducing the load from human moderators, it often fails at understanding context, like deciphering whether a biting joke may be funny to one person and yet upsetting to another.

Jigsaw, the subsidiary that helped create the tool, called Perspective, pitched it for its debut in 2017 as a way for publishers like The Economist and The Guardian to automatically filter offensive comments, leading to more civil discourse. The idea, similar to initiatives by companies like Facebook and Twitter, is that machine-learning powered software can help filter the deluge of abusive and hateful online comments that are often left online without being deleted.

But this type of content-moderation software can stumble when trying to understand the nuances of human language. Natural language processing technology, or NLP—a subset of A.I. that helps computers understand language—often fails to understand context.

Maarten Sap, a University of Washington graduate student who is studying NLP and who co-authored the paper, said that his team was interested in studying social inequality and how it’s represented through language. The researchers had read previous reports of hate-speech detection tools inadvertently flagging African-American English, raising concerns that technology was silencing the voices of African-Americans in online comments or on social media.

For their research, the academics first gathered two datasets of over 100,000 Twitter comments and that are typically used by other researchers for hate-speech detection projects. Human annotators from the data-labeling company Figure Eight (now owned by data company Appen) inspected the tweets and labeled them as being offensive, hateful, or benign.

After inspecting the datasets, the researchers noticed that the human annotators would often label tweets commonly associated with African-American vernacular as being offensive or hateful, despite the phrases being typical to that particular dialect. These phrases might contain certain words that other social groups may find offensive, like the N-word, “ass,” or “bitch.”

The researchers then used the Twitter data to train a neural network—software that learns—to recognize offensive or hateful phrases. The trained neural network then analyzed two other datasets containing Tweets in which the Twitter users’ racial identities were either known or inferred, and associated common African-American phrases as being offensive, like the human annotators.

When the researchers tested the Perspective tool in the same manner, they found that it also labeled tweets in African-American English as being toxic, Sap said.

“That’s a pretty big problem,” said Sap, pointing to the impact that biased content filters would have on online comments.

Dan Keyserling, Jigsaw’s chief operating officer, said that Jigsaw representatives are now working with the researchers who presented the paper to improve its technology.

“We really welcome this kind of research,” Keyserling said. “We are constantly attending these conferences and try to support the research community.”

The researchers also conducted a small, related experiment to learn more about how human annotators label data. They hired temporary human annotators through Amazon’s Mechanical Turk service and sent them a collection of tweets by African-Americans to label as offensive or not.

In this case, because the human annotators knew that African-Americans wrote the tweets, they were significantly less likely to classify certain phrases as being offensive. Sap speculated that these crowdsourced workers found the N-Word to be less offensive when used by African-Americans, but is “probably more offensive if it is said by a white person.” In the case of the original labeled datasets, human annotators “just didn’t know the context,” he said.

Keyserling acknowledged that the Perspective tool, like other machine-learning systems, is susceptible to bias. In fact, journalists and researchers have called out Perspective in the past for similar instances of bias against certain groups.

But, Keyserling contends that Jigsaw has made great strides in reducing biases in its technology, pointing to previous blog posts and research papers that the company has released detailing its challenges with biases.

Keyserling did not identify where the human annotators Jigsaw uses to label data for training Perspective come from, or whether Jigsaw would specifically use of any of the techniques outlined in the paper to reduce future instances of bias. This could include giving human annotators more context about people’s race when reading their comments.

However, Keyserling noted that “context really matters” and “that’s absolutely something we are paying attention to when structuring our data.”

In a follow up email, Keyserling pointed to a technique Jigsaw uses during the data training process to mitigate bias, which involves so-called model cards that detail how the machine-learning models should be used and ethical considerations to consider. In a previous blog post, Jigsaw said the technique helped reduce unintended bias in its Perspective tool that may impact people who identity as gay, homosexual, or black.

“We are always working on ways to improve our models and mitigate unintended bias on an on-going basis,” Keyserling said.

This likely won’t be the last time Perspective stumbles and develops a bias. As Jigsaw expands the use of its technology in more languages—French newspaper Le Monde recently began using a French version of the tool, for example—it will likely face similar problems with other dialects.

But Keyserling contends that Perspective is “delivering impact” and is “incredibly helpful to platforms” that are otherwise overwhelmed with hateful online comments. Consistently updating the technology to eliminate biases is a never-ending task.

“The technology will never be perfect,” Keyserling said. “That is the nature of machine-learning research.”

More must-read stories from Fortune:

—What you need to know about 8chan, the controversial site tied to the El Paso shooting
—Verizon’s unlimited plans are getting cheaper. Here’s what you should know
—What CEOs, bankers, and tech execs think about a coming recession
—How an alleged Amazon theft ring got the goods
—Boeing adds a second flight control computer to the 737 Max
Catch up with Data Sheet, Fortune‘s daily digest on the business of tech.

About the Author
By Jonathan Vanian
LinkedIn iconTwitter icon

Jonathan Vanian is a former Fortune reporter. He covered business technology, cybersecurity, artificial intelligence, data privacy, and other topics.

See full bioRight Arrow Button Icon

Latest in Tech

Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025

Most Popular

Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Finance
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam
By Fortune Editors
October 20, 2025
Rankings
  • 100 Best Companies
  • Fortune 500
  • Global 500
  • Fortune 500 Europe
  • Most Powerful Women
  • Future 50
  • World’s Most Admired Companies
  • See All Rankings
Sections
  • Finance
  • Leadership
  • Success
  • Tech
  • Asia
  • Europe
  • Environment
  • Fortune Crypto
  • Health
  • Retail
  • Lifestyle
  • Politics
  • Newsletters
  • Magazine
  • Features
  • Commentary
  • Mpw
  • CEO Initiative
  • Conferences
  • Personal Finance
  • Education
Customer Support
  • Frequently Asked Questions
  • Customer Service Portal
  • Privacy Policy
  • Terms Of Use
  • Single Issues For Purchase
  • International Print
Commercial Services
  • Advertising
  • Fortune Brand Studio
  • Fortune Analytics
  • Fortune Conferences
  • Business Development
About Us
  • About Us
  • Editorial Calendar
  • Press Center
  • Work At Fortune
  • Diversity And Inclusion
  • Terms And Conditions
  • Site Map

Latest in Tech

Brown
CybersecuritySocial Media
Mass shootings on campus give rise to a new kind of life-saving service journalism: an anonymous message board called Sidechat
By Leah Willingham and The Associated PressJanuary 8, 2026
5 hours ago
PoliticsDefense
Founder of $30 billion defense tech company Anduril embraces Trump’s threat to crack down: It’s ‘good to scare people sometimes’
By Marco Quiroz-GutierrezJanuary 8, 2026
7 hours ago
Jassy
Workplace CultureAmazon
Amazon demands proof of productivity from employees, asking for list of accomplishments
By Jake AngeloJanuary 8, 2026
9 hours ago
kappos
CommentaryEconomics
The Nobel Prize winners have a lesson for us all
By David J. KapposJanuary 8, 2026
10 hours ago
Dario Amodei sits in a white chair in front of a pink background and speaks animatedly.
AIEye on AI
AI is boosting productivity. Here’s why some workers feel a sense of loss
By Sharon GoldmanJanuary 8, 2026
10 hours ago
Mark DesJardine
CommentaryM&A
Warner Bros. Discovery’s board isn’t choosing a deal — it’s avoiding one
By Mark DesJardineJanuary 8, 2026
10 hours ago

Most Popular

placeholder alt text
Law
Amazon is cutting checks to millions of customers as part of a $2.5 billion FTC settlement. Here's who qualifies and how to get paid
By Sydney LakeJanuary 6, 2026
2 days ago
placeholder alt text
Future of Work
AI layoffs are looking more and more like corporate fiction that's masking a darker reality, Oxford Economics suggests
By Nick LichtenbergJanuary 7, 2026
1 day ago
placeholder alt text
Success
Diary of a CEO founder says he hired someone with 'zero' work experience because she 'thanked the security guard by name' before the interview
By Emma BurleighJanuary 8, 2026
11 hours ago
placeholder alt text
Future of Work
'Employers are increasingly turning to degree and GPA' in hiring: Recruiters retreat from ‘talent is everywhere,’ double down on top colleges
By Jake AngeloJanuary 6, 2026
2 days ago
placeholder alt text
Economy
Mark Cuban on the $38 trillion national debt and the absurdity of U.S. healthcare: we wouldn't pay for potato chips like this
By Nick LichtenbergJanuary 6, 2026
2 days ago
placeholder alt text
Workplace Culture
Amazon demands proof of productivity from employees, asking for list of accomplishments
By Jake AngeloJanuary 8, 2026
9 hours ago

© 2025 Fortune Media IP Limited. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | CA Notice at Collection and Privacy Notice | Do Not Sell/Share My Personal Information
FORTUNE is a trademark of Fortune Media IP Limited, registered in the U.S. and other countries. FORTUNE may receive compensation for some links to products and services on this website. Offers may be subject to change without notice.