Skip to Content
Silicon Valley

Amazon ditched AI recruitment software because it was biased against women

October 10, 2018

The data on which the artificial-intelligence algorithm was trained created a preference for male candidates.

The news: According to a report by Reuters, Amazon began developing an automated system in 2014 to rank job seekers with one to five stars. But last year, the company scrapped the project after seeing it had developed a preference for male candidates in technical roles.

Why? The AI tool was trained on 10 years’ worth of résumés the company had received. Because tech is a male-dominated industry, the majority of those résumés came from men.

The result: The system was unintentionally trained to choose male candidates over female candidates. It would reportedly penalize résumés containing the word “women’s” or the names of certain all-women colleges. Although Amazon made changes to make these terms neutral, the company lost confidence that the program was indeed gender neutral in all other areas.

Why it matters: We can’t treat artificial intelligence as inherently unbiased. Training the systems on biased data means the algorithms also become biased. If unfair AI hiring programs like this aren’t uncovered before being implemented, they will perpetuate long-standing diversity issues in business rather than solve them.

Want to keep up with how technology is changing the workplace of the future? Sign up for our future of work newsletter, Clocking In.

Keep Reading

Most Popular

AI can now create a replica of your personality

A two-hour interview is enough to accurately capture your values and preferences, according to new research from Stanford and Google DeepMind.

The 8 worst technology failures of 2024

Vertical farms, woke AI, and 23andMe made our annual list of failed tech.

Why AI could eat quantum computing’s lunch

Rapid advances in applying artificial intelligence to simulations in physics and chemistry have some people questioning whether we will even need quantum computers at all.

These AI Minecraft characters did weirdly human stuff all on their own

Hundreds of LLM-powered AI agents spontaneously made friends, invented jobs, and spread religion.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.