MIT Scholar Finds Racial Bias in Commercial Facial Analysis Programs

A new study by researchers at the Massachusetts Institute or Technology and Stanford University finds that commercially released facial analysis programs demonstrate both skin-type and gender biases.

The study found that  the computer programs had a very low error rate when determining the gender of light-skinned men. But one program had an error rate of 20 percent in determining the gender of dark-skinned women. The other two programs had an error rate of more than 34 percent when asked to identify the gender of dark-skinned women. For women who had the darkest skin, the systems failed to accurately determine their gender nearly half the time.

The findings raise questions about how today’s neural networks, which learn to perform computational tasks by looking for patterns in huge data sets, are trained and evaluated. Joy Buolamwini, a researcher in the MIT Media Lab states that “what’s important here is the method and how that method applies to other applications. The same data-centric techniques that can be used to try to determine somebody’s gender are also used to identify a person when you’re looking for a criminal suspect.”

Buolamwini is a graduate of the Georgia Institute of Technology, where she majored in computer science. She earned a master’s degree at the University of Oxford as a Rhodes Scholar and is currently at work on a Ph.D. at MIT.

A video about the research can be viewed below.

Related Articles

1 COMMENT

  1. I found this video informative, but not surprising. I used to be a TV associate Producer many years ago and one of the problems in newly integrated stations was making sure the color balance showed dark skinned reporters to their advantage.

    Inexperienced white photographers sometimes did an entire news report with the reporter as a featureless sillohette because they had not balanced the color on their camera.

    Years have passed and now I am a Senior Lecturer in France where I teach English and do LanguageTesting.

    I was surprised by the new ETS (TOEIC, TOEFL…) rule that you must photograph test candidates on a light background which of course would make dark skinned people unrecognizable given the poor quality of the cameras.

    Am I correct in thinking that based on what you have said about present technology based décisions, the making of this regulation stems from ignorance?

Leave a Reply

Get the JBHE Weekly Bulletin

Receive our weekly email newsletter delivered to your inbox

Latest News

Tuskegee University Flight School Receives $6.7 Million in Federal Funding

With a $6.7 million investment from the federal government, Tuskegee University will launch a new bachelor's degree in aviation science. The program will teach students about aviation science and technology and provide them with flight school training.

Three African Americans Appointed to University Faculty Positions

The faculty appointments are Dexter Blackman at Morgan State University in Baltimore, Stephanie Henderson at Alcorn State University in Mississippi, and Yolanda Pierce at Vanderbilt University in Nashville.

Virginia State University Approved to Launch Master’s Degree in Data Analytics

The master's degree in data analytics will prepare students to use data to make strategic technology and business decisions. The new degree program will be the 14th established master's degree at Virginia State University.

Samuel Frimpong Honored for Outstanding Contributions to Mineral Industry Education

Dr. Frimpong was honored by the Society of Mining, Metallurgy, and Exploration with the 2024 Mineral Industry Education Award. He currently serves as a professor of mineral engineering, the Robert H. Quenon Endowed Chair, and vice provost for graduate education at Missouri University of Science and Technology.

Featured Jobs