People Are Testing Out Twitter’s Photo Preview Functionality And Many Results Seem Racist

Computers don’t really make mistakes. They look like they do sometimes, but they are just following the code and inputs that they were assigned… and sometimes there are some interesting inputs involved.

Not too long ago, internauts were laughing at Google’s image search algorithms for showing African American doctors when searching for white American doctors.

Listen beautiful relax classics on our Youtube channel.

This time around, it’s Twitter and its image cropping algorithms. People have begun noticing how Twitter crops images in a way that gives preference to people with white skin as opposed to those with black skin.

Turns out, there’s an AI-based algorithm for Twitter’s photo preview that centers pics in tweets

But it seems to prefer white people over black people, despite having both in the same image as seen below

Image credits: bascule

So, it all started when Twitter user Colin Madland was troubleshooting a colleague on a related matter but on a completely different platform. Turns out, the video conferencing program Zoom seems to get rid of a black person’s head if a custom background is used. It just considers the head a part of the background.

But it wasn’t until he posted some exemplary screenshots to Twitter for further troubleshooting that he understood that Twitter is also up to its neural networking shenanigans. He noticed how, on the mobile version, his long horizontal screenshot was cropped to only include him, and not his colleague.

Having noticed this, other people started creating pictures with white and black people in them with odd, elongated ratios to force the cropping algorithm to choose a centering position. And a little bit of unofficial experimenting showed that the system tends to prefer white faces over black ones.

After one user noticing this, it seems like all of Twitter started testing this feature out

Users simply started posting elongated images that include various faces and seeing the previews

Image credits: alexhanna

A number of tweeters tested this out with stock photos, politicians, and even cartoons, namely the Simpsons characters Lenny and Carl. More often than not, the photos were cropped in a way that showed a preference for white faces.

Back in early 2018, Twitter identified the problem of tweets showing sets of photos that are off-center. They have thus implemented a mechanism that is supposed to recognize the individual elements in a photo and to center on those instead of on the center. Prior to that, it was all run on facial recognition alone, and while not all photos have faces, they needed to improve.

Results vary to some extent, but most of them, however, seem to show a racial bias in Twitter’s AI

Image credits: gnomestale

Like many new technologies, it is not perfect. According to NITS studies, it was proven that today’s facial recognition software performs worse with non-white faces with a false identification factor of 10 to 100 for Asian and African American faces. So, there’s room for improvement.

Now, this is an understandable error and it’s not like the algorithm is purely biased, as some photo cropping results showed black faces too. Liz Kelley of the Twitter Communications Team also explained that the neural network was checked for any racial and even gender bias and there was none detected. However, there is clearly a need for more analysis and testing as these results seem biased.

Besides actual politicians people tested out a bunch of stock photos, whether edited or not

Listen beautiful relax classics on our Youtube channel.

Image credits: sina_rawayama

Image credits: julinhacreicrei

Image credits: doidinhavids

CDO of Twitter, Dantley Davis, also said that it is 100% Twitter’s fault and no one should say otherwise, despite a number of people saying that this isn’t done on purpose—it’s an AI, after all—and fault is irrelevant because the focus should always be on fixing the problem.

For now, it is unsure as to what causes the algorithm to do what it does with preferring white faces over black ones. The only explanation is that it tends to look for the most prominent one in the photo.

One person tried out different colored ties and even inverted the colors with varying results

Image credits: bascule

Apparently, there is even a preference with animals just because they are of a bright color palette

Image credits: Ma_lopess2

Twitter is aware of this problem is working on fixing it

Image credits: belblueberry

While Davis explained that the AI might be focusing on other additional variables to determine where to center the photo, others also hint that the aspect ratio, the background or the color scheme might also have something to do with it. This is, however, a non-scientific explanation that is yet to be confirmed or denied.

The issue caused a bit of a ruckus on the internet with individual tweets exemplifying the bias and some even calling the algorithm racist. Said tweets gained thousands upon thousands of likes and retweets, with one particular tweet comparing Mitch McConnell or Barack Obama getting over 185,000 likes.

The AI seemed biased even with cartoons as seen with Lenny and Carl from the Simpsons

Image credits: _jsimonovski

Same with dogs

Image credits: MarkEMarkAU

The algorithm wasn’t completely biased as there were results where it preferred black people

It was speculated that it may be because of the smile or the prominent faces, which are easier for the AI to pick up

Image credits: TLopesVictorM

The only exception to the rule was this free-for-all where it was faces as well as objects

Image credits: mikaozl

What are your thoughts on this? Let us know in the comment section below!

Here are how some tweeters reacted to this


Rating People Are Testing Out Twitter’s Photo Preview Functionality And Many Results Seem Racist is 5.0 / 5 Votes: 3
Please wait...