New York – Social media giant Twitter said Monday it would investigate its image-cropping function after users complained it favored white faces over Black ones.
The image preview feature of Twitter’s mobile app automatically crops pictures that are too big to fit on the screen, selecting which parts of the image to display and which to conceal.
Prompted by a graduate student who found an image he was posting cropped out the face of a Black colleague, a San Francisco-based programmer found Twitter’s system would crop out images of President Barack Obama when posted together with images of Republican Senate Leader Mitch McConnell.
“Twitter is just one example of racism manifesting in machine learning algorithms,” the programmer, Tony Arcieri, wrote on Twitter.
Twitter is one of the world’s most popular social networks, with nearly 200 million daily users.
Other users shared similar experiments online they said showed Twitter’s cropping system favoring white people.
Twitter said in a statement, “Our team did test for bias before shipping the model and did not find evidence of racial or gender bias in our testing.”
However, it said it would look further into the issue.
“It’s clear from these examples that we’ve got more analysis to do. We’ll continue to share what we learn, what actions we take, and will open source our analysis so others can review and replicate,” Twitter said in its statement.
In a 2018 blog post, Twitter had said the cropping system was based on a “neural network” that used artificial intelligence to predict what part of a photo would be interesting to a user and crop out the rest.
A representative of Twitter also pointed to an experiment by a Carnegie Mellon University scientist who analyzed 92 images and found the algorithm favored Black faces 52 times.
But Meredith Whittaker, co-founder of the AI Now Institute that studies the social implications of artificial intelligence, said she was not satisfied with Twitter’s response.
“Systems like Twitter’s image preview are everywhere, implemented in the name of standardization and convenience,” she told Thomson Reuters Foundation.
“This is another in a long and weary litany of examples that show automated systems encoding racism, misogyny and histories of discrimination.”
In a time of both misinformation and too much information, quality journalism is more crucial than ever.
By subscribing, you can help us get the story right.