Beyond a technical bug: Biased algorithms and moderation are censoring activists on social media


(MENAFN- The Conversation) Following Red Dress Day on May 5, a day aimed to raise awareness for Missing and Murdered Indigenous Women and Girls (MMIWG), Indigenous activists and supporters of the campaign found from their Instagram accounts. In response, Instagram released a tweet saying that this was '' ,'' followed by an apology explaining that the platform '' .''

Creators, however, said that not all stories were affected.

And this is not the first time social media platforms have been under scrutiny because of their erroneous censoring of grassroots activists and racial minorities.

Many (BLM) activists were similarly frustrated when Facebook flagged their accounts, but didn''t do enough to against Black people on their platform.

So were these really about technical glitches? Or did they result from the platforms'' discriminatory and biased policies and practices? The answer lies somewhere in between.

Towards automated content moderation

Every time an activist's post is wrongly removed, there are at least three possible scenarios.

First, sometimes the platform deliberately takes down activists'' posts and accounts, usually at request of and/or in co-ordination with the government. This happened when who expressed support for the Iranian general Qassem Soleiman.

In some countries and disputed territories, such as , , platforms censored activists and journalists to allegedly .

Second, a post can be removed through a user-reporting mechanism. To handle unlawful or prohibited communication, social media platforms have indeed .

Applying community standards developed by the platform, content moderators would then review reported content and determine whether a violation had occurred. If it had, the content would be removed, and, in the case of serious or repeat infringements, the user may be temporarily suspended or permanently banned.

This mechanism is problematic. Due to the sheer volume of reports received on a daily basis, there are simply not enough moderators to review each report adequately. Also, complexities and subtleties of language pose real challenges. Meanwhile, marginalized groups reclaiming abusive terms for public awareness, such as BLM and MMIWG, can be misinterpreted as being abusive.

Further, in flagging content, users tend to rely on . User reporting approach is driven by popular opinion of a platform's users while potentially repressing the right to unpopular speech.

Such approach also emboldens , where users exercise their right to voice their opinions while actively silencing others. A notable example is the removal by Facebook of '' ,'' a multi-artist collaboration posted by Coldplay, after a number of users reported the song as ''abusive.''

Third, platforms are increasingly using artificial intelligence (AI) to help identify and remove prohibited content. The idea is that complex algorithms that use natural language processing can flag racist or violent content faster and better than humans possibly can. During the COVID-19 pandemic, social media companies are relying more on AI to cover for tens of thousands of human moderators who . Now, more than ever, algorithms decide what users can and cannot post online.

Algorithmic biases

There's an inherent belief that AI systems are less biased and can scale better than human beings. In practice, however, they''re easily disposed to error and can impose bias on a colossal systemic scale.

In two 2019 computational linguistic studies, researchers discovered that AI intended to identify hate speech may actually end up amplifying racial bias.

In , researchers found that tweets written in African American English commonly spoken by Black Americans are up to twice more likely to be flagged as offensive compared to others. Using a dataset of 155,800 tweets, found a similar widespread racial bias against Black speeches.

What's considered offensive is bound to social context; terms that are slurs when used in some settings may not be in others. Algorithmic systems lack an ability to capture nuances and contextual particularities, which may not be understood by human moderators who test data used to train these algorithms either. This means natural language processing which is often perceived as an objective tool to identify offensive content can amplify the same biases that human beings have.

Algorithmic bias may jeopardize some people who are already at risk by wrongly categorizing them as offensive, criminals or even terrorists. In mid 2020, Facebook deleted at least 35 accounts of on the pretext of terrorism while in reality, they were campaigning against violence and terrorism.




Protesters gather in Winnipeg the day after the jury delivered a not-guilty verdict in the second degree murder trial of Raymond Cormier, the man accused of killing Tina Fontaine. MMIWG activists had their posts removed from Instagram on a day to raise awareness. THE CANADIAN PRESS/John Woods

MMIWG, BLM and the Syrian cases exemplify the dynamic of '' '' where algorithms reinforce older oppressive social relations and re-install new modes of racism and discrimination.

While AI is celebrated as autonomous technology that can develop away from human intervention, it is inherently biased. The inequalities that underpin bias already exist in society and influence who gets the opportunity to build algorithms and their databases, and for what purpose. As such, algorithms do not intrinsically provide ways for marginalized people to escape discrimination, but they also reproduce new forms of inequality along social, racial and political lines.

Despite the apparent problems, algorithms are here to stay. There is no silver bullet, but one can take steps to minimize bias. First is to recognize that there's a problem. Then, making a strong commitment to root out algorithmic biases.

Bias can infiltrate the process anywhere in designing algorithms.

The inclusion of more people from diverse backgrounds within this process — Indigenous, racial minorities, women and other historically marginalized groups — is one of important steps to help mitigate the bias. In the meantime, it is important to push platforms to allow for as much transparency and public oversight as possible.

MENAFN16052021000199003603ID1102087613


Legal Disclaimer:
MENAFN provides the information “as is” without warranty of any kind. We do not accept any responsibility or liability for the accuracy, content, images, videos, licenses, completeness, legality, or reliability of the information contained in this article. If you have any complaints or copyright issues related to this article, kindly contact the provider above.