Latest stories
Tesla's in China – It's just a question of how long
You can tell a Moon Jae-in by the company he keeps
India's Teesta River funding: ambition or illusion? Harmless engagement-farming or sophisticated propaganda?At first glance, the motivation for these accounts to generate interest may appear obvious – social media engagement leads to advertising revenue. If a person sets up an account that receives inflated engagement, they may earn a share of advertising revenue from social media organizations such as Meta.
So, does the dead internet theory stop at harmless engagement farming? Or perhaps beneath the surface lies a sophisticated, well-funded attempt to support autocratic regimes, attack opponents and spread propaganda?
While the shrimp Jesus phenomenon may seem harmless (albeit bizarre), there is potentially a longer-term ploy at hand.
As these AI-driven accounts grow in followers (many fake, some real), the high follower count legitimizes the account to real users. This means that out there, an army of accounts is being created. Accounts with high follower counts could be deployed by those with the highest bid.
This is critically important, as social media is now the primary news source for many users around the world. In Australia, 46% of 18 to 24-year-olds nominated social media as their main source of news last year. This is up from 28% in 2022, taking over from traditional outlets such as radio and TV.
Bot-fueled disinformationAlready, there is strong evidence social media is being manipulated by these inflated bots to sway public opinion with disinformation – and it's been happening for years.
In 2018, a study analyzed 14 million tweets over a ten-month period in 2016 and 2017. It found bots on social media were significantly involved in disseminating articles from unreliable sources. Accounts with high numbers of followers were legitimizing misinformation and disinformation, leading real users to believe, engage and reshare bot-posted content.
This approach to social media manipulation has been found to occur after mass shooting events in the United States. In 2019, a study found bot-generated posts on X (formerly Twitter) heavily contribute to the public discussion , serving to amplify or distort potential narratives associated with extreme events.
More recently, several large-scale, pro-Russian disinformation campaigns have aimed to undermine support for Ukraine and promote pro-Russian sentiment .
Uncovered by activists and journalists, the coordinated efforts used bots and AI to create and spread fake information, reaching millions of social media users.
On X alone, the campaign used more than 10,000 bot accounts to rapidly post tens of thousands of messages of pro-Kremlin content attributed to US and European celebrities seemingly supporting the ongoing war against Ukraine.
This scale of influence is significant. Some reports have even found that nearly half of all internet traffic in 2022 was made by bots . With recent advancements in generative AI – such as OpenAI's ChatGPT models and Google's Gemini – the quality of fake content will only be improving.
Social media organizations are seeking to address the misuse of their platforms. Notably, Elon Musk has explored requiring X users to pay for membership to stop bot farms.
Sign up for one of our free newsletters The Daily ReportStart your day right with Asia Times' top stories AT Weekly ReportA weekly roundup of Asia Times' most-read stories
Social media giants are capable of removing large amounts of detected bot activity if they so chose. (Bad news for our friendly shrimp Jesus.)
Keep the dead internet in mindThe dead internet theory is not really claiming that most of your personal interactions on the internet are fake.
It is, however, an interesting lens through which to view the internet. That it is no longer for humans, by humans – this is the sense in which the internet we knew and loved is“dead”.
The freedom to create and share our thoughts on the internet and social media is what made it so powerful. Naturally, it is this power that bad actors are seeking to control.
The dead internet theory is a reminder to be skeptical and navigate social media and other websites with a critical mind.
Any interaction, trend and especially“overall sentiment” could very well be synthetic. Designed to slightly change the way in which you perceive the world.
Jake Renzella is Lecturer and Director of Studies (Computer Science), UNSW Sydney and Vlada Rozova is Research Fellow in Applied Machine Learning, The University of Melbourne
This article is republished from The Conversation under a Creative Commons license. Read the original article .
Already have an account?Sign in Sign up here to comment on Asia Times stories OR Thank you for registering!
An account was already registered with this email. Please check your inbox for an authentication link.