In a recent study, researchers from USC find that software robots pretending to be humans are influencing the political discourse on social media as never before and could threaten the very integrity of the 2016 U.S. presidential election.
By leveraging state-of-the art bot detection algorithms, the researchers have made a startling discovery:
A surprisingly high percentage of the political discussion taking place on Twitter was created by pro-Donald Trump and pro-Hillary Clinton software robots, or social bots, with the express purpose of distorting the online discussion regarding the elections.
The researchers analyzed 20 million election-related tweets created between Sept. 16 and Oct. 21.
They found that robots, rather than people, produced 3.8 million tweets, or 19 percent. Social bots also accounted for 400,000 of the 2.8 million individual users, or nearly 15 percent of the population under study.
“The presence of these bots can affect the dynamics of the political discussion in three tangible ways,” writes in a recently released paper titled, “Social Bots Distort the 2016 U.S. Presidential Election Online Discussion,” appearing the journal First Monday.
“First, influence can be redistributed across suspicious accounts that may be operated with malicious purposes. Second, the political conversation can become further polarized. Third, spreading of misinformation and unverified information can be enhanced.”
“As a result, the integrity of the 2016 U.S. presidential election could be possibly endangered.”
Interestingly, Trump’s robot-produced tweets were almost uniformly positive, boosting the candidate popularity.
By contrast only half of Clinton’s were, with the other half criticizing the nominee, according to the research paper. South Carolina produced the most fake campaign-related tweets, the study reports.
Because of social bots’ sophistication, it’s often impossible to determine who creates them, although political parties, local, national and foreign governments and “even single individuals with adequate resources could obtain the operational capabilities and technical tools to deploy armies of social bots and affect the directions of online political conversation,” the report says.
The “master puppeteers” behind influence bots, the researchers added, often create fake Twitter and Facebook profiles.
They do so by stealing online pictures, giving them fictitious names, and cloning biographical information from existing accounts.
These bots have become so sophisticated that they can tweet, retweet, share content, comment on posts, “like” candidates, grow their social influence by following legitimate human accounts and even engage in human-like conversations.
News source: University of Southern California.
Figure legend: This Knowridge.com image is for illustrative purposes only.