Summary: Researchers report Russian trolls and bots are significantly involved in promoting discord and spreading false information about vaccines on Twitter. The study reports these accounts shared anti-vaccination messages 75% more than average Twitter users.
Source: George Washington University.
Social media bots and Russian trolls promoted discord and spread false information about vaccines on Twitter, according to new research led by a School of Engineering and Applied Science researcher.
Employing tactics similar to those used during the 2016 U.S. presidential election, these Twitter accounts engaged in online debates about vaccines months before election season was underway. The research, “Weaponized Health Communication: Twitter Bots and Russian Trolls Amplify the Vaccine Debate,” was published today in the American Journal of Public Health.
The team examined thousands of tweets sent between July 2014 and September 2017. It discovered several accounts—now known to belong to the same Russian trolls who interfered in the U.S. election—as well as marketing and malware bots, tweeted about vaccines and skewed online health communications.
“The vast majority of Americans believe vaccines are safe and effective, but looking at Twitter gives the impression that there is a lot of debate. It turns out that many anti-vaccine tweets come from accounts whose provenance is unclear. These might be bots, human users or ‘cyborgs’—hacked accounts that are sometimes taken over by bots,” said David Broniatowski, a SEAS assistant professor.
“Although it’s impossible to know exactly how many tweets were generated by bots and trolls, our findings suggest that a significant portion of the online discourse about vaccines may be generated by malicious actors with a range of hidden agendas.”
The research team also includes experts from the University of Maryland and Johns Hopkins University. They found that content polluters—bot accounts that distribute malware, unsolicited commercial content and disruptive materials—shared anti-vaccination messages 75 percent more than average Twitter users.
“Content polluters seem to use anti-vaccine messages as bait to entice their followers to click on advertisements and links to malicious websites,” said Sandra Crouse Quinn, a research team member and professor in the School of Public Health at the University of Maryland. “Ironically, content that promotes exposure to biological viruses may also promote exposure to computer viruses.”
Russian trolls and more sophisticated bot accounts used a different tactic, posting equal amounts of pro- and anti-vaccination tweets. Dr. Broniatowski’s team reviewed more than 250 tweets about vaccination sent by accounts linked to the Internet Research Agency, a Russian government-backed company recently indicted by a U.S. grand jury for its attempts to interfere in the 2016 U.S. elections. The researchers found the tweets used polarizing language linking vaccination to controversial issues in American society, such as racial and economic disparities. For example, one tweet suggested that “clean” vaccines were reserved only for the elites, and questioned the safety of vaccines given to normal people.
“These trolls seem to be using vaccination as a wedge issue, promoting discord in American society,” said Mark Dredze, a professor of computer science at Johns Hopkins. “However, by playing both sides, they erode public trust in vaccination, exposing us all to the risk of infectious diseases. Viruses don’t respect national boundaries.”
Source: George Washington University
Publisher: Organized by NeuroscienceNews.com.
Image Source: NeuroscienceNews.com image is in the public domain.
Original Research: Open access research for “Weaponized Health Communication: Twitter Bots and Russian Trolls Amplify the Vaccine Debate” by David A. Broniatowski, Amelia M. Jamison, SiHua Qi, Lulwah AlKulaib, Tao Chen, Adrian Benton, Sandra C. Quinn, and Mark Dredze in American Journal of Public Health. Published August 23 2018.
Weaponized Health Communication: Twitter Bots and Russian Trolls Amplify the Vaccine Debate
To understand how Twitter bots and trolls (“bots”) promote online health content.
We compared bots’ to average users’ rates of vaccine-relevant messages, which we collected online from July 2014 through September 2017. We estimated the likelihood that users were bots, comparing proportions of polarized and antivaccine tweets across user types. We conducted a content analysis of a Twitter hashtag associated with Russian troll activity.
Compared with average users, Russian trolls (χ2(1) = 102.0; P < .001), sophisticated bots (χ2(1) = 28.6; P < .001), and “content polluters” (χ2(1) = 7.0; P < .001) tweeted about vaccination at higher rates. Whereas content polluters posted more antivaccine content (χ2(1) = 11.18; P < .001), Russian trolls amplified both sides. Unidentifiable accounts were more polarized (χ2(1) = 12.1; P < .001) and antivaccine (χ2(1) = 35.9; P < .001). Analysis of the Russian troll hashtag showed that its messages were more political and divisive.
Whereas bots that spread malware and unsolicited content disseminated antivaccine messages, Russian trolls promoted discord. Accounts masquerading as legitimate users create false equivalency, eroding public consensus on vaccination.
Public Health Implications.
Directly confronting vaccine skeptics enables bots to legitimize the vaccine debate. More research is needed to determine how best to combat bot-driven content.