Exploring Hate (Brookings)
Rhys Leahy, Nicolas Velásquez, Nicholas Johnson Restrepo, Yonatan Lupu, Beth Goldberg, Neil F. Johnson
Exploring Hate (Brookings)
Rhys Leahy, Nicolas Velásquez, Nicholas Johnson Restrepo, Yonatan Lupu, Beth Goldberg, Neil F. Johnson
Frontiers in Political Science
The current military conflict between Russia and Ukraine is accompanied by disinformation and propaganda within the digital ecosystem of social media platforms and online news sources. One month prior to the conflict’s February 2022 start, a Special Report by the U.S. Department of State had already highlighted concern about the extent to which Kremlin-funded media were feeding the online disinformation and propaganda ecosystem. Here we address a closely related issue: how Russian information sources feed into online extremist communities. Specifically, we present a preliminary study of how the sector of the online ecosystem involving extremist communities interconnects within and across social media platforms, and how it connects into such official information sources. Our focus here is on Russian domains, European Nationalists, and American White Supremacists. Though necessarily very limited in scope, our study goes beyond many existing works that focus on Twitter, by instead considering platforms such as VKontakte, Telegram, and Gab. Our findings can help shed light on the scope and impact of state-sponsored foreign influence operations. Our study also highlights the need to develop a detailed map of the full multi-platform ecosystem in order to better inform discussions aimed at countering violent extremism.
Rhys Leahy, Nicholas Johnson Restrepo, Richard Sear, Neil F. Johnson
Axios
There is a clear and growing link between Russian propaganda and online far-right extremism globally, according to a new study from researchers at the George Washington University.
Keynote talk at the International Workshop on Cyber Social Threats (CySoc) June 2022, Atlanta, GA: https://cysoc2022.github.io/
Advances in Artificial Intelligence and Machine Learning
The natural, physical and social worlds abound with feedback processes that make the challenge of modeling the underlying system an extremely complex one. This paper proposes an end-to-end deep learning approach to modelling such so-called complex systems which addresses two problems: (1) scientific model discovery when we have only incomplete/partial knowledge of system dynamics; (2) integration of graph-structured data into scientific machine learning (SciML) using graph neural networks. It is well known that deep learning (DL) has had remarkable success in leveraging large amounts of unstructured data into downstream tasks such as clustering, classification, and regression. Recently, the development of graph neural networks has extended DL techniques to graph structured data of complex systems. However, DL methods still appear largely disjointed with established scientific knowledge, and the contribution to basic science is not always apparent. This disconnect has spurred the development of physics-informed deep learning, and more generally, the emerging discipline of SciML. Modelling complex systems in the physical, biological, and social sciences within the SciML framework requires further considerations. We argue the need to consider heterogeneous, graph-structured data as well as the effective scale at which we can observe system dynamics. Our proposal would open up a joint approach to the previously distinct fields of graph representation learning and SciML.
Nicholas Gabriel, Neil F. Johnson
Calls are escalating for social media platforms to do more to mitigate extreme online communities whose views can lead to real-world harms, e.g., mis/disinformation and distrust that increased Covid-19 fatalities, and now extend to monkeypox, unsafe baby formula alternatives, cancer, abortions, and climate change; white replacement that inspired the 2022 Buffalo shooter and will likely inspire others; anger that threatens elections, e.g., 2021 U.S. Capitol attack; notions of male supremacy that encourage abuse of women; anti-Semitism, anti-LGBQT hate and QAnon conspiracies. But should ‘doing more’ mean doing more of the same, or something different? If so, what? Here we start by showing why platforms doing more of the same will not solve the problem. Specifically, our analysis of nearly 100 million Facebook users entangled over vaccines and now Covid and beyond, shows that the extreme communities’ ecology has a hidden resilience to Facebook’s removal interventions; that Facebook’s messaging interventions are missing key audience sectors and getting ridiculed; that a key piece of these online extremes’ narratives is being mislabeled as incorrect science; and that the threat of censorship is inciting the creation of parallel presences on other platforms with potentially broader audiences. We then demonstrate empirically a new solution that can soften online extremes organically without having to censor or remove communities or their content, or check or correct facts, or promote any preventative messaging, or seek a consensus. This solution can be automated at scale across social media platforms quickly and with minimal cost.
Elvira Maria Restrepo, Martin Moreno, Lucia Illari, Neil F. Johnson
Proceedings of the 2021 Conference of The Computational Social Science Society of the Americas
The COVID-19 pandemic sparked an online “infodemic” of potentially dangerous misinformation. We use machine learning to quantify COVID-19 content from opponents of establishment health guidance, in particular vaccination. We quantify this content in two different ways: number of topics and evolution of keywords. We find that, even in the early stages of the pandemic, the anti-vaccination community had the infrastructure to more effectively garner support than their pro-vaccination counterparts by exhibiting a broader array of discussion topics. This provided an advantage in terms of attracting new users seeking COVID-19 guidance online. We also find that our machine learning framework can pick up on the adaptive nature of discussions within the anti-vaccination community, tracking distrust of authorities, opposition to lockdown orders, and an interest in early vaccine trials. Our approach is scalable and hence tackles the urgent problem facing social media platforms of having to analyze huge volumes of online health misinformation. With vaccine booster shots being approved and vaccination rates stagnating, such an automated approach is key in understanding how to combat the misinformation that slows the eradication of the pandemic.
Richard Sear, Rhys Leahy, Nicholas Johnson Restrepo, Yonatan Lupu, Neil Johnson
Advances in Artificial Intelligence and Machine Learning
Not only can online hate content spread easily between social media platforms, but its focus can also evolve over time. Machine learning and other artificial intelligence (AI) tools could play a key role in helping human moderators understand how such hate topics are evolving online. Latent Dirichlet Allocation (LDA) has been shown to be able to identify hate topics from a corpus of text associated with online communities that promote hate. However, applying LDA to each day’s data is impractical since the inferred topic list from the optimization can change abruptly from day to day, even though the underlying text and hence topics do not typically change this quickly. Hence, LDA is not well suited to capture the way in which hate topics evolve and morph. Here we solve this problem by showing that a dynamic version of LDA can help capture this evolution of topics surrounding online hate. Specifically, we show how standard and dynamical LDA models can be used in conjunction to analyze the topics over time emerging from extremist communities across multiple moderated and unmoderated social media platforms. Our dataset comprises material that we have gathered from hate-related communities on Facebook, Telegram, and Gab during the time period January-April 2021. We demonstrate the ability of dynamic LDA to shed light on how hate groups use different platforms in order to propagate their cause and interests across the online multiverse of social media platforms.
Richard Sear, Rhys Leahy, Nicholas Johnson Restrepo, Yonatan Lupu, Neil F. Johnson
Forbes
Any parent knows that Facebook mom groups are a confusing mix of support, criticism and downright bad advice. But they were still mainstream communities where parents could find connection around what can often be the lonely job of raising kids.