AI and Russian propaganda: it’s not what it looks like

[Note: This article was originally published in the On Think Tanks Annual Review. It was written by David Sichinava and Dustin Gilbreath. David Sichinava is the Research Director of CRRC Georgia. Dustin Gilbreath is the Deputy Research Director of CRRC Georgia and the Communications Manager at Transparify. The views presented in this article do not reflect the views of East West Management Institute, USAID, or any related entity.]

In the think tank world, talk about artificial intelligence (AI) is common. Using it is less common. One of the underlying causes of this may be a perceived lack of familiarity with the methods. However, AI methods – including machine learning – are probably more familiar to many thinktankers than they realise. The Russian Propaganda Barometer project, recently conducted by the Caucasus Research Resource Centers (CRRC) Georgia, demonstrates the potential of these tools in think tanks for policy insight – particularly relating to discourse analysis, and developing targeting strategies.

Artificial intelligence and machine learning are more familiar than thinktankers think
To say that artificial intelligence in general, and machine learning algorithms specifically, is a dramatically changing industry would be an understatement. From optimising electricity usage in factories to deciding which advertisement to show you online, algorithms are in use all around us. In fact, algorithms have been shaping the world around us for decades.

The think tank and social science worlds are no exceptions to this. Indeed, most policy researchers will be familiar with, if not users of, algorithms like regression. Notably, this is a common tool in the machine learning world as well social science research.

Hopefully, knowing that regression is part of the machine learning toolbox will make it clear that machine learning is less foreign than many thinktankers may think.

While regression is one method in the machine learning toolbox, there are others. Although these methods are not new, this larger toolbox has only become commonly used in recent years as big data sets have become more available.

For many products and problems, machine learning solutions might be improvements on existing think tank practices. This is particularly true when it comes to developing a targeting strategy for programming, monitoring, or anything that focuses on understanding discourses.

The Russian Propaganda Barometer Project
CRRC Georgia implemented the Russian Propaganda Barometer project, funded by USAID through the East West Management Institute in 2018-2019. The project aimed to understand and monitor sources of Russian propaganda in Georgia, and to identify who was more or less likely to be vulnerable to the propaganda.

To monitor Russian propaganda, CRRC took all of the posts from public Facebook pages of potential sources of Russian propaganda (around 50,000 in total) in the Georgian language as identified by two other organizations working on the issue in addition to several pages missing from their lists. These posts were then analysed using natural language processing tools such as sentiment analysis. Network analysis was also conducted to understand the interlinkages between different sources.

One of the key insights from the project is that most of the sources of propaganda identified were in fact from far right organisations. While some of these are likely tied to Russia, an analysis of how they talked about the West and Russia suggests that most actually have more negative attitudes towards Russia than the West.

The analysis also called attention to the sharp rise in interest in the far right in Georgia. The number of interactions with far-right pages had increased by roughly 800% since 2015. While overall increasing internet use in the country likely contributed to this, it seems unlikely to be the only cause of the rise.

The results were presented in this dashboard, as well as a more traditional report. It enables users to see what the far right is talking about on a daily basis, and networks between different groups, among other metrics.

The project also aimed to inform a targeting strategy on countering anti-Western propaganda. To do so, we merged data from approximately 30 waves of CRRC and National Democratic Institute surveys that asked about a variety of preferences. From there, a ‘k-nearest neighbours’ algorithm was used to identify which groups had uncertain or inchoate foreign policy preferences. This algorithm basically identifies how similar people are based on whatever variables are included in the algorithm. Based on similarity, a prediction is then made about whatever outcome is of interest. This led to an algorithm that provided accurate predictions about two thirds of the time as to whether someone would be more or less likely to be influenced by Russian propaganda. Further research showed that the algorithm was stable in predicting whether someone was at risk of being influenced, using data that did not exist at the time of the algorithm’s creation.

The data analysis, while cutting edge in many respects, is not beyond the means of many quantitative researchers. Neither of us have MAs or PhDs in statistics: David is a geographer and Dustin is a political scientist.

While the Russian Propaganda Barometer addressed the research goals, we’d like to highlight that AI is no panacea. For the project’s success, we combined traditional think tank analysis of the situation in Georgia with AI to generate new insights.

The Russian Propaganda Barometer project is just one type of application of machine learning to policy research. There is good reason to believe more and more policy researchers will use these methods given their ubiquity in the modern world, together with the increasing availability of the large datasets needed to study these issues.  We hope that the Russian Propaganda Barometer project can serve as food for thought for others in service of this goal.

გაზიარება