\datasetname: a Multi-modal Dataset of Election Fraud Claims on Twitter

\datasetname: a Multi-modal Dataset of Election Fraud Claims on Twitter


The wide spread of unfounded election fraud claims surrounding the U.S. 2020 election had resulted in undermining of trust in the election, culminating in violence inside the U.S. capitol. Under these circumstances, it is critical to understand discussions surrounding these claims on Twitter, a major platform where the claims disseminate. To this end, we collected and release the \datasetname dataset, a multi-modal dataset with 7.6M tweets and 25.6M retweets from 2.6M users related to voter fraud claims. To make this data immediately useful for a wide area of researchers, we further enhance the data with cluster labels computed from the retweet graph, user suspension status, and perceptual hashes of tweeted images. We also include in the dataset aggregated information for all external links and YouTube videos that appear in the tweets. Preliminary analyses of the data show that Twitter’s ban actions mostly affected a specific community of voter fraud claim promoters, and exposes the most common URLs, images and YouTube videos shared in the data.


1 Cornell Tech 2 Cornell University 3 Technion
aa2776@cornell.edu, yiqing@cs.cornell.edu, hanama888@gmail.com, oamir@technion.ac.il, mor@jacobs.cornell.edu

1 Introduction

Free and fair elections are the foundation of every democracy. The 2020 presidential election in the United States was probably one of the most consequential and contentious such events. Two-thirds of the voting-eligible population voted, resulting in the highest turnout in the past 120 years Schaul et al. (2020). The Democratic Party candidate Joe Biden was elected as the president.

Unfortunately, efforts to deligitimize the election process and its results were carried out before, throughout and after the election. Mostly unfounded claims of voter fraud Frenkel (2020) were spread both through public statements by politicians, and on social media platforms. As a result, 34% of Americans say that they do not trust the election results as of December, 2020 NPR (2020). Voter fraud claims without credible evidence have great ramifications on both the integrity of the election and the stability of the U.S. democracy. On January 6th, 2021, believing that the election was ‘stolen’, mobs breached U.S. capitol while the Congress voted to certify Biden as the winner of the election.

Social media platforms like Facebook, Twitter, YouTube and Reddit play a significant role in political events Vitak et al. (2011); Allcott and Gentzkow (2017), and the 2020 election was no exception Ferrara et al. (2020). In particular, Twitter has been the focus of public and media attention as a prominent public square where ideas are adopted and claims – false or true – are spread Vosoughi et al. (2018); Grinberg et al. (2019). It is thus important to understand the participants, discussions, narratives, and allegations around voter fraud claims on this specific platform.

In this work, we release \datasetname, a multi-modal Twitter dataset of 7.6M tweets and 25.6M retweets that are related to voter fraud claims. Using a manually curated set of keywords (e.g., “voter fraud” and “#stopthesteal”) that was further expanded using a data-driven approach, we streamed Twitter activities between October 23rd and December 16th, 2020. We performed various validations on the limits of our stream, given Twitter’s API constraints Morstatter et al. (2013), and estimate that we were able to retrieve around 60% of the data containing our crawled keywords.

We further enhanced the \datasetname dataset in order to make it accessible for a broader set of researchers and future research: (1) We cluster users according to their retweeting dynamics and release the cluster labels; (2) Given Twitter’s widespread post-election suspension action, we crawl and include the user status as of January 10th, 2021; (3) We compute and share the perceptual hashes of 168K images that appeared in the data; (4) We aggregate and share metadata about 138K external links that appeared in the tweets, including 12K unique YouTube videos. Our dataset also allows researchers to calculate the amount of Twitter interactions with the collected tweets, users, and media items, including number of retweets and quotes from various clusters, or from suspended users.

A preliminary analysis finds a significant cluster of users who were promoting the election fraud related claims, with nearly 7.8% of them suspended in January. The suspensions focused on a specific community within the cluster. A simple analysis of the distribution of images, based on visual similarity, exposes that the most broadly shared (by number of tweets) and the most retweeted images are different. Although recent research has shown that voter fraud claims are pushed mainly by mass media Benkler et al. (2020), we also find that external links referenced by promoters of the claims point mostly to low-quality news websites, streaming services, and YouTube videos. Some of the widespread videos claiming ‘evidence’ of voter fraud were published by surprisingly small channels. Most strikingly, all of the top ten channels and videos spreading voter fraud claims were still available on YouTube as of January 11th, 2021.

We believe that the release of \datasetname, the largest public dataset of Twitter discussions around the voter fraud claims, with the enhanced labels and data, will help the broad research community better understand this important topic at a critical time.

2 Data Collection

Our data collection process involved streaming Twitter data using a data-driven manually curated set of keywords and hashtags. We report on the span and volume of the collected data, as well as on analyses estimating its coverage.

2.1 Streaming Twitter data

We used a data-driven approach to generate a list of keywords and hashtags related to election fraud claims in an iterative manner. We started with a single phrase and two derived keywords: voter fraud and #voterfraud. We first used a convenience sample of 11M political tweets consisting of the tweets of 2,262 U.S. political candidates and the replies to those tweets, collected between July 21st and Oct 22nd, 2020 using the Twitter Streaming API Twitter (2019). We then identified hashtags that co-occur with our meta-seed keywords, voter fraud and voterfraud. We selected all hashtags that appeared in at least 10 tweets and co-occurred with either of the meta-seed keywords at least 50% of the time. From the resulting set, we manually filtered out those that were not directly relevant to voter fraud. To this end, two members of the research team reviewed the hashtags, including, if needed, searching for them on Twitter to see whether they produce relevant results. Only the hashtags that were agreed on by both evaluators were added, resulting in an initial set of hashtags that was added to the two original keywords.

We computed the Jaccard coefficient between each of our seed hashtags and all other hashtags that appeared in the new stream. We added to our set all hashtags that had a Jaccard coefficient greater than 0.001 with any of the seed hashtags. Three members of the team again reviewed this list by 1) excluding hashtags that were not related to voter fraud, 2) adding corresponding keywords of the hashtags (e.g. #discardedballots corresponds to discarded ballots), and 3) adding relevant hashtags or keywords that the researchers observed while searching for hashtags from the generated list. Both the seed list and the final list of keywords and hashtags we used for streaming are included in Appendix A (Table 3).

We collected data using the Twitter streaming API Twitter (2019). The \datasetname dataset includes tweets from 17:00, October 23rd, 2020 to 13:00 December 16th, 2020. We expanded the keywords list on Oct. 31st with additional keywords, and added #stopthesteal as it started trending on November 3rd. While streaming, we stored each tweet’s metadata (e.g., user ID, text, timestamp). We also downloaded all image media items included in the tweets. In total, we collected 3,781,524 original tweets, 25,566,698 retweets, and 3,821,579 quote tweets (i.e. tweets that include a reference to another tweet) discussing election fraud claims. Note that quote tweets are included in the Twitter stream when either the new tweet or the referenced (quoted) tweet include one of the keywords or hashtags on the list. In total, we collected tweets from 2,559,018 users who posted, shared or quoted one or more tweets with these keywords.

2.2 Coverage Analysis

Since the Twitter streaming API provides only a sample of the tweets, especially for large-volume keywords Morstatter et al. (2013), we performed multiple tests to evaluate and estimate the coverage of the \datasetname dataset. This analysis suggests that the dataset covers over 60% of the content shared on Twitter using the keywords we tracked.

Retweet and quote coverage. We evaluated the coverage of retweet and quote tweets by comparing the counts of these objects in the stream to Twitter’s metadata. When a new retweet for an original tweet appears in the stream, the API returns the tweet’s metadata including the current retweet count and quote count of the original tweet. In other words, if an original tweet is retweeted, it will appear in the stream as a retweet , and the metadata for will include the total number of retweets of so far. From this metadata, it is easy to define the retweet coverage as the ratio of the total number of retweets ( objects) streamed and stored in our dataset, over the sum of all retweet counts of the original tweets, returned by the API in the latest retweet of each original tweet. The quote coverage is defined analogously. According to this analysis, the \datasetname dataset captured 63.2% of the retweets and 62.6% of the quote tweets. These findings compare favorably with previous work that shows a single API client captures only 37.6% of the retweets through the Streaming API Morstatter et al. (2013).

Comparison with #Election2020. To further evaluate the coverage on the voter fraud tweets, we compared our dataset with a previously published Twitter dataset of the U.S. 2020 election Chen et al. (2020). The creators of the #Election2020 dataset used the streaming API to track 168 keywords that are broadly related to the election and 57 accounts that are tied to candidates running for president.

As in \datasetname, the keyword ‘voter fraud’ was also used to collect data for #Election2020. We used this overlap to estimate our coverage. First, we can directly compare the relative volume and overlap between the ‘voter fraud’ tweets in both datasets. We expect \datasetname to have a higher volume of such tweets because of its more focused set of keywords. Second, if we assume sampling for both streams is independent and random, we could estimate the coverage of \datasetname by looking at the proportion of #Election2020 tweets that also appear in our data.

To this end, we extracted all tweets and retweets that contain this keyword from both datasets posted on two days following the November 3rd election data: November 6th and November 13th. The analysis, performed on December 17th, was limited to two days as we had to obtain the content of the tweets of the #Election2020 dataset by “hydrating” them (i.e. using the tweet IDs to get the full tweet text using the Twitter API). We were unable to hydrate the full data, presumably due to inactive accounts and deleted tweets. The hydration yielded 92.4% of the #Election2020 data on November 6th (a total of 1.4M tweets/3.5M retweets), and 91.1% of the data on November 13th (1.3M tweets/3M retweets).

In total, our \datasetname data includes 45,322 ‘voter fraud’ related tweets on November 6th, 2.3 times as much as recorded in #Election2020. The ratio is even higher on November 13th, when we obtained 47,313 tweets, 3.1 times as much as in #Election2020. Figure 1 breaks down the coverage by dates (separated by rows), in the two datasets (by different colors). From left to right, the bars show the percentages of tweets that are available only in our dataset (dark blue), that are available in both datasets (light blue), and that are available only in #Election2020 (yellow). On any given day, the \datasetname dataset contains substantially more tweets related to voter fraud, as compared to #Election2020, especially when the estimated total volume is lower. On November 13th (second row), \datasetname  contained 95.7% of the combined data (left two bars) while #Election2020 only collected 30.7% (right two bars) of the tweets. These numbers also indicate that \datasetname’s sample includes 32.1% of the related samples in #Election2020 on November 6th and 85.9% on November 13th. We acknowledge that these two numbers are not consistent, presumably because of November 6th’s much higher volume of activity. If these samples are indeed independent, though, it means that our lower bound of coverage is November 6th’s 32.1%.

Figure 1: Coverage comparison between our dataset and #Election2020 for tweets containing ‘voter fraud’.

Based on these coverage analyses, we conclude that \datasetname is, at the time of submission, the largest known public Twitter dataset of voter fraud claims and discussions.

3 Data Enhancement

To ensure the reusability of our data, we took the following steps to enhance the dataset. In addition to raw streaming data, we clustered users according to the retweet dynamics and release the cluster labels. We also queried Twitter for the user status on 10th of January, and share the user status as active/not-found/suspended. Furthermore, to enable research on visual misinformation, we encode all images shared in the tweets with perceptual hash. Finally, we release the URLs, and the metadata of the YouTube videos that appeared in our dataset.

Community Users Relative size % of users
0 860,976 45.6% 1%
1 437,783 23.2% 4.6%
2 342,184 18.1% 14.1%
3 33,857 1.8% 1.5%
4 23,414 1.2% 1.6%
Figure 2: (a) Community statistics. (b) Retweet graph colored by communities. (c) Suspension status (orange: suspended users).

Retweet Graph Communities. Retweet networks have been frequently analyzed in previous works in order to understand political conversations on Twitter Arif et al. (2018); Cherepnalkoski and Mozetič (2016). Using community detection algorithms, researchers are able to study key players, sharing patterns and content on different sides of a discussion surrounding a heated political topic.

We first constructed a retweet graph of the \datasetname dataset, where nodes represent users and directed edges correspond to retweets between the users. The direction of an edge corresponds to the direction of the information spreading in the retweet relation. Edges are weighted according to the number of times the corresponding source user has been retweeted. The resulting network consists of 1,887,736 nodes and 16,718,884 edges.

To detect communities within the graph, we used the Infomap community detection algorithm Bohlin et al. (2014), which captures the flow of information in directed networks. Using the default parameters, the algorithm produces thousands of clusters. By excluding all clusters that contain fewer than 1% of the nodes we are left with 90% of all nodes1 which are clustered into five communities (see Figure 0(a)).

In Figure 1(a), we visualize the retweet network using the Force Atlas 2 layout in Gephi Bastian et al. (2009), using a random sample of 44,474 nodes and 456,372 edges. The nodes are colored according to their computed community as described in Figure 0(a). Edges are colored by their source node. The visualization indicates that the nodes are split between two distinct clusters - community 0 (blue) on the left and communities 1, 2, 3 and 4 on the right. By examining the top users in each community, we conclude that community 0 mostly consists of accounts that detract the voter fraud claims, while the communities on the right consist of accounts that promote the voter fraud claims. Most of the tweets from these users are written in English, except for users in Community 3 who mainly post tweets in Japanese and users in Community 4 who write in Spanish. Community 2 is more deeply embedded in the promoter cluster compared to Community 1, as we observe tweets from Community 1 being retweeted by Community 0 on the left, but not from Community 2. We include the list of top 5 Twitter accounts in each community by the number of community retweets in the Appendix.

For brevity, in the following analyses, we refer to the cluster on the left as the detractor cluster, and the cluster with community 1,2,3,4 on the right as the promoter cluster. Note that due to the partisan nature of the U.S. politics, most of the promoter users are likely aligned with right-leaning politics, and detractor users align with left-leaning politics.

To identify users that are prominent within each of these two cluster, we calculate the closeness centrality of the user nodes in each cluster. In a retweet network this metric can be interpreted as a user’s ability to spread information to other users in the network Okamoto et al. (2008). We compute the top-k closeness centrality to find the 10,000 most central nodes within the detractor and promoter clusters Bisenius et al. (2017).

We release the author’s community label of each tweet, the community label of each user, and a user’s closeness centrality in the detractor and promoter clusters. We also include two additional metrics - retweet count by community and quote count by community . For a tweet , the retweet count by community is the total number of retweets it received from each user in community . The metric is computed analogously for quotes.

Labeling Suspended and Deleted Users When the electoral college were set to confirm the election results on January 6th, 2021, the allegations of voter fraud took a dramatic turn, which culminated in the storming of the US Capitol. Subsequently, Twitter took a harder stance on moderating content on their platform and suspended at least 70,000 accounts that were engaged in propagating conspiracy theories and sharing QAnon-content Twitter (2021). This ban has substantial implications for researchers seeking to understand the spread of voter fraud allegations on Twitter, since the Twitter API does not allow the “hydration” of Tweets from suspended users. In order to understand the distribution of suspensions within our dataset, we queried the updated user status of all users in our dataset on January 10th, a few days following the ban. The Twitter API returns a user status that indicates if the user is active, suspended or not found (presumably deleted). In total, 3.9% of the accounts (99,884 accounts) in our data were suspended.

We enhance the \datasetname dataset by labeling tweets and users that were suspended. This metadata will enable both research and ease hydration by allowing hydraters to skip content that is no longer available. We also include two additional metrics for each tweet: retweet count by suspended users and quote count by suspended users.

Due to its immense public interest, we have retained the full data we retrieved from the 99,884 suspended users including 1,240,405 tweets and 6,246,245 retweets. This detailed data is not part of \datasetname. However, we will distribute an anonymized version of this data to published academic researchers upon request.

Images. Because of their persuasive power and ease of spread, there is a growing interest in analyzing how visual misinformation spreads both within a platform or across platforms Zannettou et al. (2018); Highfield and Leaver (2016); Paris and Donovan (2019); Moreira et al. (2018); Zannettou et al. (2020). However, visual information such as images or videos is difficult for many researchers to study due to computational and storage costs. Here, we make the information about image content shared in \datasetname easier to use by sharing perceptual hash values for these images. With these numeric hash values, researchers can easily find duplicates and near-duplicate images in tweets, without working directly with cumbersome image content. To this end, we download all image media items that were posted in the tweets in the streaming data, and encode them with three different types of perceptual hashes.

Common perceptual hashes are binary strings designed such that the Hamming distance Zauner et al. (2011) between two hashes is close if and only if the two corresponding images are perceptually similar. In other words, an image that is only slightly transformed, for example, by re-sizing, cropping, or rotation, will have a similar hash value to the original image. However, as the definition of perceptual similarity is often subjective and the underlying algorithms are often different, various hashing functions have different performance characteristics dealing with various types of image transformations. Therefore, we encode the images in our dataset with three perceptual hash functions: the Perceptive Hash (pHash), the Average Hash (aHash), and the Wavelet Hash (wHash) Petrov (2017); Zauner et al. (2011).

In total, our streamed tweets included 201,259 image URLs, 167,696 of them were retrieved during streaming. We provide some more details about the distribution of these images in Section 5.

External links. Misinformation campaigns are known to use broad cross-platform information, often via links to other sites Wilson and Starbird (2020); Golovchenko et al. (2020). Hence, we extracted and publish the set of external (non-Twitter) URLs that were referenced in the tweets. For ease of use, we resolved URLs that point to a redirected location (e.g. bit.ly URLs) to their final destination URL. Our streamed tweets included references to a total of 138,411 unique URLs, appearing in 609,513 tweets.

Since a large portion (over 12%) of all URLs in the data were YouTube links, we further enhanced the data with YouTube-specific metadata. A key motivation for this specific focus was the known role YouTube plays generally in spreading misinformation Hussein et al. (2020); Papadamou et al. (2020) and specifically its role in the 2020 election and voter fraud claims Kaplan (2020); Wakabayashi,Daisuke (2020). For each YouTube video that was shared in the collected tweets, we used YouTube’s Data API YouTube (2021), to retrieve the video’s title, description, as well as the id and title of the channel that posted it. We retrieved the YouTube metadata on Jan 1st, 2021. On that data, out of the 13,611 unique video ids that we have queried, 1,608 were no longer available resulting in 12,003 YouTube URLs with full additional metadata.

4 Data Sharing and Format

Our \datasetname dataset is available for download under FAIR principles Wilkinson et al. (2016) in CSV format2. The data includes “item data” tables for tweets, retweets, and users keyed by Twitter assigned IDs and augmented with additional metadata as described below. The data also includes the images that appear in the dataset, indexed by randomly genenerated unique IDs. Finally, the data includes aggregated tables for URLs and for YouTube videos including the information described in Section 3. The dataset tables and the fields they contain are summarized on Github3.

The \datasetname dataset conforms with FAIR principles. The dataset is Findable as it is publicly available on Figshare, with a digital object identifier (DOI): 10.6084/m9.figshare.13571084. It is also Accessible since it can be accessed by anyone in the world through the link. The datset is in csv format, hence it is Interoperable. We release the full dataset with descriptions detailed in this paper, as well as an online tool to explore the dataset at http://voterfraud2020.io, making the dataset Reusable to the research community.

The tables for Tweets and Retweets contain the full set of items that were collected, including from suspended users. These tables do not include raw tweet data beyond the ID, according to Twitter’s ToS. However, to support use of the data without being required to download (“hydrate”) the full set of tweets, we augment the Tweets table with several key properties. For each tweet we provide the number of total retweets as computed by Twitter (retweet/quote_count_metadata), as well as the number of retweets and quotes we streamed for this tweet from users in each of the five main communities (retweet/quote_count_community_X, X ranging from 0 to 4). Note that the latter do not add up to the Twitter metadata due to the coverage issues listed in Section 2.2. The Tweet table properties also include the user_community (0–4) for the user who posted the tweet, computed using methods listed in Section 3. Some of the Twitter accounts were not clustered into one of the five main communities. In this case, the user_community label is null. With this augmentation, researchers using this dataset could very quickly, for example, select and then hydrate a subset of the most retweeted tweets from non-suspended users in Community 2. As the tweet itself and the ID of the user who tweeted it is not available until hydration, Twitter’s users’ privacy is preserved.

The Users table is similarly augmented with aggregate information about the importance of the user in the dataset, including the community that they belong to, their centrality in the two meta-clusters, detractor and promoter (closeness_centrality_detractor_cluster and closeness_centrality_promoter_cluster), and the amount of attention (retweets and quotes) they received from other users in the different communities. We also note whether, according to the data we collected, the user had been suspended. With this data, interested researchers can quickly focus their attention and research on the main actors in each community.

The Images table includes all the image media items retrieved in the stream, their unique media ID, and the ID of the tweet in which the image was shared. We augment this table with the image hash using three types of perceptual hash functions – aHash, pHash and wHash, as detailed in Section 3. This augmentation, together with the link to the Tweet ID, will allow researchers to quickly identify and hydrate popular images using the tweet metadata. They can also quickly identify and get information for images that are similar to any other arbitrary image, by computing and comparting the perceptual hash values.

The two aggregate tables, the URLs table and the YouTube Videos table again provide information about the popularity of the object in the dataset: aggregate retweet and quote counts, both using the Twitter metadata and the count of objects in our stream from the various communities. In addition, these tables are augmented with metadata about the item (URL or YouTube video) as noted in Section 3.

5 Data Analysis

We performed a preliminary analysis of our dataset and its different modalities – tweets and users, images, external links – to demonstrate its potential interest and provide some initial guiding insights about the data.

Figure 3: Temporal overview of the dataset showing number of streamed tweets, quotes and retweets per day. The shaded regions mark the expansions of the keyword set.
(a) (b) (c)
Tweets Retweets (total) Retweets (in cluster) Tweets Retweets (total) Retweets (in cluster) Tweets Retweets (total) Retweets (in cluster)
15 24,399 18,020 11 20,104 10,424 34 28,833 10,250
Figure 4: Top three most retweeted images in the promoter cluster: (a)–(c), with the number of tweets, retweets as in metadata, and retweets by users in the cluster. Image (c) was cropped to fit the figure.

Tweets and users. Figure 3 shows the amount of retweets (green), original tweets (blue) and quote tweets (yellow) in the \datasetname dataset over the time (X-axis) of the data collection. Three shaded regions, from left to right, mark the expansion of our set of keywords on October 31st (light blue, region b) and November 3rd (light green, region c). The Y-axis specifies the daily count. In general, except for the large increase after the election date (November 3rd, dotted vertical line), the volume of the stream remains roughly the same. On average, there are 170,938 tweets, 576,136 retweets, and 85,488 quote tweets per day after the election.

Our manual inspection shows that top tweets retweeted by the detractor cluster often condemn the alleged voter fraud claims, while top tweets on the promoter cluster indeed make voter fraud claims. Not surprisingly, among the top ten most retweeted tweets in the promoter cluster, nine were tweeted by President Trump. We refer readers to our project website for more details about popular tweets.

While the promoter clusters seems rather homogeneous (Figure 1(a)), users in Community 2 (yellow) stand out in both their level of activity and the rate in which they were suspended. Community 2 was highly active in our dataset. For example, Community 2 comprises 18.1% of the users, but contributed 68% of the \datasetname tweets, and 74% of the retweets. Moreover, 14% of Community 2’s users were suspended by Twitter by the time we collected the account status data as described above, a much higher rate than the other communities, as shown in Figure 0(a). In total, Community 2 was responsible for 46.1% of all suspensions amongst the users we associated with the top five communities. The suspension effect, and its focus on Community 2, can also be observed in Figure 1(b).

A full analysis of the suspended accounts and their network communities, and the potential impact of the suspension is out of scope for this dataset paper, but can be easily performed using the data we share in \datasetname. For example, the data shows that 35% of the promoter cluster users that were retweeted more than 1,000 times (1,596 in total) were suspended.

To conclude, our preliminary analysis shows that alleged election fraud claims mostly circulate in the promoter cluster, and in particular in Community 2 within the cluster. The most popular tweets (by retweet counts) supporting such claims often come from prominent accounts. The recent moderation efforts from Twitter seem to have effected the most active community that engaged in fraud related misinformation, and did not broadly target all accounts involved in promoting such claims.


We conducted a preliminary examination of matching and repeated images in \datasetname to analyze the distribution of images related to voter fraud claims. Our data, using the perceptual hash functions described in Section 3, allows tracking of duplicate and near-duplicate images that were posted in multiple tweets. In this analysis, we experimented with three perceptual hash functions and refer to two images as matching if they have an identical perceptual hash value.

In Figure 4(a), we show the cumulative distribution of the number of unique perceptual hashes in \datasetname (Y-axis), with hash values sorted based on the number of unique tweets in which they appear, from the highest to the lowest (X-axis). For example, according to pHash, the 1,000 images shared in the largest number of unique tweets appeared together in 25,019 different tweets (not including retweets). Although in general the results are similar when using different hash functions, pHash is the most “conservative” in terms of assigning matches.

Overall, our results are similar when using different hash functions. For example, there are 109,312 (out of 167,696) images with the same pHash value. Of these, 17,831 were shared in more than one tweet, an average of 4.27 times. In other words, 34% of the images instances in \datasetname tweets appear in more than one tweet. Figure 4(b) presents the image that appeared in most number of unique tweets: the same perceptual hash value appeared in over 1,000 tweets, according to all three hash functions.

We further investigate the popularity of images, defined by number of retweets, in particular, within the promoter and detractor clusters. After grouping images by the same pHash value, we present in Figure 4 the top three images that have been retweeted in the promoter cluster. Also note that despite the high values of metadata retweets and cluster retweets, all these “popular” images appeared in only a few original tweets in our data. For example, image (a) appeared in 15 tweets, whose metadata retweet (as returned from the API) counts add up to 24,399 in total, and was retweeted (as recorded in our dataset) from users in the promoter cluster 18,020 times. We note that images (a) and (b) were also the top two images retweeted by users in the suspended users set, with 5,547 and 3,122 retweets in that set, respectively (recall that as almost all suspended users belong to the promoter cluster). As expected, the most retweeted images in the two clusters are quite different. The three most retweeted images in the detractor cluster (not included for lack of space) have somewhat lower spread, appearing in tweets that were retweeted 10743, 6425, 3411 times (based on metadata). The top image is a screenshot of the NY Times front page of Nov 11th, 2020 reporting that top election officials across the country have not identified any fraud.

The analysis presented above can be easily extended with less-strict image similarity matching by calculating the Hamming distance between a pair of perceptual hash values. In this initial analysis, we used a strict sense of similarity, treating images as similar only when they share the same perceptual hash values.


Figure 5: (a) The cumulative number of repeated images by hash matches. (b) The most tweeted image.
promoter cluster detractor cluster
Domain Retweets Domain Retweets
pscp.tv 51,822 washingtonpost.com 11,220
youtube.com 44,031 rawstory.com 9,267
thegatewaypundit.com 35,967 cnn.com 4,139
davidharrisjr.com 18,793 independent.co.uk 3,882
foxnews.com 17,332 nytimes.com 3,746
theepochtimes.com 15,297 newsweek.com 3,496
thedcpatriot.com 14,958 news.yahoo.com 2,899
thefederalist.com 13,288 deadstate.org 2,409
djhjmedia.com 11,816 theguardian.com 2,232
justthenews.com 11,149 politicususa.com 2,032
Table 1: Top 10 domains being retweeted in the promoter and the detractor clusters respectively, as well as the number of retweets by users in these clusters.

URLs. We conduct preliminary analyses of the external links that have been included in the \datasetname tweets. Table 1 lists the top 10 domains that have been shared inside the detractor and promoter clusters respectively. Most of the links shared by users in the detractor clusters are to mainstream news media, such as the Washington Post, CNN, and the New York Times. The rest are other news-related websites. The links shared by users in the promoter cluster mostly point to low-quality news-related websites.

The most shared domain in the promoter cluster is pscp.tv, a live video streaming app that’s owned by Twitter. YouTube stands out as the second most retweeted domain among the promoter users. This trend is reflected in multiple news reports, warning of the significant role that YouTube plays in spreading false information related to voter fraud claims Frenkel (2020). The majority of the top 10 most retweeted videos by the promoter users falsely claim evidence of widespread election fraud. The users spreading these videos had significant overlap with the January (or earlier) suspension action by Twitter. For eight of these videos, around of the retweets of tweets sharing those videos were by accounts later suspended by Twitter.

A scan of the top 10 YouTube channels retweeted in the promoter cluster shows that they were relatively large (millions of subscribers), though there are also several smaller channels. For example, the most retweeted channel, Precinct 13, has only 3.67K subscribers, has a video that appeared in 88 tweets and have been retweeted over 9K times.

Despite YouTube’s announcement that it will take actions against content creators who falsely claim the existence of widespread voter fraud4, as of Jan 11th, the top 10 channels and videos listed in our tables are still available on YouTube.

6 Related Work and Datasets

We review prior work using Twitter data analysing politically related events, with an emphasis on those that have released a public dataset.

In particular, previous works had used and published Twitter data to study U.S. elections. Using tweets collected during the 2016 U.S. election, researchers have analysed information operations run by social bots Rizoiu et al. (2018), characterized the dissemination of misinformation Vosoughi et al. (2018) and its exposure to American voters Grinberg et al. (2019). Work in Hua et al. (2020a, b) characterized adversarial interaction against political candidates during the 2018 U.S. general election and shared 1.7M tweets interacting with political candidates.

Focusing on the U.S. 2020 election, research studied false claims regarding mail-in ballots Benkler et al. (2020) before the election as the COVID-19 pandemic made it hard to vote in person. Closest to our work is the #Election2020 dataset Chen et al. (2020), which streamed a broad set of Twitter data for both political candidates’ tweets and related keywords. As discussed above, although some of the voter fraud related keywords were included in their data collection process, our \datasetname dataset contains more than 2.3 times as much of the related data in #Election2020, for overlapping streaming keywords, presumably because of our more focused stream. Our stream also included a broader set of fraud-claim related keywords.

In order to help understand the dissemination of misinformation cross platforms, Brena et al. (2019); Hui et al. (2018) used news articles as queries and released the tweets pointing to these articles. In 2018, Twitter published a list of accounts that the platform suspects to be related with Russia’s government controlled Internet Research Agency Twitter (2018). This release enabled a number of studies that deepened our understanding of foreign information manipulation in the U.S. Arif et al. (2018); Im et al. (2020); Badawy et al. (2018).

Most of the previous works that released Twitter datasets only included the tweet IDs, in accordance with Twitter’s Terms of Service. We keep to that practice, and augment the data without sharing tweet content, as detailed above, making our multi-modal dataset more accessible and useful to the research community.

7 Discussion and Conclusions

The voter fraud allegations to discredit the U.S. 2020 presidential elections are likely to form one of the most consequential misinformation campaigns in modern history. It is critical to allow a diverse set of researchers to provide a deeper understanding of this effort, which will continue to have national and global impact for years to come. To enable that contribution, it is important to provide a public and accessible archive of this campaign on various social media platforms, including Twitter as we do in \datasetname.

The \datasetname dataset has the potential to benefit the research community, and to further inform the public regarding both Twitter activities around the voter fraud claims, as well as Twitter’s response. Yet, the data has some limitations. We could not possibly capture the full extent of the voter fraud claims on Twitter, as our dataset was constructed by using matching keywords. Further, as discussed above, we do not have full coverage even for the keywords we tracked, though we estimate that we have a majority of the tweets with those keywords. Nevertheless, the breadth of the data enables various types of investigation using both the tweet data, as well as the aggregated data of URLs, videos and images used in the campaign. We propose three major categories of such investigation.

First, researchers can use the dataset to study the spread, reach, and dynamics of the voter fraud campaign on Twitter. Researchers can describe and analyze the participants, including the activities of political candidates using information from orthogonal data sets of candidate accounts 5, and the interaction between public figures and other accounts spreading claims and promoting certain narratives. Further, the data can help expose how different public figures spread different claims, for example the claims regarding the Dominion voting machines, what kind of engagement such narratives received. The data can also be used to understand the role of bots and other coordinated activities and campaigns in spreading this information. In general, the dataset can provide for analysis of the distribution of attention to these claims and how it spreads – via images, tweets, URLs – including comparison among different pre-computed communities and clusters.

Second, we include auxiliary data – URLs including YouTube links, and image hashes – that can help researchers examine other sources of information and their roles in spreading these claims. For example, using the image hash values that were encoded using publicly available algorithms, researchers can easily map images not just within the Twitter data, but also within the larger web ecosystem. Researchers may combine our dataset with datasets that are collected from other social media platforms to examine how visual misinformation spread cross platforms (e.g., Zannettou et al. (2018); Moreira et al. (2018)).

A third potential area of investigation is Twitter’s response to the voter fraud claims. A specific question is the characterization of the suspended users, who are primarily part of a specific community even within the the group promoting voter fraud claims as shown above. Researchers can use the data to both understand Twitter’s non-public response, and its potential effectiveness, or even simulate the effectiveness of hypothetical earlier bans of the same population. As noted above, while Twitter’s terms forbid us from publicly sharing full data for these suspended users – the \datasetname tweets for these users are no longer available on Twitter by their ID – we will make these tweets available privately to published academic researchers, as we believe these tweets are of immense and justified public interest.

The publicly released \datasetname data was collected and made available according to Twitter’s Terms of Service for academic researchers, following established guidelines for ethical Twitter data use Rivers and Lewis (2014). By limiting to the Tweet IDs as the main data element, the dataset does not expose information about users whose data had been removed from the service. The only content in our data that is directly tied to a Tweet ID is the hash of the images for tweets that included them. Even though that hash, theoretically, can be tied to an image from another source, in absence of the original tweet the image will not be associated with any user account. We believe that this minor disclosure risk is justified given the potential benefits of this data.

Appendix A Appendix

Community 0
ID Handle Active Status Retweets
32871086 kylegriffin1 active 76,302
1640929196 mmpadellan active 74,393
255812611 donwinslow active 69,796
216776631 BernieSanders active 60,961
15952856 AriBerman active 58,222
a) Community 1
25073877 realDonaldTrump suspended 1,560,373
187680645 LLinWood suspended 1,057,805
586707638 SidneyPowell1 suspended 633,273
240454812 GenFlynn suspended 334,197
1812055789 CodeMonkeyZ suspended 274,210
b) Community 2
2922345639 DonnaWR8 suspended 38,388
259260816 zeusFanHouse suspended 36,347
393190233 LeahR77 suspended 33,352
951302891708583936 TheRISEofROD suspended 32,992
32804484 Bubblebathgirl active 27,787
c) Commmunity 3
835040085573689346 ganaha_masako active 12,480
1128981340848656384 KadotaRyusho active 6,890
796450109986902016 yamatogokorous active 5,716
1166577240601239552 mei98862477 active 5,347
109458204 kohyu1952 active 5,244
d) Community 4
3393186119 FernandoAmandi active 4,217
1126414392080232449 POTUS_Trump_ESP active 2,981
1195348350620622850 TDN_NOTICIAS active 2,459
98294131 1VAFI active 1,802
1068238181282267137 Gamusina77 active 1,638
Table 2: Top 5 Users in each community sorted by retweets from other users.
Seed list #abolishdemocratparty #ballotharvasting #ballotvoterfraud #cheatingdemocrats #democratvoterfraud #gopvoterfraud #ilhanballotharvesting #ilhanomarballotharvesting #ilhanomarvoterfraud #mailinvoterfraud #stopvoterfraud #voterfraud #voterfraudbymail #voterfraudisreal
Filtered #abolishdemocratparty
Generated from the seed list #ballotharvesting #voterid #ilhanomarforprison #stopgopvoterfraud #ilhanomar #nancypelosiabusingpower #nancypelosimustresign #junkmailballots #traresforcongress #immigrationfraud #votebymailfraud #ballotfraud #exposed #votersuppression #ilhanresign #voteinperson #votebymail #video #lockherup #nomailinvoting #ilhanomarelectionfraud #taxfraud #ballotharvesting #massivemailinballots #arrestilhanomar #obamagate #ilhanomarlockherup #buyingvotes #2020election #campaignfraud #homewrecker #voteinperson #minneapolis #absenteeballots #trump2020 #arrestilhanomar #absenteeballot #darktolight #wwg1wga #terrorist #daveygravyspirualsavage #trump #fraud #liar #pizzagate #republicans #qproof #theawakening #voteatthepolls #marriedherbrother #glasshouses #sheepnomore #voteyouout #cheater #georgesoros #georgia #vote #walkaway #thegreatawakening #qanon #evil #savethechildren
Keywords list 10/24 #ballotfraud #ballotharvesting #ballotvoterfraud #cheatingdemocrats #democratvoterfraud #ilhanomarballotharvesting #ilhanomarvoterfraud #mailinvoterfraud #nomailinvoting #stopgopvoterfraud #stopvoterfraud #votebymailfraud #voterfraud #voterfraudisreal
Added on 10/31 #discardedballots #electionfraud #electioninterference #electiontampering #gopvoterfraud #hackedvotingmachines ‘destroyed ballots’ ‘discarded ballots’ ‘election fraud’ ‘election interference’ ‘election tampering’ ‘hacked voting machine’ ‘pre-filled ballot’ ‘stolen ballots’ ‘ballot fraud’ ‘ballot harvesting’ ‘cheating democrats’ ‘democrats cheat’ ‘harvest ballot’ ‘vote by mail fraud’ ‘voter fraud’
Added on 11/03 #stopthesteal
Table 3: Hashtags and keywords related to election fraud.


  1. Since the graph only includes retweeting and retweeted users, this number corresponds to 73.8% of all users in our dataset.
  2. https://figshare.com/account/projects/96518/articles/13571084
  3. https://github.com/sTechLab/VoterFraud2020
  4. see: https://twitter.com/YouTubeInsider/status/1347231471212371970
  5. https://github.com/vegetable68/Midterm-2020-candidates


  1. Social media and fake news in the 2016 election. Journal of economic perspectives 31 (2), pp. 211–36. Cited by: §1.
  2. Acting the part: examining information operations within# blacklivesmatter discourse. Proceedings of the ACM on Human-Computer Interaction 2 (CSCW), pp. 1–27. Cited by: §3, §6.
  3. Analyzing the digital traces of political manipulation: the 2016 russian interference twitter campaign. In 2018 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), pp. 258–265. Cited by: §6.
  4. Gephi: an open source software for exploring and manipulating networks. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 3. Cited by: §3.
  5. Mail-in voter fraud: anatomy of a disinformation campaign. Available at SSRN. Cited by: §1, §6.
  6. Computing Top- k Closeness Centrality in Fully-dynamic Graphs. pp. 1–26. External Links: arXiv:1710.01143v1 Cited by: §3.
  7. Community Detection and Visualization of Networks with the Map Equation Framework. In Measuring Scholarly Impact: Methods and Practice, Y. Ding, R. Rousseau and D. Wolfram (Eds.), pp. 3–34. External Links: Document, ISBN 978-3-319-10377-8 Cited by: §3.
  8. News sharing user behaviour on twitter: a comprehensive data collection of news articles and social interactions. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 13, pp. 592–597. Cited by: §6.
  9. # election2020: the first public twitter dataset on the 2020 us presidential election. arXiv preprint arXiv:2010.00600. Cited by: §2.2, §6.
  10. Retweet networks of the European Parliament: evaluation of the community structure. Applied Network Science 1 (1), pp. 1–20. External Links: Document, ISSN 23648228, Link Cited by: §3.
  11. Characterizing social media manipulation in the 2020 us presidential election. First Monday. Cited by: §1.
  12. How misinformation ‘superspreaders’ seed false election theories. Note: \urlhttps://www.nytimes.com/2020/11/23/technology/election-misinformation-facebook-twitter.html[Online; accessed 4-Jan-2021] Cited by: §1, §5.
  13. Cross-platform state propaganda: russian trolls on twitter and youtube during the 2016 us presidential election. The International Journal of Press/Politics, pp. 1940161220912682. Cited by: §3.
  14. Fake news on twitter during the 2016 us presidential election. Science 363 (6425), pp. 374–378. Cited by: §1, §6.
  15. Instagrammatics and digital methods: studying visual social media, from selfies and gifs to memes and emoji. Communication Research and Practice 2 (1), pp. 47–62. Cited by: §3.
  16. Characterizing twitter users who engage in adversarial interactions against political candidates. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–13. Cited by: §6.
  17. Towards measuring adversarial twitter interactions against candidates in the us midterm elections. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 14, pp. 272–282. Cited by: §6.
  18. The hoaxy misinformation and fact-checking diffusion network. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 12. Cited by: §6.
  19. Measuring misinformation in video search platforms: an audit study on youtube. Proceedings of the ACM on Human-Computer Interaction 4 (CSCW1), pp. 1–27. Cited by: §3.
  20. Still out there: modeling and identifying russian troll accounts on twitter. In 12th ACM Conference on Web Science, pp. 1–10. Cited by: §6.
  21. YouTube has allowed conspiracy theories about interference with voting machines to go viral. Note: \urlhttps://www.mediamatters.org/google/youtube-has-allowed-conspiracy-theories-about-interference-voting-machines-go-viral[Online; accessed 5-Jan-2021] Cited by: §3.
  22. Image provenance analysis at scale. IEEE Transactions on Image Processing 27 (12), pp. 6109–6123. Cited by: §3, §7.
  23. Is the sample good enough? Comparing data from twitter’s streaming API with Twitter’s firehose. Proceedings of the 7th International Conference on Weblogs and Social Media, ICWSM 2013, pp. 400–408. External Links: 1306.5204 Cited by: §1, §2.2, §2.2.
  24. Poll: just a quarter of republicans accept election outcome. Note: \urlhttps://www.npr.org/2020/12/09/944385798/poll-just-a-quarter-of-republicans-accept-election-outcome[Online; accessed 5-Jan-2021] Cited by: §1.
  25. Ranking of Closeness Centrality for Large-Scale Social Networks Ranking of Closeness Centrality for Large-Scale Social Networks. (June). External Links: Document, ISBN 9783540693116 Cited by: §3.
  26. ” It is just a flu”: assessing the effect of watch history on youtube’s pseudoscientific video recommendations. arXiv preprint arXiv:2010.11638. Cited by: §3.
  27. Deepfakes and cheap fakes. United States of America: Data & Society. Cited by: §3.
  28. Wavelet image hash in python. Note: \urlhttps://fullstackml.com/wavelet-image-hash-in-python-3504fdd282b5[Online; accessed 15-Jan-2021] Cited by: §3.
  29. Ethical research standards in a world of big data. F1000Research 3. Cited by: §7.
  30. # debatenight: the role and influence of socialbots on twitter during the 1st 2016 us presidential debate. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 12. Cited by: §6.
  31. 2020 turnout is the highest in over a century. Note: \urlhttps://www.washingtonpost.com/graphics/2020/elections/voter-turnout/[Online; accessed 5-Jan-2021] Cited by: §1.
  32. Update on twitter’s review of the 2016 us election. Note: \urlhttps://blog.twitter.com/official/en_us/topics/company/2018/2016-election-update.html[Online; accessed 7-Jan-2021] Cited by: §6.
  33. Twitter standard api. Note: \urlhttps://developer.twitter.com/en/docs/tweets/filter-realtime/overview[Online; accessed 15-Jan-2019] Cited by: §2.1, §2.1.
  34. An update, following the riots in washington, dc. Note: \urlhttps://blog.twitter.com/en_us/topics/company/2021/protecting–the-conversation-following-the-riots-in-washington–.html[Online; accessed 12-Jan-2021] Cited by: §3.
  35. It’s complicated: facebook users’ political participation in the 2008 election. CyberPsychology, behavior, and social networking 14 (3), pp. 107–114. Cited by: §1.
  36. The spread of true and false news online. Science 359 (6380), pp. 1146–1151. Cited by: §1, §6.
  37. Election misinformation continues staying up on youtube.. Note: \urlhttps://www.nytimes.com/2020/11/10/technology/election-misinformation-continues-staying-up-on-youtube.html[Online; accessed 5-Jan-2021] Cited by: §3.
  38. The fair guiding principles for scientific data management and stewardship. Scientific data 3 (1), pp. 1–9. Cited by: §4.
  39. Cross-platform disinformation campaigns: lessons learned and next steps. Harvard Kennedy School Misinformation Review 1 (1). Cited by: §3.
  40. YouTube data api — google developers. Note: \urlhttps://developers.google.com/youtube/v3[Online; accessed 8-Jan-2021] Cited by: §3.
  41. On the origins of memes by means of fringe web communities. In Proceedings of the Internet Measurement Conference 2018, pp. 188–202. Cited by: §3, §7.
  42. Characterizing the use of images in state-sponsored information warfare operations by russian trolls on twitter. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 14, pp. 774–785. Cited by: §3.
  43. Rihamark: perceptual image hash benchmarking. In Media watermarking, security, and forensics III, Vol. 7880, pp. 78800X. Cited by: §3.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description