Examining a hate speech corpus
for hate speech detection and popularity prediction
As research on hate speech becomes more and more relevant every day, most of it is still focused on hate speech detection. By attempting to replicate a hate speech detection experiment performed on an existing Twitter corpus annotated for hate speech, we highlight some issues that arise from doing research in the field of hate speech, which is essentially still in its infancy. We take a critical look at the training corpus in order to understand its biases, while also using it to venture beyond hate speech detection and investigate whether it can be used to shed light on other facets of research, such as popularity of hate tweets.
Keywords: hate speech, machine learning, feature analysis, corpus bias, ephemeral data, replicability
Examining a hate speech corpus
for hate speech detection and popularity prediction
|Filip Klubička, Raquel Fernández|
|School of Computing, Dublin Institute of Technology|
|Institute of Logic, Language and Computation, University of Amsterdam|
The Internet, likely one of humanity’s greatest inventions, facilitates the sharing of ideas and knowledge, as well as online discussion and user interaction. All these are positive features but, as with any tool, whether they are used in a positive or negative manner depends largely on the people that use them. Consequently, and especially when user anonymity is added to the mix, online discussion environments can become abusive, hateful and toxic. To help identify, study, and ultimately curb this problem, such negative environments and the language used within are being studied under the name hate speech.
Research on hate speech has become quite prominent in recent years, with dedicated workshops and conferences,111A few recent examples: https://europa.eu/newsroom/events/conference-online-hate-speech_en
http://likestiltnorden2017.regjeringen.no/language/en/nordic-hate-speech-conference/ and even being featured on LREC2018’s list of hot topics. However, hate speech research is still in its infancy. In part, this is due to the following challenges:
The term hate speech is difficult to define. ?) say that “hate speech lies in a complex nexus with freedom of expression, group rights, as well as concepts of dignity, liberty, and equality. For this reason, any objective definition (i.e., that can be easily implemented in a computer program) can be contested.” Generally, the current consensus among researchers seems to be that hate speech can be seen as a phenomenon encompassing issues such as: personal attacks, attacks on a specific group or minority, and abusive language targeting specific group characteristics (e.g., ethnicity, religion, gender, sexual orientation).
Creating resources for studying hate speech is far from trivial. Hate speech comprises a very small fraction of online content, and on most social platforms it is heavily moderated. For example, ?) report that in their corpus of comments on Yahoo! articles collected between April 2014 and April 2015, the percentage of abusive comments is around 3.4% on Finance articles and 10.7% on News. Since the phenomenon is elusive, researchers often use lists of offensive terms to collect datasets with the aim to increase the likelihood of catching instances of hate speech [Davidson et al., 2017, Waseem and Hovy, 2016]. This filtering process, however, has the risk of producing corpora with a variety of biases, which may go undetected.
Finally, hate speech is present in user-generated content that is not under the control of the researcher. Social media data is typically collected by public APIs that may lead to inconsistent results. For example, ?) find that the Twitter Search API yields a smaller dataset than the Stream API when using the same filtering parameters. Furthermore, users might delete their profiles or moderate their own questionable content themselves. Thus, datasets on which research experiments are performed are ephemeral, which makes replication of results very difficult.
In this paper, we focus on the latter two points. We consider a particular hate speech corpus – a Twitter corpus collected by ?), which has been gaining traction as a resource for training hate speech detection models [Waseem and Hovy, 2016, Gambäck and Utpal, 2017, Park and Fung, 2017] – and analyse it critically to better understand its usefulness as a hate speech resource. In particular, we make the following contributions:
We report the outcome of a reproduction experiment, where we attempt to replicate the results by ?) on hate speech detection using their Twitter corpus.
We use the corpus to study a novel aspect related to hate speech: the popularity of tweets containing hate speech. To this end, we develop models for the task of predicting whether a hate tweet will be interacted with and perform detailed feature analyses.
We perform a quantitative and qualitative analysis of the corpus to analyse its possible biases and assess the generality of the results obtained for the hate speech detection and popularity tasks.
2. Replication: hate speech detection results
We aim to replicate the results on hate speech detection by ?) using the hate speech Twitter corpus created by the authors.222https://github.com/zeerakw/hatespeech The dataset is a useful resource as it is one of few freely available corpora for hate speech research; it is manually annotated and distinguishes between two types of hate speech – sexism and racism – which allows for more nuanced insight and analysis. Additionally, as a Twitter corpus, it provides opportunity for any type of analysis and feature examination typical for Twitter corpora, such as user and tweet metadata, user interaction, etc.
2.1. Corpus in numbers
Here we provide just a brief quantitative overview of the corpus, whereas a more detailed qualitative analysis is presented in Section 4.. The original dataset contains 16907 annotated tweets. However, as is common practice with Twitter corpora, the corpus was only made available as a set of annotated tweet IDs, rather than the tweets themselves. To obtain the actual tweets and corresponding metadata, we used the Tweepy Twitter API wrapper.333http://tweepy.readthedocs.io/en/v3.5.0/ Given that the corpus was initially collected and annotated in 2016, there have been some changes in the availability of tweets by the time we extracted in in May 2017. Table 1 presents the distribution of annotations in the corpus in its original version and the version that was used for this paper. A tweet in the corpus can have three labels (None, Racism, Sexism). It is possible that a tweet has multiple labels, in the case that it contains both racism and sexism (this only happens in 8 tweets in the original dataset, so it is not a widespread phenomenon in this corpus.)
The dataset is quite unbalanced, but this is reflective of the unbalanced distribution of hate speech ‘in the wild’, and speaks to why it is so difficult to do research on hate speech in the first place: it is an elusive phenomenon. This, combined with the fact that users might delete their profiles or moderate their own questionable content themselves, makes available data scarce, and makes every Twitter corpus smaller over time, and consequently, less valuable and more prone to mistakes when attempting a replicative study.
2.2. Experimental setup
As with any replication study, our aim here is to mimic the original experimental setup as closely as possible, in hopes of obtaining same or comparable results. Unfortunately, this effort is already potentially hindered by the fact that the Twitter corpus has shrunk over time. However, the difference is not too large, and we expect it not to have a significant impact on the results.
A much more prominent obstacle is the lack of certain implementation details in the original paper that make reproduction difficult. At several points in the pipeline, we were left to our own devices and resort to making educated guesses as to what may have been done, due to the lack of comprehensive documentation. More specifically, there are two important aspects of the pipeline that present us with this problem: the algorithm and the features.
?) state that they use a logistic regression classifier for their hate speech prediction task. What is not mentioned is which implementation of the algorithm is used, how the model was fit to the data, whether the features were scaled, and whether any other additional parameters had been used.
Due to its popularity and accessibility, we opt for the Scikit-learn [Pedregosa et al., 2011] Python implementation of the logistic regression algorithm.444http://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html In addition, after fitting the model, we do not do additional scaling of the features when working with just n-grams (as these are already scaled when extracted), but we do scale our other features using the scaling function.555http://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.scale.html
?) explore several feature types: they employ n-gram features – specifically, they find that character n-grams of lengths up to 4 perform best – and in addition, they combine them with gender information, geographic location information and tweet length, finding that combining n-gram features with gender features yields slightly better results than just n-gram features do, while mixing in any of the other features results in slightly lower scores.
As a rule of thumb, we would attempt to replicate the best performing setup (character n-grams in combination with gender). However, this proved to be difficult, as user gender information is not provided by Twitter (hence it cannot be scraped from the Twitter API) and has not been made available by the authors along with their dataset. However, they do describe how they went about procuring the gender information for themselves (by performing semi-automatic, heuristics-based annotation), but only managed to annotate about 52% of the users. This, in combination with the fact that in the original experiment the F1 score improvement when gender is considered is minor (0.04 points) and not statistically significant, led us to focus our efforts on replicating only the experiments involving n-gram features.
However, extracting the n-gram features is also shown to be a nontrivial task, as the original paper does not state how the features are encoded: whether it is using a bag-of-n-grams approach, a frequency count approach, or a TF-IDF measure for each n-gram. We opt for TF-IDF because it is most informative, and just as easy to implement as the more basic approaches.666http://scikit-learn.org/stable/modules/generated/sklearn.feature_extraction.text.TfidfVectorizer.html
2.3. Evaluation and results
The original paper states the use of 10-fold cross-validation for model evaluation purposes, without specifying a particular implementation. For the sake of consistency, we again opt for the Scikit-learn implementation.777http://scikit-learn.org/stable/modules/generated/sklearn.model_selection.cross_val_score.html
We compare the results of our setup to the results of the original experiment. In addition, we also compare evaluations of a system trained on various other features (which we will describe in Section 3.) extracted from the tweets and their metadata. The results are presented in Table 2.
Examining the table reveals that our best attempt at replicating the original experiment, with logistic regression trained only on character n-grams, yields an F1-score that is 0.03 points lower than the original. Such a drop is to be expected, considering that our version of the dataset was smaller and that we had to fill in some gaps in the procedure ourselves, likely resulting in slight procedural mismatches. However, the drop is not large, and might indicate a stable, consistent result.
When looking at the performance of classifiers trained on features extracted from tweets and their metadata, they significantly underperform, with a 6 point drop compared to our replicated experiment, and a 9 point drop compared to the original results. This adds a strong confirmation of an observation made in the original study, namely that n-gram features are the most predictive compared to any other types of features.
3. New experiment: popularity prediction
To date, most research on hate speech within the NLP community has been done in the area of automatic detection using a variety of techniques, from lists of prominent keywords [Warner and Hirschberg, 2012] to regression classifiers as seen in the previous section [Nobata et al., 2016, Waseem and Hovy, 2016], naive Bayes, decision trees, random forests, and linear SVMs [Davidson et al., 2017], as well as deep learning models with convolutional neural networks [Gambäck and Utpal, 2017, Park and Fung, 2017].
Our intent in this section is to explore hate speech beyond just detection, using the Twitter corpus by ?). Given that Twitter is a platform that enables sharing ideas, and given that extreme ideas have a tendency to intensely spread through social networks [Brady et al., 2017], our question is: how does the fact that a tweet is a hate tweet affect its popularity?
3.1. Related work
To our knowledge there has not been any work relating tweet popularity with hate speech. However, there is a significant body of work dealing with tweet popularity modeling and prediction. Many papers explore features that lead to retweeting. ?) perform an extensive analysis of features that affect retweetability, singling out two groups of features: content and contextual features. Similarly, ?) train a model to predict the number of retweets using two types of features: user features and tweet features. They also compute information gain scores for their features and build a feature-weighted model. They compare the performance of two algorithms: logistic regression and SVM and find that SVM works better, yielding an F-score of 0.71. In addition, some of the related work also relies on temporal features: ?) predict the total number of retweets a given amount of time after posting, using a Bayesian model based on features of early retweet times and follower graphs. Similarly, ?) predict the number of retweets, using binary and multi-class classifiers. They use a more varied set of features, and aside from temporal features, they use content, topical and graph features, as well as user metadata.
We do not have temporal data at our disposal, nor are we at this stage interested in predicting the exact number of retweets at any given point. We are more concerned with investigating how hate speech comes into play regarding tweet popularity, if at all.
3.2. Popularity analysis
As surveyed above, most of the related work on tweet popularity focuses solely on retweets as indicators of popularity. However, while this is probably the clearest indicator, users can interact with tweets in a number of other ways. For this reason, in the present work we also consider other potential measures of popularity; namely, number of tweet replies and number of ‘likes’ (formerly called ‘favorites’).
Given these distributions, we opt for framing the problem as a binary classification task: we wish to determine whether a tweet receives a reaction (retweet, like, response) at least once, or not at all.
But before we go into prediction, we wish to investigate whether there is a significant difference between hate speech and non-hate speech tweets regarding the number of times a tweet was liked/retweeted/replied to. Thus, to determine whether these differences are statistically significant, we employ the chi-squared () statistical significance test.
When examining likes and replies, the test yields -values of <0.0001, meaning that tweets containing hate speech in the corpus are both liked and replied to significantly less than non-hate speech tweets are. In other words, if a tweet contains hate speech, it is less likely to be liked and replied to. However, when examining the difference in the number of retweets, the -value comes out as 0.5967. This means that we cannot dismiss the null hypothesis, or rather, that whether a tweet contains hate speech or not, does not impact its retweetability either way.
3.3. Popularity prediction
We use an large set of features inspired by related work [Waseem and Hovy, 2016, Sutton et al., 2015, Suh et al., 2010, Zaman et al., 2013, Hong et al., 2011, Zhang et al., 2012, Cheng et al., 2014, Ma et al., 2013, Zhao et al., 2015]. We divide our features into three groups: Tweet features (metadata about the the tweet itself), user features (metadata about the author of a tweet) and content features (features derived from the content of the tweet), with the largest number of features falling into the latter group. The features are listed in Table 5.
|Tweet features||User features|
Models and results.
We train a logistic regression classifier, as well as a linear SVM classifier to compare their performances. We also train separate models for likes and for retweets. One pair of models was trained on the whole corpus, and two additional pairs of classifiers were trained on just the hate speech portion and non-hate speech portion of the corpus respectively.
We tested all models using 10-fold cross validation, holding out 10% of the sample for evaluation to help prevent overfitting. All modeling and evaluation was performed using Scikit-learn [Pedregosa et al., 2011]. The evaluation results are presented in Table 6. We also make our feature dataset, and our training and evaluation scripts available to the community for transparency and reproduction purposes.888The dataset is comprised of anonymized tweet IDs with extracted content features. Link to GitHub repository: https://github.com/GreenParachute/hate-speech-popularity.
Interestingly, our classifiers are consistently better at predicting retweets than likes. Given that they are trained on the same features, this indicates that the nature of these two activities is different, in spite of the fact that they intuitively seem very similar.
Furthermore, it seems that the linear regression model seems to perform slightly better overall than the SVM model on both prediction tasks (likes and retweets).
In order to investigate which features are most informative for the task, we perform feature ablation according to our feature groups. Some notable results show that removing author metadata from the feature set reduces the performance of the model.999As our analysis in Section 4. will reveal, this seems a consequence of a strong bias towards a handful of overproductive authors in the corpus. However, the biggest takeaway for now is the feature’s impact on the model. Our SVM model’s average accuracy drops by 0.04 points if the feature is omitted from the feature set, whereas omitting many of the other features decreases performance scores by 0.02 points at most, if at all.
Inspired by ?), we also calculate information gain for all features. The top most informative features for each task (predicting whether a tweet will be liked/retweeted) and for each setup (full dataset/non-hate dataset/hate dataset) according to the information gain (IG) measure are presented in Tables 7, 8 and 9.
Given the context of this paper and the nature of the corpus, it is interesting to note that the feature does not appear anywhere near the top of the IG rankings, indicating that it is not very informative in regards to predicting whether a tweet will be liked or retweeted.
On a broader note, although the feature lists are more or less similar across the different dataset splits, there is a marked difference between the retweeting and liking lists, in each split. Features that are very informative for retweeting, but not for liking, are whether the tweet contains uppercase tokens, and, most notably, whether the tweet is a reply. This is in line with our findings in the feature ablation study, confirming that there is a strong link between the possibility of retweeting and whether or not the tweet in question is a reply. Our interpretation of this discrepancy is that original, stand-alone ideas (tweets) might be more likely to be picked up and passed on (retweeted), than a turn in a twitter conversation thread would be. In addition, these overall IG measurements also indicate that there is an inherent qualitative difference between the acts of liking and retweeting.
4. Corpus analysis
As the field of hate speech research is yet to mature, with disagreement about what exactly the phenomenon entails [Waseem et al., 2017] and without a unified annotation framework [Fišer et al., 2017], it is warranted to look at the data and examples in more detail, with considerations for potential shortcomings. In Section 2., we pointed out the ephemeral nature of the corpus by ?), common to all Twitter datasets. In this section, we analyse other characteristics of the corpus related to the challenges of data collection for hate speech analysis we mentioned in the Introduction (point 2), which can result in undesirable biases.
Given the small fraction of online content comprised of hate speech, collecting a significant amount of examples is an extremely difficult task. At present, it is not feasible to collect a large sample of tweets and then manually label them as hate or non hate, as the fraction of instances labeled with the positive class will be negligible. The only way to model the phenomenon is to target tweets already likely to contain hate speech.
Driven by this rationale, the authors of the corpus have obtained their dataset by performing an initial manual search of common slurs and terms used pertaining to religious, sexual, gender, and ethnic minorities. The full list of terms they queried for is not very long: MKR, asian drive, femi-nazi, immigrant, nigger, sjw, WomenAgainstFeminism, blameonenotall, islam terrorism, notallmen, victimcard, victim card, arab terror, gamergate, jsil, racecard, race card. In the results obtained from these queries, they identified frequently occurring terms in tweets that contain hate speech and references to specific entities (such as MKR, addressed further below). In addition to this, they identified a small number of prolific users from these searches.
This manner of tweet collection allowed the authors to obtain quite a considerable amount of data. However, this approach to data collection inevitably introduces many biases into the dataset, as will be demonstrated further in this section.
Qualitative observations on tweet content.
According to the annotation guidelines devised by ?) for the purpose of annotating this corpus, a tweet is tagged as offensive if it: (1) uses a sexist or racial slur, (2) attacks a minority, (3) seeks to silence a minority, (4) criticizes a minority (without a well founded argument), (5) promotes, but does not directly use, hate speech or violent crime, (6) criticizes a minority and uses a straw man argument, (7) blatantly misrepresents truth or seeks to distort views on a minority with unfounded claims, (8) shows support of problematic hashtags (e.g. #BanIslam, #whoriental, #whitegenocide), (9) negatively stereotypes a minority, (10) defends xenophobia or sexism, (11) the tweet is ambiguous (at best); and contains a screen name that is offensive as per the previous criteria; and is on a topic that satisfies any of the above criteria.
Though at first glance specific and detailed, these criteria are quite broad and open to interpretation. This was likely done to cover as many hate speech examples as possible – a thankless task, as hate speech data is scarce to begin with. However, due to this same breadth, the corpus contains some potential false positives. The most jarring example of this being that, if a user quotes a tweet containing hate speech (by prepending the quoted text with “RT”), the quoter’s tweet is tagged as hate speech. Certainly, the user could have quoted the original tweet in support of its message, and even if not, one could argue that they do perpetuate the original hateful message by quoting it. On the other hand, it is just as likely that the user is quoting the tweet not to make an endorsement, but a neutral response. It is even more likely that the user’s response is an instance of counterspeech — interaction used to challenge hate speech [Wright et al., 2017]. Manual inspection shows that there are instances of both such phenomena in the corpus, yet all those tweets are tagged as hate speech. In fact, 30% of hate speech tweets in the corpus contain the token ’RT’, indicating they are actually retweets. This could pose a problem further down the line when extrapolating information about hate speech users. Addressing this issue would at the very least require going through tweets with quotes and relabeling them, if not altogether rethinking the annotation guidelines, or rather, being more mindful of the semantics at play during annotation.
In spite of the broad guidelines, however, it seems that the actual hate speech examples end up falling on quite a narrow spectrum. Even though the tweets were semi-automatically picked based on a wide variety of keywords likely to identify hate speech, the tag ‘racism’ is in fact used as an umbrella term to label not only hate based on race/ethnicity, but also religion, specifically Islam. Indeed, the majority of the tweets tagged as racist are, in fact, islamophobic, and primarily written by a user with an anti-Islam handle (as per guideline 11). Though it is stated in the original paper which seed words were used to collect the data (which included both racist and islamophobic terms), it is undeniable that the most frequent words in the racist portion of the corpus refer to islamophobia (which is also explicitly stated by the authors themselves). This is not wrong, of course, but it begs the question of why the authors did not choose a more specific descriptor for the category, especially given that the term ‘racism’ typically sparks different connotations, ones that, in this case, do not accurately reflect the content of the actual data.
When it comes to sexist tweets, they are somewhat more varied than those annotated as racist. However, they do contain a similar type of bias: 13.6% of the tweets tagged as sexist contain the hashtag and/or handle MKR/MyKitchenRules. My Kitchen Rules is an Australian competitive cooking game show which is viewed less for the cooking and skill side of the show than for the gossip and conflict which certain contestants are encouraged to cause.101010https://en.wikipedia.org/wiki/My_Kitchen_Rules#Criticism It seems to be a popular discussion topic among fans of the show on Twitter, and apparently prompts users to make sexist remarks regarding the female contestants. There is nothing inherently problematic about this being included in a corpus of hate speech, but it cannot be disregarded that more than a tenth of the data on sexism is constrained to an extremely specific topic domain, which might not make for the most representative example of sexism on Twitter.
Distribution of users vs. tweet content
Another interesting dimension of the corpus that we observe is the distribution of users in relation to the hate speech annotations – an aspect that could be important for our analysis of popularity presented in Section 3.
There are 1858 unique user IDs in the corpus. Thus many of the 16907 tweets were written by the same people. As a simplistic approximation, we can (very tentatively) label every user that is the author of at least one tweet containing hate speech as a hate user; and users that, in the given dataset, have not produced any tweets containing hate speech we label as non-hate users. Of course, this does have certain drawbacks, as we cannot know that a user does not produce hate speech outside the sample we are working with, but it does provide at least an approximation of a user’s production of hate tweets in the sample. Using this heuristic, the distribution of users in the corpus in regards to whether they produce hate speech or not is presented in Table 10.
|Racist and sexist||3|
A really striking discrepancy immediately jumps out when looking at Table 10: there is a total of 5 users responsible for the 1942 racist tweets, as opposed to the 523 users responsible for the 3126 sexist tweets. Assuming normal distribution (which is certainly the wrong assumption), on average there are 388 racist tweets per racist user, while there is an average of 6 sexist tweets per sexist user. The actual distribution, however, is extremely skewed – the bulk of all the hate speech data is distributed between three users: one user who produced 1927 tweets tagged as racist, and two users who respectively produced 1320 and 964 tweets tagged as sexist. This is illustrated in Figure 1.
Such a distribution renders any attempt at generalization or modeling of racist tweets moot, as the sample cannot be called representative of racism as such, but only of the Twitter production of these 5 users.111111However, the data might still be useful when looked at in bulk with sexism, as it might reinforce the similarities they both share stemming from the fact that they are types of hate speech. Similarly, the fact that most of the tweets tagged as sexist belong to the same two users considerably skews this subset of the data.
All of these points deserve due consideration. The imbalances with respect to distribution of users were certainly considered while we worked with the data. In an attempt to reduce them, we did not distinguish between racist and sexist tweets in our analysis in both Sections 2. and 3. (even though we were tempted to do so), but rather treated them all as simply hate speech tweets. Additionally, it is possible that the insights and biases presented in this section might even call into question the relevance of the findings from Section 3., as the popularity modeled there is likely reflecting the popularity of the particular Twitter users in the corpus rather than of hate speech tweets as such.
This paper has provided an overview of several research directions involving hate speech:
A critical look at a publicly available hate speech dataset.
An attempt at replicating and confirming already established hate speech detection findings.
Pushing the research space in a new direction: popularity prediction.
Overall, we analyzed a currently popular hate speech dataset, pointed out considerations that have to be made while working such data, and observed that it is biased on several levels. This does not render it useless, but it is important to keep these biases in mind while using this resource and while drawing any sort of conclusions from the data.
As far as replicability goes, the resource does allow one to model hate speech (as biased as it may be), but not without a certain degree of difficulty. We achieve system evaluation scores of 0.71 in terms of F1 score, which is slightly lower than the original results of 0.74 F1 score on the same setup. The differences and gaps in implementation showcase a common trend in scientific publishing - the general problem of reproducing results due to publications not providing sufficient information to make the experiments they describe replicable without involving guessing games. And only when attempting to reproduce a study can one truly realize how much detail is so easily omitted or overlooked, simply due to lack of awareness.
When it comes to popularity prediction, we determine that hate speech negatively impacts the likelihood of likes and replies, but does not affect likelihood of retweets. However, training only on the hate speech portion of the data does seem to boost our model’s performance in retweet prediction. These findings, as well as the evaluation scores and feature analyses, are only the first stepping stone in a long line of future work that can be done to better understand the impact of hate speech on social media and how it spreads.
Possibilities include employing social graph mining and network analysis, perhaps using user centrality measures as features in both hate speech and popularity prediction tasks. In addition, reframing the task as not just a binary prediction task, but rather fitting a regression model to predict the exact number of likes, retweets and replies, would certainly be preferable and more informative, and could lead to a better understanding of how hate speech behaves on Twitter.
What is clear is that hate speech is a very nuanced phenomenon and we are far from knowing everything there is to know about it. Resources are scarce and far from perfect, and much more work and careful consideration are needed, as well as much cleaning, fine-tuning, discussion and agreement on what hate speech even is, if we are to build better resources and successfully model and predict hate speech, or any of its aspects.
This research was co-funded by the Erasmus+ programme of the European Union and conducted while the first author was visiting the ILLC in Amsterdam. In addition, the research was supported by the ADAPT Centre for Digital Content Technology which is funded under the SFI Research Centres Programme (Grant 13/RC/2106) and is co-funded under the European Regional Development Fund.
6. Bibliographical References
- Brady et al., 2017 Brady, W. J., Wills, J. A., Jost, J. T., Tucker, J. A., and Van Bavel, J. J. (2017). Emotion shapes the diffusion of moralized content in social networks. Proceedings of the National Academy of Sciences, 114(28):7313–7318.
- Cheng et al., 2014 Cheng, J., Adamic, L. A., Dow, P. A., Kleinberg, J. M., and Leskovec, J. (2014). Can cascades be predicted? CoRR, abs/1403.4608.
- Davidson et al., 2017 Davidson, T., Warmsley, D., Macy, M., and Weber, I. (2017). Automated hate speech detection and the problem of offensive language. In International AAAI Conference on Web and Social Media.
- Fišer et al., 2017 Fišer, D., Ljubešić, N., and Erjavec, T. (2017). Legal framework, dataset and annotation schema for socially unacceptable online discourse practices in slovene. In Proceedings of ALW1: 1st Workshop on Abusive Language Online.
- Gambäck and Utpal, 2017 Gambäck, B. and Utpal, K. S. (2017). Using convolutional neural networks to classify hate-speech. In Proceedings of ALW1: 1st Workshop on Abusive Language Online.
- González-Bailón et al., 2014 González-Bailón, S., Wang, N., Rivero, A., Borge-Holthoefer, J., and Moreno, Y. (2014). Assessing the bias in samples of large online networks. Social Networks, 38:16–27.
- Hong et al., 2011 Hong, L., Dan, O., and Davison, B. D. (2011). Predicting popular messages in twitter. In Proceedings of the 20th international conference companion on World wide web, pages 57–58. ACM.
- Ma et al., 2013 Ma, Z., Sun, A., and Cong, G. (2013). On predicting the popularity of newly emerging hashtags in twitter. Journal of the American Society for Information Science and Technology, 64, 07.
- Nobata et al., 2016 Nobata, C., Tetreault, J., Thomas, A., Mehdad, Y., and Chang, Y. (2016). Abusive language detection in online user content. In Proceedings of the 25th International Conference on World Wide Web, WWW ’16, pages 145–153, Republic and Canton of Geneva, Switzerland. International World Wide Web Conferences Steering Committee.
- Park and Fung, 2017 Park, J. H. and Fung, P. (2017). One-step and two-step classification for abusive language detection on twitter. In Proceedings of ALW1: 1st Workshop on Abusive Language Online.
- Pedregosa et al., 2011 Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., and Duchesnay, E. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830.
- Silva et al., 2016 Silva, L. A., Mondal, M., Correa, D., Benevenuto, F., and Weber, I. (2016). Analyzing the targets of hate in online social media. CoRR, abs/1603.07709.
- Suh et al., 2010 Suh, B., Hong, L., Pirolli, P., and Chi, E. H. (2010). Want to be retweeted? large scale analytics on factors impacting retweet in twitter network. In Social computing (socialcom), 2010 ieee second international conference on, pages 177–184. IEEE.
- Sutton et al., 2015 Sutton, J., Gibson, C. B., Spiro, E. S., League, C., Fitzhugh, S. M., and Butts, C. T. (2015). What it takes to get passed on: message content, style, and structure as predictors of retransmission in the boston marathon bombing response. PLoS one, 10(8):e0134452.
- Warner and Hirschberg, 2012 Warner, W. and Hirschberg, J. (2012). Detecting hate speech on the world wide web. In Proceedings of the Second Workshop on Language in Social Media, LSM ’12, pages 19–26, Stroudsburg, PA, USA. Association for Computational Linguistics.
- Waseem and Hovy, 2016 Waseem, Z. and Hovy, D. (2016). Hateful symbols or hateful people? predictive features for hate speech detection on twitter. In Proceedings of the NAACL Student Research Workshop, pages 88–93, San Diego, California, June. Association for Computational Linguistics.
- Waseem et al., 2017 Waseem, Z., Davidson, T., Warmsley, D., and Weber, I. (2017). Understanding abuse: A typology of abusive language detection subtasks. In Proceedings of ALW1: 1st Workshop on Abusive Language Online.
- Wright et al., 2017 Wright, L., Ruths, D., Dillon, K. P., Saleem, H. M., and Benesch, S. (2017). Vectors for counterspeech on twitter. In Proceedings of ALW1: 1st Workshop on Abusive Language Online.
- Zaman et al., 2013 Zaman, T., Fox, E. B., and Bradlow, E. T. (2013). A bayesian approach for predicting the popularity of tweets. CoRR, abs/1304.6777.
- Zhang et al., 2012 Zhang, Y., Xu, Z., and Yang, Q. (2012). Predicting popularity of messages in twitter using a feature-weighted model. International Journal of Advanced Intelligence.
- Zhao et al., 2015 Zhao, Q., Erdogdu, M. A., He, H. Y., Rajaraman, A., and Leskovec, J. (2015). SEISMIC: A self-exciting point process model for predicting tweet popularity. CoRR, abs/1506.02594.