Cardiff University | Prifysgol Caerdydd ORCA
Online Research @ Cardiff 
WelshClear Cookie - decide language by browser settings

Hate speech, machine classification and statistical modelling of information flows on Twitter: interpretation and communication for policy decision making

Burnap, Peter and Williams, Matthew Leighton 2014. Hate speech, machine classification and statistical modelling of information flows on Twitter: interpretation and communication for policy decision making. Presented at: Internet, Policy & Politics, Oxford, UK, 26 September 2014.

[img]
Preview
PDF - Accepted Post-Print Version
Download (388kB) | Preview

Abstract

In 2013, the murder of Drummer Lee Rigby in Woolwich, UK led to an extensive public social media reaction. Given the extreme terrorist motive and public nature of the actions it was feasible that the public response could include written expressions of hateful and antagonistic sentiment towards a particular race, ethnicity and religion, which can be interpreted as ‘hate speech’. This provided motivation to study the spread of hate speech on Twitter following such a widespread and emotive event. In this paper we present a supervised machine learning text classifier, trained and tested to distinguish between hateful and/or antagonistic responses with a focus on race, ethnicity or religion; and more general responses. We used human annotated data collected from Twitter in the immediate aftermath of Lee Rigby’s murder to train and test the classifier. As “Big Data” is a growing topic of study, and its use is in policy and decision making is being constantly debated at present, we discuss the use of supervised machine learning tools to classify a sample of “Big Data”, and how the results can be interpreted for use in policy and decision making. The results of the classifier are optimal using a combination of probabilistic, rule-based and spatial based classifiers with a voted ensemble meta-classifier. We achieve an overall F-measure of 0.95 using features derived from the content of each tweet, including syntactic dependencies between terms to recognise “othering” terms, incitement to respond with antagonistic action, and claims of well founded or justified discrimination against social groups. We then demonstrate how the results of the classifier can be robustly utilized in a statistical model used to forecast the likely spread of hate speech in a sample of Twitter data.

Item Type: Conference or Workshop Item (Paper)
Date Type: Completion
Status: Published
Schools: Computer Science & Informatics
Last Modified: 02 Oct 2014 13:01
URI: http://orca-mwe.cf.ac.uk/id/eprint/65227

Citation Data

Cited 1 time in Google Scholar. View in Google Scholar

Actions (repository staff only)

Edit Item Edit Item

Full Text Downloads from ORCA for this publication

Top Downloads of this item by Country

Monthly Full Text Downloads of this item

More statistics for this item...