######################################################################## ### Crowdsourced Sentiment Dataset for ISF Journal Paper: ### ### Regional Sentiment Bias in Social Media Reporting During Crises ### ######################################################################## Sources: > Paper: https://doi.org/10.1007/s10796-018-9827-x > Data: http://dx.doi.org/10.5525/gla.researchdata.584 Information: This dataset contains sentiment labels generated using the crowdflower crowdsourcing platform for Twitter tweets collected during the Paris attacks and Berlin attacks events in English, French and German. Paris Attacks > English: Paris_English_All.csv > French: Paris_French_Full.csv Berlin Attacks > English: Berlin_EN_All.csv > German: Berlin_German_All.csv Please use the citation below if using this dataset: @Article{Smith2018, author="Smith, Karin Sim and McCreadie, Richard and Macdonald, Craig and Ounis, Iadh", title="Regional Sentiment Bias in Social Media Reporting During Crises", journal="Information Systems Frontiers", year="2018", month="Feb", day="28", abstract="Crisis events such as terrorist attacks are extensively commented upon on social media platforms such as Twitter. For this reason, social media content posted during emergency events is increasingly being used by news media and in social studies to characterize the public's reaction to those events. This is typically achieved by having journalists select `representative' tweets to show, or a classifier trained on prior human-annotated tweets is used to provide a sentiment/emotion breakdown for the event. However, social media users, journalists and annotators do not exist in isolation, they each have their own context and world view. In this paper, we ask the question, `to what extent do local and international biases affect the sentiments expressed on social media and the way that social media content is interpreted by annotators'. In particular, we perform a multi-lingual study spanning two events and three languages. We show that there are marked disparities between the emotions expressed by users in different languages for an event. For instance, during the 2016 Paris attack, there was 16{\%} more negative comments written in the English than written in French, even though the event originated on French soil. Furthermore, we observed that sentiment biases also affect annotators from those regions, which can negatively impact the accuracy of social media labelling efforts. This highlights the need to consider the sentiment biases of users in different countries, both when analysing events through the lens of social media, but also when using social media as a data source, and for training automatic classification models.", issn="1572-9419", doi="10.1007/s10796-018-9827-x", url="https://doi.org/10.1007/s10796-018-9827-x" }