Kicking Prejudice: Large Language Models for Racism Classification in Soccer Discourse on Social Media

SANTOS, Guto Leoni, DOS SANTOS, Vitor Gaboardi, KEARNS, Colm, SINCLAIR, Gary, BLACK, Jack, DOIDGE, Mark, FLETCHER, Thomas, KILVINGTON, Dan, ENDO, Patricia Takako, LISTON, Katie and LYNN, Theo (2024). Kicking Prejudice: Large Language Models for Racism Classification in Soccer Discourse on Social Media. In: GUIZZARDI, Giancarlo, FLAVIA, Santoro, HARALAMBOS, Mouratidis and PNINA, Soffer, (eds.) Advanced Information Systems Engineering: 36th International Conference, CAiSE 2024, Limassol, Cyprus, June 3–7, 2024, Proceedings. Lecture Notes in Computer Science, 14663 . Springer, Cham, 547-562. [Book Section]

Documents
33888:644176
[thumbnail of Black (2024c) Santos et al. (2024).pdf]
Preview
PDF
Black (2024c) Santos et al. (2024).pdf - Accepted Version
Available under License All rights reserved.

Download (587kB) | Preview
Abstract
In the dynamic space of Twitter, now called X, interpersonal racism surfaces when individuals from dominant racial groups engage in behaviours that diminish and harm individuals from other racial groups. It can be manifested in various forms, including pejorative name-calling, racial slurs, stereotyping, and microaggressions. The consequences of racist speech on social media are profound, perpetuating social division, reinforcing systemic inequalities, and undermining community cohesion. In the specific context of football discourse, instances of racism and hate crimes are well-documented. Regrettably, this issue has seamlessly migrated to the football discourse on social media platforms, especially Twitter. The debate on Internet freedom and social media moderation intensifies, balancing the right to freedom of expression against the imperative to protect individuals and groups from harm. In this paper, we address the challenge of detecting racism on Twitter in the context of football by using Large Language Models (LLMs). We fine-tuned different BERT-based model architectures to classify racist content in the Twitter discourse surrounding the UEFA European Football Championships. The study aims to contribute insights into the nuanced language of hate speech in soccer discussions on Twitter while underscoring the necessity for context-sensitive model training and evaluation. Additionally, Explainable Artificial Intelligence (XAI) techniques, specifically the Integrated Gradient method, are used to enhance transparency and interpretability in the decision-making processes of the LLMs, offering a comprehensive approach to mitigating racism and offensive language in online sports discourses.
More Information
Statistics

Downloads

Downloads per month over past year

View more statistics

Metrics

Altmetric Badge

Dimensions Badge

Share
Add to AnyAdd to TwitterAdd to FacebookAdd to LinkedinAdd to PinterestAdd to Email

Actions (login required)

View Item View Item