Abstract
Detecting inappropriate language in online platforms is vital for maintaining a safe and respectful digital environment, especially in the context of hate speech prevention. However, defining what constitutes inappropriate language can be highly subjective and context-dependent, varying from person to person. This study presents the outcomes of a comprehensive examination of the subjectivity involved in assessing inappropriateness within conversational contexts. Different annotation methods, including expert annotation, crowd annotation, ChatGPT-generated annotation, and lexicon-based annotation, were applied to English Reddit conversations. The analysis revealed a high level of agreement across these annotation methods, with most disagreements arising from subjective interpretations of inappropriate language. This emphasizes the importance of implementing content moderation systems that not only recognize inappropriate content but also understand and adapt to diverse user perspectives and contexts. The study contributes to the evolving field of hate speech annotation by providing a detailed analysis of annotation differences in relation to the subjective task of judging inappropriate words in conversations.
Original language | English |
---|---|
Title of host publication | Proceedings of the Fourth Workshop on Threat, Aggression & Cyberbullying @ LREC-COLING-2024 |
Editors | Ritesh Kumar, Atul Kr. Ojha, Atul Kr. Ojha, Shervin Malmasi, Bharathi Raja Chakravarthi, Bornini Lahiri, Siddharth Singh, Shyam Ratan |
Publisher | ELRA and ICCL |
Pages | 96-104 |
Number of pages | 9 |
ISBN (Electronic) | 9782493814470 |
Publication status | Published - 2024 |
Event | 4th Workshop on Threat, Aggression and Cyberbullying, TRAC 2024 - Torino, Italy Duration: 20 May 2024 → … |
Conference
Conference | 4th Workshop on Threat, Aggression and Cyberbullying, TRAC 2024 |
---|---|
Country/Territory | Italy |
City | Torino |
Period | 20/05/24 → … |
Bibliographical note
Publisher Copyright:© 2024 ELRA Language Resource Association.
Keywords
- Inappropriate Language
- Online Content Moderation
- Subjectivity in Annotation