Hate speech relies heavily on cultural influences, leading to varying individual interpretations. For that reason, we propose a Semantic Componential Analysis (SCA) framework for a cross-cultural and cross-domain analysis of hate speech definitions. We create the first dataset of hate speech definitions encompassing 493 definitions from more than 100 cultures, drawn from five key domains: online dictionaries, academic research, Wikipedia, legal texts, and online platforms. By decomposing these definitions into semantic components,our analysis reveals significant variation across definitions, yet many domains borrow definitions from one another without taking into account the target culture. We conduct zero-shot model experiments using our proposed dataset, employing three popular open-sourced LLMs to understand the impact of different definitions on hate speech detection. Our findings indicate that LLMs are sensitive to definitions: responses for hate speech detection change according to the complexity of definitions used in the prompt.

Korre, A., Muti, A., Ruggeri, F., Barrón-Cedeño, A. (2025). Untangling Hate Speech Definitions: A Semantic Componential Analysis Across Cultures and Domains. Association for Computational Linguistics.

Untangling Hate Speech Definitions: A Semantic Componential Analysis Across Cultures and Domains

AiKaterini Korre
Primo
;
Arianna Muti
Secondo
;
Federico Ruggeri
Penultimo
;
Alberto Barrón-Cedeño
Ultimo
2025

Abstract

Hate speech relies heavily on cultural influences, leading to varying individual interpretations. For that reason, we propose a Semantic Componential Analysis (SCA) framework for a cross-cultural and cross-domain analysis of hate speech definitions. We create the first dataset of hate speech definitions encompassing 493 definitions from more than 100 cultures, drawn from five key domains: online dictionaries, academic research, Wikipedia, legal texts, and online platforms. By decomposing these definitions into semantic components,our analysis reveals significant variation across definitions, yet many domains borrow definitions from one another without taking into account the target culture. We conduct zero-shot model experiments using our proposed dataset, employing three popular open-sourced LLMs to understand the impact of different definitions on hate speech detection. Our findings indicate that LLMs are sensitive to definitions: responses for hate speech detection change according to the complexity of definitions used in the prompt.
2025
Findings of the Association for Computational Linguistics: NAACL 2025
3184
3198
Korre, A., Muti, A., Ruggeri, F., Barrón-Cedeño, A. (2025). Untangling Hate Speech Definitions: A Semantic Componential Analysis Across Cultures and Domains. Association for Computational Linguistics.
Korre, Aikaterini; Muti, Arianna; Ruggeri, Federico; Barrón-Cedeño, Alberto
File in questo prodotto:
File Dimensione Formato  
2025.findings-naacl.175.pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 304.76 kB
Formato Adobe PDF
304.76 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1005871
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact