Exploration of How Hate Is Framed on Social Media
Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Abstract
Understanding how hate is framed in multimodal social media content is crucial for developing interpretable and robust hate detection systems. We present the MM-HateFrames Dataset, a large-scale resource encoding 2,298 Hate Frames (HFs) and their corresponding rationales discovered from two benchmark datasets—Hateful Memes and MMHS150K—comprising over 11K+ social media multimodal posts. This allowed us to explore several generative and non-generative methods to automatically discover the way hate is framed when relying on MM-HateFrames, including clustering-based methods and large multimodal models (LMMs) under zero-shot and few-shot settings. Experimental evaluations show that few-shot LMMs prompting generates the most coherent and sound frame articulations. The MM-HateFrames Dataset provides a valuable foundation for future research in hate speech understanding, frame articulation, and explainable multimodal NLP, enabling models to interpret not only whether content is hateful but also how hate is conceptually framed.