Papers
arxiv:2005.04790

The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes

Published on May 10, 2020
Authors:
,
,
,
,
,

Abstract

A new challenge set for detecting hate speech in multimodal memes demonstrates the difficulty of the task, showing that state-of-the-art models significantly underperform compared to humans.

AI-generated summary

This work proposes a new challenge set for multimodal classification, focusing on detecting hate speech in multimodal memes. It is constructed such that unimodal models struggle and only multimodal models can succeed: difficult examples ("benign confounders") are added to the dataset to make it hard to rely on unimodal signals. The task requires subtle reasoning, yet is straightforward to evaluate as a binary classification problem. We provide baseline performance numbers for unimodal models, as well as for multimodal models with various degrees of sophistication. We find that state-of-the-art methods perform poorly compared to humans (64.73% vs. 84.7% accuracy), illustrating the difficulty of the task and highlighting the challenge that this important problem poses to the community.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2005.04790
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2005.04790 in a model README.md to link it from this page.

Datasets citing this paper 7

Browse 7 datasets citing this paper

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2005.04790 in a Space README.md to link it from this page.

Collections including this paper 1