May 12, 2020
We’ve built and are now sharing a dataset designed specifically to help AI researchers develop new systems to identify multimodal hate speech. This content combines different modalities, such as text and images, making it difficult for machines to understand.
The Hateful Memes dataset contains 10,000+ new multimodal examples created by Facebook AI. We licensed images from Getty Images so that researchers can use the dataset to support their work. We are also releasing the code for baseline-trained models.
We are also launching the Hateful Memes Challenge, a first-of-its-kind online competition hosted by DrivenData with a $100,000 total prize pool. The challenge has been accepted as part of the NeurIPS 2020 competition track.
In order for AI to become a more effective tool for detecting hate speech, it must be able to understand content the way people do: holistically. When viewing a meme, for example, we don’t think about the words and photo independently of each other; we understand the combined meaning together. This is extremely challenging for machines, however, because it means they can’t just analyze the text and the image separately. They must combine these different modalities and understand how the meaning changes when they are presented together. To catalyze research in this area, Facebook AI has created a dataset to help build systems that better understand multimodal hate speech. Today, we are releasing this Hateful Memes dataset to the broader research community and launching an associated competition, hosted by DrivenData with a $100,000 prize pool.
The challenges of harmful content affect the entire tech industry and society at large. As with our work on initiatives like the Deepfake Detection Challenge and the Reproducibility Challenge, Facebook AI believes the best solutions will come from open collaboration by experts across the AI community.
We continue to make progress in improving our AI systems to detect hate speech and other harmful content on our platforms, and we believe the Hateful Memes project will enable Facebook and others to do more to keep people safe.
The Hateful Memes dataset consists of more than 10,000 newly created examples of multimodal content. The memes were selected in such a way that strictly unimodal classifiers would struggle to classify them correctly (as illustrated in the examples below). We also designed the dataset specifically to overcome common challenges in AI research, such as the lack of examples to help machines learn to avoid false positives. It covers a wide variety of both the types of attacks and the groups and categories targeted. (More information on the dataset is available in this paper.)
We started with real examples of hateful memes that were shared online. We then built the dataset to address the needs of the AI research community.
To provide researchers with a dataset with clear licensing terms, we licensed assets from Getty Images. We worked with trained third-party annotators to create new memes similar to existing ones that had been shared on social media sites. The annotators used Getty Images’ collection of stock images to replace the original visuals while still preserving the semantic content. For example, if the original meme had a photo of a desert, we picked a similar desert photo from Getty. If no suitable replacement image could be found, the meme was discarded. We’d like to thank Getty for its help and partnership.
Annotators reviewed each meme to make sure it required multimodal understanding. As with the merely mean examples shown in this blog post, considering just the text or image in isolation would mean missing the meme’s true meaning.
To make sure the classification decisions are actionable, we created the examples in the Hateful Memes dataset using a clear and specific definition of hate speech:
“A direct or indirect attack on people based on characteristics, including ethnicity, race, nationality, immigration status, religion, caste, sex, gender identity, sexual orientation, and disability or disease. We define attack as violent or dehumanizing (comparing people to non-human things, e.g., animals) speech, statements of inferiority, and calls for exclusion or segregation. Mocking hate crime is also considered hate speech.”
Specially trained annotators classified the meme examples using this definition, which mirrors Facebook’s Community Standards. This in turn will help ensure that systems trained with Hateful Memes data will work effectively in real-life production applications.
Our examples also cover a wide variety of protected categories (such as religion, gender, and sexual orientation) and types of attacks (such as inciting violence or portraying types of people as criminals or terrorists). The distribution in the dataset reflects the real-world distribution found in the original examples.
The dataset also contains multimodal memes that are similar to hateful examples but are actually harmless. These examples, known as benign confounders, will help researchers address potential biases in classification systems and build systems that avoid false positives.
Multimodal content, such as the memes in Hateful Memes dataset, is difficult to classify with machine learning algorithms, as decisions are often more subtle than in unimodal cases and require real-life context and common sense.
To address this challenge, the research community is focused on building tools that take the different modalities present in a particular piece of content and then fuse them early in the classification process. This approach enables the system to analyze the different modalities together, like people do.
This approach contrasts with late-fusion systems, which are easier to build but less effective at understanding complex multimodal content.
Using the Hateful Memes dataset, we established baselines for the community using several well-known model architectures. We tested two unimodal systems and several multimodal systems. In late fusion, we trained the models separately and averaged their two scores during inference to get a prediction. In mid-fusion, we concatenated the BERT and ResNet-152 representations and fed them into a two-layer classifier (ConcatBERT). Finally, we used several BERT-derived architectures that fuse image and text understanding earlier in the process: a supervised multimodal bi-transformer model (MMBT), and state-of-the-art self-supervised (ViLBERT and Visual BERT). In addition to sharing the benchmark data, we are releasing the code for these models here through MMF, a multimodal framework, written in Facebook’s PyTorch deep learning library.
We’re pleased to be able to work with DrivenData to launch the Hateful Memes Challenge, which invites participants to create models trained on the Hateful Memes dataset. Both the challenge and the dataset launch today. More information for researchers is available here, and the accompanying paper is available here. For researchers considering attending NeurIPS in December, we’re also pleased to announce that the Hateful Memes Challenge has been accepted for the NeurIPS 2020 competition track. Journalists can request access to sample content in order to review the dataset and its characteristics. (The competition is subject to official rules, available here. See the competition website for eligibility, entry dates, submission requirements, evaluation metrics, and prizing.)
Like spam and other adversarial challenges, the problem of hateful content will continue to evolve. By providing a dataset expressly made to help researchers tackle this problem, along with a common benchmark and a community competition, we are confident that the Hateful Memes Challenge will spur faster progress across the industry in dealing with hateful content and help advance multimodal machine learning more broadly.
Applied Research Scientist