Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhance Multimodal Model Performance with Data Augmentation: Facebook Hateful Meme Challenge Solution (2105.13132v2)

Published 25 May 2021 in cs.LG and cs.CY

Abstract: Hateful content detection is one of the areas where deep learning can and should make a significant difference. The Hateful Memes Challenge from Facebook helps fulfill such potential by challenging the contestants to detect hateful speech in multi-modal memes using deep learning algorithms. In this paper, we utilize multi-modal, pre-trained models VilBERT and Visual BERT. We improved models' performance by adding training datasets generated from data augmentation. Enlarging the training data set helped us get a more than 2% boost in terms of AUROC with the Visual BERT model. Our approach achieved 0.7439 AUROC along with an accuracy of 0.7037 on the challenge's test set, which revealed remarkable progress.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yang Li (1142 papers)
  2. Zinc Zhang (1 paper)
  3. Hutchin Huang (1 paper)
Citations (1)

Summary

We haven't generated a summary for this paper yet.