This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
the rise of multimodal content as a way to express
ideas and emotions. As a result, a brand new type
of message was born: meme. A meme is typically
formed by an image and a short piece of text on
top of it, embedded as part of the image. Memes
are typically innocent and designed to look funny.
WARNING: This paper contains meme examples andwords that are offensive in nature.
Over time, memes started being used for harm-
ful purposes in the context of contemporary politi-
cal and socio-cultural events, targeting individuals,
groups, businesses, and society as a whole. At
the same time, their multimodal nature and often
camouflaged semantics make their analysis highly
challenging (Sabat et al., 2019).
Meme analysis. The proliferation of memes
online and their increasing importance have led
to a growing body of research on meme analy-
sis (Sharma et al., 2020a; Reis et al., 2020; Pra-
manick et al., 2021). It has also been shown that
off-the-shelf multimodal tools may be inadequate
to unfold the underlying semantics of a meme as
(i) memes are often context-dependent, (ii) the vi-
sual and the textual content are often uncorrelated,
and (iii) meme images are mostly morphed, and the
embedded text is sometimes hard to extract using
standard OCR tools (Bonheme and Grzes, 2020).
The dark side of memes. Recently, there has
been a lot of effort to explore the dark side of
memes, e.g., focusing on hate (Kiela et al., 2020)
and offensive (Suryawanshi et al., 2020) memes.
However, the harm a meme can cause can be much
broader. For instance, the meme1 in Figure 1c is
neither hateful nor offensive, but it is harmful to the
media shown on the top left (ABC, CNN, etc.), as it
compares them to China, suggesting that they adopt
strong censorship policies. In short, the scope of
harmful meme detection is much broader, and it
may encompass other aspects such as cyberbully-
ing, fake news, etc. Moreover, harmful memes have
a target (e.g., news organization such as ABC and
CNN in our previous example), which requires sep-
arate analysis not only to decipher their underlying
semantics, but also to help with the explainability
of the detection models.
1In order to avoid potential copyright issues, all memeswe show in this paper are our own recreation of existingmemes, using images with clear licenses.
2784
(a) [0] (b) [2,0] (c) [1,1] (d) [2,2] (e) [2,3]
Figure 1: Examples from our HarMeme dataset. The labels are in the format [Intensity, Target]. For
Intensity, {0, 1, 2} correspond to harmless, partially harmful, and very harmful, respectively. For Target,
{0, 1, 2, 3} correspond to individual, organization, community, and society, respectively. Examples 1b and 1c are
harmful, but neither hateful, nor offensive. Example 1d is both harmful and offensive. Source (a); Source (b);
Table 5: Performance for target identification of harm-
ful memes (†human accuracy on the test set).
6.2 Target Identification for Harmful Memes
Table 5 shows the results for the target identifi-
cation task. This is an imbalanced 4-class classi-
fication problem, and the majority class baseline
yields 46.60% accuracy. The unimodal models per-
form relatively better here, achieving 63%− 70%
accuracy; their F1 Macro and MMAE scores are
also above the majority class. However, the overall
performance of the unimodal models is poor. In-
corporating multimodal signals with fine-grained
fusion improves the results substantially, and ad-
vanced multimodal fusion techniques with multi-
modal pre-training perform much better than sim-
ple late fusion with unimodal pre-training. More-
over, V-BERT COCO outperforms ViLBERT CC
by 8% of F1 score and by nearly 0.3 of MMAE.
6.3 Human Evaluation
To understand how human subjects perceive these
tasks, we further hired a different set of experts
(not the annotators) to label the test set. We ob-
served 86% − 91% accuracy on average for both
tasks, which is much higher than V-BERT, the best-
performing model. This shows that their is a po-
tential for enriched multimodal models that better
understand the ingrained semantics of the memes.
(a) Very harmful meme (b) LIME output - image
(c) LIME output - text
(d) Harmless meme (e) LIME output - image
Figure 5: Example of explanation by LIME on both
visual and textual modalities and visualization of bias
in V-BERT for both tasks.
6.4 Side-by-side Diagnostics and Anecdotes
Since the HarMeme dataset was compiled of
memes related to COVID-19, we expected that
models with enriched contextual knowledge and
sophisticated technique would have superior per-
formance. Thus, to comprehend the interpretability
of V-BERT (the best model), we used LIME (Lo-
cally Interpretable Model-Agnostic Explanations)
(Ribeiro et al., 2016), a consistent model-agnostic
explainer to interpret the predictions.
2791
We chose two memes from the test set to analyze
the potential explanability of V-BERT. The first
meme, which is shown in Figure 5a, was manually
labeled as very harmful, and V-BERT successfully
classified it, with prediction probabilities of 0.651,
0.260, and 0.089 corresponding to the very harm-
ful, the partially harmful, and the harmless classes
respectively. Figure 5b highlights the most con-
tributing super-pixels to the very harmful (green)
class. As expected, the face of Donald Trump,
as highlighted by the green pixels, prominently
contributed to the prediction. Figure 5c demon-
strates the contribution of different meme words
to the model prediction. We can see that words
like CORONA and MASK have significant contribu-
tions to the very harmful class, thus supporting the
lexical analysis of HarMeme as shown in Table 3.
The second meme, which is shown in Figure 5d,
was manually labeled as harmless, but V-BERT in-
correctly predicted it to be very harmful. Figure 5e
shows that, similarly to the previous example, the
face of Donald Trump contributed to the prediction
of the model. We looked closer into our dataset,
and we found that it contained many memes with
the image of Donald Trump, and that the majority
of these memes fall under the very harmful category
and targeted and individual. Therefore, instead of
leaning the underlying semantics of one particular
meme, the model easily got biased by the presence
of Donald Trump’s image and blindly classified the
meme as very harmful.
7 Conclusion and Future Work
We presented HarMeme, the first large-scale
benchmark dataset, containing 3,544 memes, re-
lated to COVID-19, with annotations for degree of
harmfulness (very harmful, partially harmful, or
harmless), as well as for the target of the harm (an
individual, an organization, a community, or soci-
ety). The evaluation results using several unimodal
and multimodal models highlighted the importance
of modeling the multimodal signal (for both tasks)
—(i) detecting harmful memes and (ii) detecting
their targets—, and indicated the need for more
sophisticated methods. We also analyzed the best
model and identified its limitations.
In future work, we plan to design new multi-
modal models and to extend HarMeme with exam-
ples from other topics, as well as to other languages.
Alleviating the biases in the dataset and in the mod-
els are other important research directions.
Ethics and Broader Impact
User Privacy. Our dataset only includes memes
and it does not contain any user information.
Biases. Any biases found in the dataset are un-
intentional, and we do not intend to do harm to
any group or individual. We note that determining
whether a meme is harmful can be subjective, and
thus it is inevitable that there would be biases in
our gold-labeled data or in the label distribution.
We address these concerns by collecting examples
using general keywords about COVID-19, and also
by following a well-defined schema, which sets
explicit definitions during annotation. Our high
inter-annotator agreement makes us confident that
the assignment of the schema to the data is correct
most of the time.
Misuse Potential. We ask researchers to be
aware that our dataset can be maliciously used to
unfairly moderate memes based on biases that may
or may not be related to demographics and other in-
formation within the text. Intervention with human
moderation would be required in order to ensure
that this does not occur.
Intended Use. We present our dataset to encour-
age research in studying harmful memes on the
web. We distribute the dataset for research pur-
poses only, without a license for commercial use.
We believe that it represents a useful resource when
used in the appropriate manner.
Environmental Impact. Finally, we would also
like to warn that the use of large-scale Transform-
ers requires a lot of computations and the use
of GPUs/TPUs for training, which contributes to
global warming (Strubell et al., 2019). This is a bit
less of an issue in our case, as we do not train such
models from scratch; rather, we fine-tune them on
relatively small datasets. Moreover, running on a
CPU for inference, once the model has been fine-
tuned, is perfectly feasible, and CPUs contribute
much less to global warming.
Acknowledgments
The work was partially supported by the Wipro
research grant and the Infosys Centre for AI, IIIT
Delhi, India. It is also part of the Tanbih mega-
project, developed at the Qatar Computing Re-
search Institute, HBKU, which aims to limit the
impact of “fake news,” propaganda, and media bias
by making users aware of what they are reading.
2792
References
Firoj Alam, Stefano Cresci, Tanmoy Chakraborty, Fab-rizio Silvestri, Dimiter Dimitrov, Giovanni Da SanMartino, Shaden Shaar, Hamed Firooz, and PreslavNakov. 2021. A survey on multimodal disinforma-tion detection. arXiv 2103.12541.
Stefano Baccianella, Andrea Esuli, and Fabrizio Sebas-tiani. 2009. Evaluation Measures for Ordinal Re-gression. In Proceedings of the Ninth InternationalConference on Intelligent Systems Design and Appli-cations, ISDA ’09, pages 283–287, Pisa, Italy.
Victoria Bobicev and Marina Sokolova. 2017. Inter-annotator agreement in sentiment analysis: Machinelearning perspective. In Proceedings of the Inter-national Conference Recent Advances in NaturalLanguage Processing, RANLP ’17, pages 97–102,Varna, Bulgaria.
Lisa Bonheme and Marek Grzes. 2020. SESAM atSemEval-2020 task 8: Investigating the relationshipbetween image and text in sentiment analysis ofmemes. In Proceedings of the Fourteenth Workshopon Semantic Evaluation, SemEval ’20, pages 804–816, Barcelona, Spain.
Darren Crovitz and Clarice Moran. 2020. AnalyzingDisruptive Memes in an Age of International Inter-ference. The English Journal, 109(4):62–69.
Abhishek Das, Japsimar Singh Wahi, and Siyao Li.2020. Detecting hate speech in multi-modal memes.arXiv 2012.14891.
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, KaiLi, and Li Fei-Fei. 2009. ImageNet: A large-scalehierarchical image database. In Proceedings of theIEEE Conference on Computer Vision and PatternRecognition, CVPR ’09, pages 248–255, Miami, FL,USA.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, andKristina Toutanova. 2019. BERT: Pre-training ofdeep bidirectional transformers for language under-standing. In Proceedings of the 2019 Conference ofthe North American Chapter of the Association forComputational Linguistics: Human Language Tech-nologies, NAACL-HLT ’19, pages 4171–4186, Min-neapolis, MN, USA.
Dimitar Dimitrov, Bishr Bin Ali, Shaden Shaar, FirojAlam, Fabrizio Silvestri, Hamed Firooz, PreslavNakov, and Giovanni Da San Martino. 2021a. De-tecting propaganda techniques in memes. In Pro-ceedings of the Joint Conference of the 59th An-nual Meeting of the Association for ComputationalLinguistics and the 11th International Joint Con-ference on Natural Language Processing, ACL-IJCNLP ’21.
Dimitar Dimitrov, Bishr Bin Ali, Shaden Shaar, FirojAlam, Fabrizio Silvestri, Hamed Firooz, PreslavNakov, and Giovanni Da San Martino. 2021b.
SemEval-2021 Task 6: Detection of persuasion tech-niques in texts and images. In Proceedings of theInternational Workshop on Semantic Evaluation, Se-mEval ’21.
Marc J. Dupuis and Andrew Williams. 2019. Thespread of disinformation on the Web: An ex-amination of memes on social networking. InProceedings of the 2019 IEEE SmartWorld,Ubiquitous Intelligence Computing, AdvancedTrusted Computing, Scalable Computing Commu-nications, Cloud Big Data Computing, Internetof People and Smart City Innovation, Smart-World/SCALCOM/UIC/ATC/CBDCom/IOP/SCI ’19,pages 1412–1418.
Zhe Gan, Yen-Chun Chen, Linjie Li, Chen Zhu,Yu Cheng, and Jingjing Liu. 2020. Large-scale ad-versarial training for vision-and-language represen-tation learning. In Proceedings of the 34th Con-ference on Neural Information Processing Systems,NeurIPS ’20, pages 1–15, Vancouver, Canada.
Sunil Gundapu and Radhika Mamidi. 2020. Gunda-pusunil at SemEval-2020 task 8: Multimodal Mem-otion Analysis. In Proceedings of the FourteenthWorkshop on Semantic Evaluation, SemEval ’20,pages 1112–1119, Barcelona, Spain.
Yingmei Guo, Jinfa Huang, Yanlong Dong, andMingxing Xu. 2020. Guoym at SemEval-2020task 8: Ensemble-based classification of visuo-lingual metaphor in memes. In Proceedings of theFourteenth Workshop on Semantic Evaluation, Se-mEval ’20, pages 1120–1125, Barcelona, Spain.
Alan Hawkes. 1971. Spectra of some self-excitingand mutually exciting point processes. Biometrika,58:83–90.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, andJian Sun. 2016. Deep residual learning for imagerecognition. In Proceedings of the IEEE Confer-ence on Computer Vision and Pattern Recognition,CVPR ’16, pages 770–778, Las Vegas, NV, USA.
Anthony Hu and Seth Flaxman. 2018. Multimodalsentiment analysis to explore the structure of emo-tions. In Proceedings of the 24th ACM SIGKDD In-ternational Conference on Knowledge Discovery &Data Mining, ACMKDD ’18, pages 350–358, Lon-don, UK.
Gao Huang, Zhuang Liu, Laurens Van Der Maaten, andKilian Q. Weinberger. 2017. Densely connected con-volutional networks. In Proceedings of the 2017IEEE Conference on Computer Vision and PatternRecognition, CVPR ’17, pages 2261–2269, Hon-olulu, HI, USA.
Kimmo Karkkainen and Jungseock Joo. 2021. Fair-face: Face attribute dataset for balanced race, gen-der, and age for bias measurement and mitigation. InProceedings of the IEEE/CVF Winter Conference onApplications of Computer Vision, WACV ’21, pages1548–1558.
2793
Douwe Kiela, Suvrat Bhooshan, Hamed Firooz, andDavide Testuggine. 2019. Supervised multimodalbitransformers for classifying images and text.arXiv 1909.02950.
Douwe Kiela, Hamed Firooz, Aravind Mohan, VedanujGoswami, Amanpreet Singh, Pratik Ringshia, andDavide Testuggine. 2020. The hateful memes chal-lenge: Detecting hate speech in multimodal memes.In Advances in Neural Information Processing Sys-tems, volume 33 of NeurIPS ’20, pages 2611–2624,Vancouver, Canada.
Diederik P Kingma and Jimmy Ba. 2014. Adam: Amethod for stochastic optimization. In Proceedingsof the 3rd International Conference on LearningRepresentations, ICLR ’14, San Diego, CA, USA.
Jure Leskovec, Lars Backstrom, and Jon Kleinberg.2009. Meme-tracking and the dynamics of the newscycle. In Proceedings of the 15th ACM SIGKDDInternational Conference on Knowledge Discoveryand Data Mining, KDD ’09, page 497–506, NewYork, USA.
Liunian Harold Li, Mark Yatskar, Da Yin, Cho-JuiHsieh, and Kai-Wei Chang. 2019. VisualBERT: Asimple and performant baseline for vision and lan-guage. arxiv 1908.03557.
Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang,Xiaowei Hu, Lei Zhang, Lijuan Wang, HoudongHu, Li Dong, Furu Wei, Yejin Choi, and JianfengGao. 2020. Oscar: Object-semantics aligned pre-training for vision-language tasks. In Proceedingsof the 16th European Computer Vision Conference,ECCV ’20, pages 121–137, Glasgow, UK.
Tsung-Yi Lin, Michael Maire, Serge Belongie, JamesHays, Pietro Perona, Deva Ramanan, Piotr Dollar,and C Lawrence Zitnick. 2014. Microsoft COCO:Common Objects in Context. In Proceedingsof the European Conference on Computer Vision,ECCV ’14, pages 740–755, Zurich, Switzerland.
Chen Ling, Ihab AbuHilal, Jeremy Blackburn, Emil-iano De Cristofaro, Savvas Zannettou, and GianlucaStringhini. 2021. Dissecting the meme magic: Un-derstanding indicators of virality in image memes.In Proceedings of the 24th ACM Conference onComputer-Supported Cooperative Work and SocialComputing, CSCW ’21.
Phillip Lippe, Nithin Holla, Shantanu Chandra, San-thosh Rajamanickam, Georgios Antoniou, EkaterinaShutova, and Helen Yannakoudakis. 2020. A multi-modal framework for the detection of hateful memes.arXiv 2012.12871.
Zehao Liu, Emmanuel Osei-Brefo, Siyuan Chen, andHuizhi Liang. 2020. UoR at SemEval-2020 task8: Gaussian mixture modelling (GMM) based sam-pling approach for multi-modal memotion analysis.In Proceedings of the Fourteenth Workshop on Se-mantic Evaluation, SemEval ’20, pages 1201–1207,Barcelona, Spain.
Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee.2019. ViLBERT: Pretraining task-agnostic visi-olinguistic representations for vision-and-languagetasks. In Proceedings of the Conference on Neu-ral Information Processing Systems, NeurIPS ’19,pages 13–23, Vancouver, Canada.
Alexandros Mittos, Savvas Zannettou, Jeremy Black-burn, and Emiliano De Cristofaro. 2020. “And wewill fight for our race!” A measurement study ofgenetic testing conversations on Reddit and 4chan.In Proceedings of the Fourteenth International AAAIConference on Web and Social Media, ICWSM ’20,pages 452–463, Atlanta, GA, USA.
Terufumi Morishita, Gaku Morio, Shota Horiguchi,Hiroaki Ozaki, and Toshinori Miyoshi. 2020. Hi-tachi at SemEval-2020 task 8: Simple but effectivemodality ensemble for meme emotion recognition.In Proceedings of the Fourteenth Workshop on Se-mantic Evaluation, SemEval ’20, pages 1126–1134,Barcelona, Spain.
Gretel Liz De la Pena Sarracen, Paolo Rosso, and Anas-tasia Giachanou. 2020. PRHLT-UPV at SemEval-2020 task 8: Study of multimodal techniques formemes analysis. In Proceedings of the FourteenthWorkshop on Semantic Evaluation, SemEval ’20,pages 908–915, Barcelona, Spain.
Shraman Pramanick, Md Shad Akhtar, and TanmoyChakraborty. 2021. Exercise? I thought you said‘extra fries’: Leveraging sentence demarcations andmulti-hop attention for meme affect analysis. pro-ceedings of the Fifteenth International AAAI Confer-ence on Web and Social Media.
Julio C. S. Reis, Philipe Melo, Kiran Garimella, Jus-sara M. Almeida, Dean Eckles, and Fabrıcio Ben-evenuto. 2020. A dataset of fact-checked imagesshared on WhatsApp during the Brazilian and Indianelections. In Proceedings of the International AAAIConference on Web and Social Media, ICWSM ’20,pages 903–908.
Marco Tulio Ribeiro, Sameer Singh, and CarlosGuestrin. 2016. “Why should I trust you?” Explain-ing the predictions of any classifier. In Proceed-ings of the 22nd ACM SIGKDD International Con-ference on Knowledge Discovery and Data Mining,KDD ’16, pages 1135–1144, San Francisco, CA,USA.
Benet Oriol Sabat, Cristian Canton-Ferrer, and XavierGiro-i-Nieto. 2019. Hate speech in pixels: Detec-tion of offensive memes towards automatic modera-tion. arXiv 1910.02334.
Vlad Sandulescu. 2020. Detecting hateful memes us-ing a multimodal deep ensemble. arXiv 2012.13235.
2794
Chhavi Sharma, Deepesh Bhageria, William Scott,Srinivas PYKL, Amitava Das, Tanmoy Chakraborty,Viswanath Pulabaigari, and Bjorn Gamback. 2020a.SemEval-2020 task 8: Memotion analysis- thevisuo-lingual metaphor! In Proceedings of theFourteenth Workshop on Semantic Evaluation, Se-mEval ’20, pages 759–773, Barcelona, Spain.
Mayukh Sharma, Ilanthenral Kandasamy, and W.b. Vas-antha. 2020b. Memebusters at SemEval-2020 task8: Feature fusion model for sentiment analysis onmemes using transfer learning. In Proceedings ofthe Fourteenth Workshop on Semantic Evaluation,SemEval ’20, pages 1163–1171, Barcelona, Spain.
Piyush Sharma, Nan Ding, Sebastian Goodman, andRadu Soricut. 2018. Conceptual captions: Acleaned, hypernymed, image alt-text dataset for au-tomatic image captioning. In Proceedings of the56th Annual Meeting of the Association for Com-putational Linguistics, ACL ’18, pages 2556–2565,Melbourne, Australia.
Karen Simonyan and Andrew Zisserman. 2015. Verydeep convolutional networks for large-scale imagerecognition. In Proceedings of the 3rd InternationalConference on Learning Representations, ICLR ’15,San Diego, CA, USA.
Amanpreet Singh, Vedanuj Goswami, and Devi Parikh.2020. Are we pretraining it right? Digging deeperinto visio-linguistic pretraining. arXiv 2004.08744.
Emma Strubell, Ananya Ganesh, and Andrew McCal-lum. 2019. Energy and policy considerations fordeep learning in NLP. In Proceedings of the 57thAnnual Meeting of the Association for Computa-tional Linguistics, ACL ’19, pages 3645–3650, Flo-rence, Italy.
Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, LeweiLu, Furu Wei, and Jifeng Dai. 2020. VL-BERT:pre-training of generic visual-linguistic representa-tions. In Proceedings of the 8th International Con-ference on Learning Representations, ICLR ’20, Ad-dis Ababa, Ethiopia.
Shardul Suryawanshi, Bharathi Raja Chakravarthi, Mi-hael Arcan, and Paul Buitelaar. 2020. Multimodalmeme dataset (MultiOFF) for identifying offensivecontent in image and text. In Proceedings of the Sec-ond Workshop on Trolling, Aggression and Cyber-bullying, LREC-TRAC ’20, pages 32–41, Marseille,France.
Hao Tan and Mohit Bansal. 2019. LXMERT: Learningcross-modality encoder representations from trans-formers. In Proceedings of the 2019 Conference onEmpirical Methods in Natural Language Processingand the 9th International Joint Conference on Nat-ural Language Processing, EMNLP-IJCNLP ’17,pages 5100–5111, Hong Kong, China.
Chi Thang Duong, Remi Lebret, and Karl Aberer. 2017.Multimodal classification for analysing social media.arXiv 1708.02099.
Ashish Vaswani, Noam Shazeer, Niki Parmar, JakobUszkoreit, Llion Jones, Aidan N Gomez, LukaszKaiser, and Illia Polosukhin. 2017. Attention isall you need. In Proceedings of the Annual Con-ference on Neural Information Processing Systems,NeurIPS ’17, pages 5998–6008, Long Beach, CA,USA.
Yuping Wang, Fatemeh Tahmasbi, Jeremy Black-burn, Barry Bradlyn, Emiliano De Cristofaro,David Magerman, Savvas Zannettou, and GianlucaStringhini. 2020. Understanding the use of fauxtog-raphy on social media. arXiv 2009.11792.
Saining Xie, Ross Girshick, Piotr Dollar, Zhuowen Tu,and Kaiming He. 2017. Aggregated residual trans-formations for deep neural networks. In Proceed-ings of the IEEE Conference on Computer Visionand Pattern Recognition, CVPR ’17, pages 1492–1500, Honolulu, HI, USA.
Fei Yu, Jiji Tang, Weichong Yin, Yu Sun, Hao Tian,Hua Wu, and Haifeng Wang. 2021. ERNIE-ViL:Knowledge enhanced vision-language representa-tions through scene graph. In Proceedings of theThirty-Fifth AAAI Conference on Artificial Intelli-gence, AAAI ’21, pages 3208–3216.
Li Yuan, Jin Wang, and Xuejie Zhang. 2020. YNU-HPCC at SemEval-2020 task 8: Using a parallel-channel model for memotion analysis. In Proceed-ings of the Fourteenth Workshop on Semantic Eval-uation, SemEval ’20, pages 916–921, Barcelona,Spain.
Savvas Zannettou, Tristan Caulfield, Barry Bradlyn,Emiliano De Cristofaro, Gianluca Stringhini, andJeremy Blackburn. 2020a. Characterizing the use ofimages in state-sponsored information warfare oper-ations by Russian trolls on Twitter. In Proceedingsof the Fourteenth International AAAI Conference onWeb and Social Media, ICWSM ’20, pages 774–785,Atlanta, GA, USA.
Savvas Zannettou, Joel Finkelstein, Barry Bradlyn, andJeremy Blackburn. 2020b. A quantitative approachto understanding online antisemitism. In Proceed-ings of the Fourteenth International AAAI Confer-ence on Web and Social Media, ICWSM ’20, pages786–797, Atlanta, GA, USA.
Xiayu Zhong. 2020. Classification of multimodal hatespeech – the winning solution of hateful memes chal-lenge. arXiv 2012.01002.
Luowei Zhou, Hamid Palangi, Lei Zhang, HoudongHu, Jason Corso, and Jianfeng Gao. 2020. Uni-fied vision-language pre-training for image caption-ing and VQA. Proceedings of the AAAI Conferenceon Artificial Intelligence, pages 13041–13049.
Yi Zhou and Zhenhao Chen. 2020. Multimodal learn-ing for hateful memes detection. arXiv 2011.12870.
Ron Zhu. 2020. Enhance multimodal transformer withexternal label and in-domain pretrain: Hateful memechallenge winning solution. arXiv 2012.08290.
2795
A Implementation Details and
Hyper-Parameter Values
We trained all the models using the Pytorch frame-
work on an NVIDIA Tesla T4 GPU with 16 GB of
dedicated memory and with CUDA-10 and cuDNN-
11 installed. For the unimodal models, we imported
all the pre-trained weights from the TORCHVI-
SION.MODELS10 subpackage of PyTorch. We ini-
tialized the non pre-trained weights randomly with
a zero-mean Gaussian distribution with a standard
deviation of 0.02. To minimize the impact of the
label imbalance in the loss calculation, we assigned
larger weights to the minority class. We trained our
models using the Adam optimizer (Kingma and Ba,
2014) and the negative log-likelihood loss as the
objective function. Table A.1 gives the values of
all hyper-parameters we used for training.
We trained the models end-to-end for the two
classification tasks, i.e., the memes that were clas-
sified as Very Harmful or Partially Harmful in the
first classification stage were sent to the second
stage for target identification.
B Annotation Guidelines
B.1 What do we mean by harmful memes?
The entrenched meaning of harmful memes is
targeted towards a social entity (e.g., an individ-
ual, an organization, a community, etc.), likely
to cause calumny/vilification/defamation depend-
ing on their background (bias, social background,
educational background, etc.). The harm caused
by a meme can be in the form of mental abuse,
psycho-physiological injury, proprietary damage,
emotional disturbance, compensated public image.
A harmful meme typically attacks celebrities or
well-known organizations, with the intent to ex-
pose their professional demeanor.
Characteristics of harmful memes:
• Harmful memes may or may not be offensive,
hateful, or biased in nature.
• Harmful memes expose vices, allegations, and
other negative aspects of an entity based on veri-