Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Do Androids Laugh at Electric Sheep? Humor "Understanding" Benchmarks from The New Yorker Caption Contest (2209.06293v2)

Published 13 Sep 2022 in cs.CL and cs.CV

Abstract: Large neural networks can now generate jokes, but do they really "understand" humor? We challenge AI models with three tasks derived from the New Yorker Cartoon Caption Contest: matching a joke to a cartoon, identifying a winning caption, and explaining why a winning caption is funny. These tasks encapsulate progressively more sophisticated aspects of "understanding" a cartoon; key elements are the complex, often surprising relationships between images and captions and the frequent inclusion of indirect and playful allusions to human experience and culture. We investigate both multimodal and language-only models: the former are challenged with the cartoon images directly, while the latter are given multifaceted descriptions of the visual scene to simulate human-level visual understanding. We find that both types of models struggle at all three tasks. For example, our best multimodal models fall 30 accuracy points behind human performance on the matching task, and, even when provided ground-truth visual scene descriptors, human-authored explanations are preferred head-to-head over the best machine-authored ones (few-shot GPT-4) in more than 2/3 of cases. We release models, code, leaderboard, and corpus, which includes newly-gathered annotations describing the image's locations/entities, what's unusual in the scene, and an explanation of the joke.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (72)
  1. Miriam Amin and Manuel Burghardt. 2020. A survey on approaches to computational humor generation. In The 4th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature.
  2. Issa Annamoradnejad and Gohar Zoghi. 2020. ColBERT: Using BERT sentence embedding for humor detection. arXiv preprint arXiv:2004.12765.
  3. Salvatore Attardo. 2008. A primer for the linguistics of humor. The primer of humor research, 8:101–55.
  4. Michael Billig. 2005. Laughter and ridicule: Towards a social critique of humour. Sage.
  5. Kim Binsted and Graeme Ritchie. 1994. An implemented model of punning riddles. In AAAI.
  6. Large dataset and language model fun-tuning for humor recognition. In ACL.
  7. Language models are few-shot learners. NeurIPS.
  8. Towards multimodal sarcasm detection (an _Obviously_ perfect paper). In ACL.
  9. FLUTE: figurative language understanding and textual explanations. In EMNLP.
  10. Punny captions: Witty wordplay in image descriptions. In NAACL.
  11. We are humor beings: Understanding and predicting visual humor. In CVPR.
  12. PaLM: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311.
  13. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR.
  14. Analyzing humor in newspaper comic strips using verbal-visual analysis. Lingua Cultura, 12(4):383–388.
  15. Christiane Fellbaum. 1998. WordNet: An Electronic Lexical Database. Bradford Books.
  16. Sigmund Freud. 1905. Jokes and their Relation to the Unconscious, volume 8 of The Standard Edition of the Complete Psychological Works of Sigmund Freud. Hogarth, London.
  17. William F. Fry. 1963. Sweet madness: A study of humor. Pacific Books, Palo Alto.
  18. Charles R. Gruner. 1978. Understanding laughter: The workings of wit & humor. Nelson-Hall, Chicago.
  19. Kilem Gwet. 2014. Handbook of Inter-Rater reliability: The Definitive Guide to Measuring the Extent of Agreement Among Raters, 4th edition edition. Advanced Analytics, LLC.
  20. Humor knowledge enriched transformer for understanding multimodal humor. In AAAI.
  21. UR-FUNNY: a multimodal language dataset for understanding humor. In EMNLP.
  22. The curious case of neural text degeneration. In ICLR.
  23. The New Yorker cartoon caption contest dataset.
  24. NEXT: A system for real-world development, evaluation, and application of active learning. In NeurIPS.
  25. Random walk factoid annotation for collective discourse. In ACL.
  26. Large language models are zero-shot reasoners. In NeurIPS.
  27. Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text summarization branches out.
  28. Microsoft COCO: Common objects in context. In ECCV.
  29. Tania Lombrozo. 2006. The structure and function of explanations. Trends in Cognitive Sciences, 10(10):464–470.
  30. Few-shot self-rationalization with natural language prompts. In Findings of NAACL.
  31. Rada Mihalcea and Stephen Pulman. 2009. Characterizing humour: An exploration of features in humorous texts. In Proceedings of the 8th International Conference on Computational Linguistics and Intelligent Text Processing, page 337–347, Berlin, Heidelberg. Springer-Verlag.
  32. Rada Mihalcea and Carlo Strapparava. 2005. Making computers laugh: Investigations in automatic humor recognition. In EMNLP.
  33. Rada Mihalcea and Carlo Strapparava. 2006. Technologies that make you smile: Adding humor to text-based applications. IEEE Intelligent Systems, 21(5):33–39.
  34. Harvey Mindess. 1971. Laughter and Liberation. Nash.
  35. The pudding caption contest experiments. https://pudding.cool/projects/caption-contest/. Accessed: 2022-04-01.
  36. Matthijs P. Mulder and Antinus Nijholt. 2002. Humour research: State of the art. Centre for Telematics and Information Technology, University of Twente.
  37. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748.
  38. OpenAI. 2023. Gpt-4 technical report.
  39. BLEU: a method for automatic evaluation of machine translation. In ACL.
  40. Pytorch: An imperative style, high-performance deep learning library. In NeurIPS.
  41. Multimodal humor dataset: Predicting laughter tracks for sitcoms. In WACV.
  42. Hypothesis only baselines in natural language inference. In *SEM.
  43. Matt Post. 2018. A call for clarity in reporting BLEU scores. In WMT.
  44. Humor in collective discourse: Unsupervised funniness detection in the New Yorker cartoon caption contest. In LREC.
  45. Learning transferable visual models from natural language supervision. In ICML.
  46. Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR.
  47. Victor Raskin. 1979. Semantic mechanisms of humor. In Annual Meeting of the Berkeley Linguistics Society, volume 5, pages 325–335.
  48. Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters. In KDD.
  49. Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence embeddings using Siamese BERT-networks. In EMNLP.
  50. Arthur Schopenhauer. 1818. The world as will and idea, volume 1.
  51. Inside jokes: Identifying humorous cartoon captions. In KDD.
  52. News2meme: An automatic content generator from news based on word subspaces from text and image. In Conference on Machine Vision Applications.
  53. Thomas R Shultz. 1976. A cognitive-developmental analysis of humour. Transaction Publishers.
  54. Oliviero Stock and Carlo Strapparava. 2003. Getting serious about the development of computational humor. In IJCAI.
  55. Rajesh Shanmuga Sundaram. 2018. Generation of Humorous Caption for Cartoon Images Using Deep Learning. Ph.D. thesis, Texas A&M University-Commerce.
  56. Chenhao Tan. 2022. On the diversity and limits of human explanations. In NAACL.
  57. A KL-LUCB algorithm for large-scale crowdsourcing. In NeurIPS.
  58. How reasonable are common-sense reasoning tasks: A case-study on the Winograd schema challenge and SWAG. In EMNLP.
  59. Villy Tsakona. 2009. Language and image interaction in cartoons: Towards a multimodal theory of humor. Journal of Pragmatics, 41(6):1171–1188.
  60. Alessandro Valitutti. 2011. How many jokes are really funny? In Human-Machine Interaction in Translation: Proceedings of the 8th International NLPCS Workshop.
  61. Attention is all you need. NeurIPS.
  62. David Wallace. 2022. Lecture notes for MIT 2.00b toy product design: Innovation and associations.
  63. Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework. In ICML.
  64. William Yang Wang and Miaomiao Wen. 2015. I can has cheezburger? a nonparanormal approach to combining textual and visual information for predicting and generating popular meme descriptions. In NAACL.
  65. Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
  66. White, E. B. 1941. Preface. In E. B. White and Katherine S. White, editors, A Subtreasury Of American Humor, page xvii. The original version of this quote appeared as a preview in The Saturday Review (1941), credited to both Whites. But, the quote appears in the preface to A Subtreasury (1941) with authorship solely credited to E.B.. We thus credited the quote itself to E.B., and credited both E.B. and K.S. as editors of the anthology in which it appears in non-preview form.
  67. Reframing human-AI collaboration for generating free-text explanations. In NAACL.
  68. Hannah Wilson. 2019. Project four - nobody knows you’re a bot.
  69. Transformers: State-of-the-art natural language processing. In EMNLP: System Demonstrations.
  70. Neural joking machine: Humorous image captioning. In CVPR Language & Vision Workshop.
  71. Michael Zelenko and Frank Bi. 2015. On the internet, nobody knows you’re a machine.
  72. Socratic models: Composing zero-shot multimodal reasoning with language. arXiv preprint arXiv:2204.00598.
Citations (67)

Summary

  • The paper introduces novel tasks to assess AI humor understanding using the New Yorker Caption Contest framework.
  • The study employs both multimodal and language-only models, revealing a 30-point accuracy gap in matching tasks compared to human performance.
  • The research underscores the complexity of humor comprehension and the need for culturally-informed, multimodal AI systems.

Humor "Understanding" Benchmarks from The New Yorker Caption Contest

The paper "Do Androids Laugh at Electric Sheep? Humor 'Understanding' Benchmarks from The New Yorker Caption Contest" presents a meticulous analysis of AI systems' ability to comprehend humor. The authors have devised a series of tasks based on the New Yorker Cartoon Caption Contest to test AI's humor understanding capabilities. These tasks involve matching a cartoon with its corresponding caption, assessing the quality of captions, and explaining why certain captions are humorous. The paper reveals significant gaps in AI performance compared to human benchmarks, underscoring the complexity of humor understanding even with advanced AI systems like GPT-4.

Research Design and Methods

The paper employs both language-only and multimodal models to address three specific tasks:

  1. Matching Task: Models are required to match captions to cartoons from a set of distractors.
  2. Quality Ranking Task: Models need to assess the quality of captions, distinguishing winners from non-finalists.
  3. Explanation Task: Models generate explanations for why captions are funny, simulating deeper humor comprehension.

For multimodal evaluation, the authors leverage CLIP and OFA models, which integrate visual cues with language processing. In contrast, language-only models, such as T5 and various GPT versions, simulate "visual understanding" using rich descriptions rather than actual images.

Key Findings

The authors report that both models struggle significantly with these tasks. The best-performing multimodal models lag by 30 accuracy points compared to human performance in matching tasks. In explanation tasks, human-authored explanations are preferred over those of models two-thirds of the time, even when using advanced LLMs like GPT-4.

The discrepancy highlights the nuanced difficulty AI systems face in humor comprehension, which is characterized by subtle and often culturally-dependent associations between text and imagery.

Implications and Future Directions

The findings of this paper have dual implications. Practically, they suggest that while AI technologies have advanced considerably, subtleties in human-like tasks such as humor understanding remain challenging. Theoretically, the research emphasizes the value of integrating multimodal learning to capture the nuanced interplay between linguistic and visual stimuli that humor often requires.

From a speculative standpoint, closing this gap may necessitate more sophisticated models incorporating wider cultural and experiential data. This indicates a future where AI could assist more comprehensively in creative processes, contributing substantively to tasks that require deep context understanding, such as writing, education, and entertainment.

Conclusion

The research presents significant contributions to AI's understanding of humor, introducing benchmarks that mirror human performance for more realistic assessments. The continued development of such tasks will be vital in advancing AI’s capabilities in comprehending increasingly complex human characteristics, thereby propelling advancements in AI-human collaborative environments. The paper serves as a foundational step in realizing AI systems that can one day fully engage in the multifaceted domain of human humor, emphasizing that humor understanding remains a formidable challenge in the evolution of intelligent systems.

Youtube Logo Streamline Icon: https://streamlinehq.com