Loading Now

Summary of Getting Serious About Humor: Crafting Humor Datasets with Unfunny Large Language Models, by Zachary Horvitz et al.


Getting Serious about Humor: Crafting Humor Datasets with Unfunny Large Language Models

by Zachary Horvitz, Jingru Chen, Rahul Aditya, Harshvardhan Srivastava, Robert West, Zhou Yu, Kathleen McKeown

First submitted to arxiv on: 23 Feb 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper investigates the potential of large language models (LLMs) in generating synthetic data for humor detection. Specifically, it explores whether LLMs can “unfun” jokes by editing texts to make them non-humorous. The authors benchmark their approach on an existing human dataset and demonstrate that current LLMs, such as GPT-4, are capable of successfully generating unfunny versions of humorous texts. This ability is evaluated both subjectively by humans and objectively using the downstream task of humor detection. Furthermore, the paper extends its approach to a code-mixed English-Hindi humor dataset, where it finds that GPT-4’s synthetic data is highly rated by bilingual annotators and provides challenging adversarial examples for humor classifiers.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper looks at how computers can help us detect what’s funny or not. Right now, detecting humor is hard because we don’t have enough good examples of jokes with non-joke versions to compare them to. The authors want to see if big language models can create these “unfun” jokes themselves. They tested their idea on a dataset of human-labeled humorous texts and found that the computers were pretty good at making jokes not funny anymore! They even tried it out on jokes in two languages, English and Hindi, and people thought the computer-made unfunny jokes were really well done. This could help us create better ways to tell whether something is meant to be funny or not.

Keywords

* Artificial intelligence  * Gpt  * Synthetic data