The Metaphor Understanding Challenge Dataset (MUNCH) is designed to assess the metaphor understanding capabilities of Large Language Models (LLMs). It consists of over 10k paraphrases for sentences containing metaphors, along with 1.5k instances of inapt paraphrases. The dataset covers various genres like academic, news, fiction, and conversation, offering different levels of novelty in metaphorical expressions. Experiments with LLaMA and GPT-3.5 demonstrate the challenging nature of MUNCH for LLMs. The dataset aims to evaluate whether models can perform full metaphor interpretation or rely on lexical similarity.
Til et andet sprog
fra kildeindhold
arxiv.org
Vigtigste indsigter udtrukket fra
by Xiaoyu Tong,... kl. arxiv.org 03-19-2024
https://arxiv.org/pdf/2403.11810.pdfDybere Forespørgsler