The Metaphor Understanding Challenge Dataset (MUNCH) is designed to assess the metaphor understanding capabilities of Large Language Models (LLMs). It consists of over 10k paraphrases for sentences containing metaphors, along with 1.5k instances of inapt paraphrases. The dataset covers various genres like academic, news, fiction, and conversation, offering different levels of novelty in metaphorical expressions. Experiments with LLaMA and GPT-3.5 demonstrate the challenging nature of MUNCH for LLMs. The dataset aims to evaluate whether models can perform full metaphor interpretation or rely on lexical similarity.
إلى لغة أخرى
من محتوى المصدر
arxiv.org
الرؤى الأساسية المستخلصة من
by Xiaoyu Tong,... في arxiv.org 03-19-2024
https://arxiv.org/pdf/2403.11810.pdfاستفسارات أعمق