Papers
arxiv:2311.09761

MAFALDA: A Benchmark and Comprehensive Study of Fallacy Detection and Classification

Published on Nov 16, 2023
Authors:
,
,
,
,

Abstract

MAFALDA is a benchmark for fallacy classification that combines existing datasets, introduces a new taxonomy and annotation scheme, and evaluates language models and human performance in detecting and classifying fallacies under zero-shot conditions.

AI-generated summary

We introduce MAFALDA, a benchmark for fallacy classification that merges and unites previous fallacy datasets. It comes with a taxonomy that aligns, refines, and unifies existing classifications of fallacies. We further provide a manual annotation of a part of the dataset together with manual explanations for each annotation. We propose a new annotation scheme tailored for subjective NLP tasks, and a new evaluation method designed to handle subjectivity. We then evaluate several language models under a zero-shot learning setting and human performances on MAFALDA to assess their capability to detect and classify fallacies.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2311.09761
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 2

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2311.09761 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2311.09761 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.