Back to Main Conference 2024
LREC-COLING 2024main

Argument Quality Assessment in the Age of Instruction-Following Large Language Models

Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

DOI:10.63317/2v24be3iq8pn

Abstract

The computational treatment of arguments on controversial issues has been subject to extensive NLP research, due to its envisioned impact on opinion formation, decision making, writing education, and the like. A critical task in any such application is the assessment of an argument’s quality - but it is also particularly challenging. In this position paper, we start from a brief survey of argument quality research, where we identify the diversity of quality notions and the subjectiveness of their perception as the main hurdles towards substantial progress on argument quality assessment. We argue that the capabilities of instruction-following large language models (LLMs) to leverage knowledge across contexts enable a much more reliable assessment. Rather than just fine-tuning LLMs towards leaderboard chasing on assessment tasks, they need to be instructed systematically with argumentation theories and scenarios as well as with ways to solve argument-related problems. We discuss the real-world opportunities and ethical issues emerging thereby.

Details

Paper ID
lrec2024-main-0135
Pages
pp. 1519-1538
BibKey
wachsmuth-etal-2024-argument
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
2522-2686
ISBN
979-10-95546-34-4
Conference
Joint International Conference on Computational Linguistics, Language Resources and Evaluation
Location
Turin, Italy
Date
20 May 2024 25 May 2024

Authors

  • HW

    Henning Wachsmuth

  • GL

    Gabriella Lapesa

  • EC

    Elena Cabrio

  • AL

    Anne Lauscher

  • JP

    Joonsuk Park

  • EV

    Eva Maria Vecchi

  • SV

    Serena Villata

  • TZ

    Timon Ziegenbein

Links