Back to DELITE 2024
LREC-COLING 2024workshop

Pitfalls of Conversational LLMs on News Debiasing

Proceedings of the First Workshop on Language-driven Deliberation Technology (DELITE) @ LREC-COLING 2024

DOI:10.63317/4r2sfz6qga4y

Abstract

This paper addresses debiasing in news editing and evaluates the effectiveness of conversational Large Language Models in this task. We designed an evaluation checklist tailored to news editors’ perspectives, obtained generated texts from three popular conversational models using a subset of a publicly available dataset in media bias, and evaluated the texts according to the designed checklist. Furthermore, we examined the models as evaluator for checking the quality of debiased model outputs. Our findings indicate that none of the LLMs are perfect in debiasing. Notably, some models, including ChatGPT, introduced unnecessary changes that may impact the author’s style and create misinformation. Lastly, we show that the models do not perform as proficiently as domain experts in evaluating the quality of debiased outputs.

Details

Paper ID
lrec2024-ws-delite-4
Pages
pp. 33-38
BibKey
baris-schlicht-etal-2024-pitfalls
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
N/A
ISBN
N/A
Workshop
Proceedings of the First Workshop on Language-driven Deliberation Technology (DELITE) @ LREC-COLING 2024
Location
undefined, undefined
Date
20 May 2024 25 May 2024

Authors

  • IB

    Ipek Baris Schlicht

  • DA

    Defne Altiok

  • MT

    Maryanne Taouk

  • LF

    Lucie Flek

Links