Back to Main Conference 2002
LREC 2002main

Evaluation of Machine Learning Methods for Natural Language Processing Tasks

Proceedings of the Third International Conference on Language Resources and Evaluation (LREC 2002)

DOI:10.63317/2wmcmskiy5tm

Abstract

We show that the methodology currently in use for comparing symbolic supervised learning methods applied to human language technology tasks is unreliable. We show that the interaction between algorithm parameter settings and feature selection within a single algorithm often accounts for a higher variation in results than differences between different algorithms or information sources. We illustrate this with experiments on a number of linguistic datasets. The consequences of this phenomenon are far-reaching, and we discuss possible solutions to this methodological problem.

Details

Paper ID
lrec2002-main-094
Pages
N/A
BibKey
daelemans-hoste-2002-evaluation
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
N/A
Conference
Third International Conference on Language Resources and Evaluation
Location
Las Palmas, Spain
Date
29 May 2002 31 May 2002

Authors

  • WD

    Walter Daelemans

  • VH

    Véronique Hoste

Links