Back to Main Conference 2008
LREC 2008main

Applying Automated Metrics to Speech Translation Dialogs

Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC 2008)

DOI:10.63317/4oxfceja8bfr

Abstract

Over the past five years, the Defense Advanced Research Projects Agency (DARPA) has funded development of speech translation systems for tactical applications. A key component of the research program has been extensive system evaluation, with dual objectives of assessing progress overall and comparing among systems. This paper describes the methods used to obtain BLEU, TER, and METEOR scores for two-way English-Iraqi Arabic systems. We compare the scores with measures based on human judgments and demonstrate the effects of normalization operations on BLEU scores. Issues that are highlighted include the quality of test data and differential results of applying automated metrics to Arabic vs. English.

Details

Paper ID
lrec2008-main-143
Pages
N/A
BibKey
condon-etal-2008-applying
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
2-9517408-4-0
Conference
Sixth International Conference on Language Resources and Evaluation
Location
Marrakech, Morocco
Date
28 May 2008 30 May 2008

Authors

  • SC

    Sherri Condon

  • JP

    Jon Phillips

  • CD

    Christy Doran

  • JA

    John Aberdeen

  • DP

    Dan Parvaz

  • BO

    Beatrice Oshika

  • GS

    Greg Sanders

  • CS

    Craig Schlenoff

Links