LREC 2000 2nd International Conference on Language Resources & Evaluation
 

Previous Paper   Next Paper

Title Lessons Learned from a Task-based Evaluation of Speech-to-Speech Machine Translation
Authors Levin Lori (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Bartlog Boris (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Font Llitjos Ariadna (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Gates Donna (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Lavie Alon (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, alavie@cs.cmu.edu, www.is.cs.cmu.edu)
Wallace Dorcas (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Watanabe Taro (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Woszczyna Monika (Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA 15213 USA, lsl@cs.cmu.edu)
Keywords Speech-to-Speech Machine Translation, Task-Based Evaluation
Session Session SO5 - Evaluation of Dialogue
Full Paper 215.ps, 215.pdf
Abstract For several years we have been conducting Accuracy Based Evaluations (ABE) of the JANUS speech-to-speech MT system (Gates et al., 1997) which measure quality and fidelity of translation. Recently we have begun to design a Task Based Evaluation for JANUS (Thomas, 1999) which measures goal completion. This paper describes what we have learned by comparing the two types of evaluation. Both evaluations (ABE and TBE) were conducted on a common set of user studies in the semantic domain of travel planning.