Back to Main Conference 2002
LREC 2002main

Automatic Evaluation: Using a DATE Dialogue Act Tagger for User Satisfaction and Task Completion Prediction

Proceedings of the Third International Conference on Language Resources and Evaluation (LREC 2002)

DOI:10.63317/2m58wpesch86

Abstract

The objective of the DARPA Communicator project is to support rapid, cost-effective development of multi-modal speech-enabled dialogue systems with advanced conversational capabilities. During the course of the Communicator program, we have been involved in developing methods for measuring progress towards the program goals  and assessing advances in the component technologies required to achieve such goals.  Our goal has been to develop a lightweight evaluation paradigm for heterogeneous systems. In this paper, we utilize the Communicator evaluation corpus from 2001 and build on previous work applying the PARADISE evaluation framework to establish a  baseline for fully automatic system evaluation. We train a regression tree to predict User Satisfaction using a random 80 of the dialogues for training. The metrics (features) we use  for prediction are a fully  automatic Task Success Measure, Efficiency Measures, and System Dialogue Act Behaviors extracted from the dialogue logfiles using the DATE (Dialogue Act Tagging for Evaluation) tagging scheme. The learned tree with the DATE metrics has a correlation of 0.614 (R

Details

Paper ID
lrec2002-main-143
Pages
N/A
BibKey
hastie-etal-2002-automatic
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
N/A
Conference
Third International Conference on Language Resources and Evaluation
Location
Las Palmas, Spain
Date
29 May 2002 31 May 2002

Authors

  • HH

    Helen Wright Hastie

  • RP

    Rashmi Prasad

  • MW

    Marilyn Walker

Links