Back to Main Conference 2008
LREC 2008main

In-car Speech Data Collection along with Various Multimodal Signals

Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC 2008)

DOI:10.63317/25sdgvxpnix3

Abstract

In this paper, a large-scale real-world speech database is introduced along with other multimedia driving data. We designed a data collection vehicle equipped with various sensors to synchronously record twelve-channel speech, three-channel video, driving behavior including gas and brake pedal pressures, steering angles, and vehicle velocities, physiological signals including driver heart rate, skin conductance, and emotion-based sweating on the palms and soles, etc. These multimodal data are collected while driving on city streets and expressways under four different driving task conditions including two kinds of monologues, human-human dialog, and human-machine dialog. We investigated the response timing of drivers against navigator utterances and found that most overlapped with the preceding utterance due to the task characteristics and the features of Japanese. When comparing utterance length, speaking rate, and the filler rate of driver utterances in human-human and human-machine dialogs, we found that drivers tended to use longer and faster utterances with more fillers to talk with humans than machines.

Details

Paper ID
lrec2008-main-505
Pages
N/A
BibKey
ozaki-etal-2008-car
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
2-9517408-4-0
Conference
Sixth International Conference on Language Resources and Evaluation
Location
Marrakech, Morocco
Date
28 May 2008 30 May 2008

Authors

  • AO

    Akira Ozaki

  • SH

    Sunao Hara

  • TK

    Takashi Kusakawa

  • CM

    Chiyomi Miyajima

  • TN

    Takanori Nishino

  • NK

    Norihide Kitaoka

  • KI

    Katunobu Itou

  • KT

    Kazuya Takeda

Links