Reconsidering Read and Spontaneous Speech: Causal Perspectives on the Generation of Training Data for Automatic Speech Recognition

Philipp Gabler, Bernhard Geiger, Barbara Schuppler, Roman Kern*

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

Abstract

Superficially, read and spontaneous speech—the two main kinds of training data for automatic speech recognition—appear as complementary, but are equal: pairs of texts and acoustic signals. Yet, spontaneous speech is typically harder for recognition. This is usually explained by different kinds of variation and noise, but there is a more fundamental deviation at play: for read speech, the audio signal is produced by recitation of the given text, whereas in spontaneous speech, the text is transcribed from a given signal. In this review, we embrace this difference by presenting a first introduction of causal reasoning into automatic speech recognition, and describing causality as a tool to study speaking styles and training data. After breaking down the data generation processes of read and spontaneous speech and analysing the domain from a causal perspective, we highlight how data generation by annotation must affect the interpretation of inference and performance. Our work discusses how various results from the causality literature regarding the impact of the direction of data generation mechanisms on learning and prediction apply to speech data. Finally, we argue how a causal perspective can support the understanding of models in speech processing regarding their behaviour, capabilities, and limitations.

Original languageEnglish
Article number137
JournalInformation
Volume14
Issue number2
DOIs
Publication statusPublished - Feb 2023

Keywords

  • annotation
  • automatic speech recognition
  • causality
  • data generation processes
  • speaking styles

ASJC Scopus subject areas

  • Information Systems

Fingerprint

Dive into the research topics of 'Reconsidering Read and Spontaneous Speech: Causal Perspectives on the Generation of Training Data for Automatic Speech Recognition'. Together they form a unique fingerprint.

Cite this