An analysis of prosodic boundary detection in German and Austrian German read speech

Barbara Schuppler, Bogdan Ludusan

Publikation: Beitrag in Buch/Bericht/KonferenzbandBeitrag in einem KonferenzbandBegutachtung

Abstract

With speech annotation being one of the most time-consuming and costly aspects of speech corpora development,there is a significant interest in the development of automatic annotation tools. The present study focuses on variant-independent prosodic boundary annotations for German. We test a previously proposed unsupervised approach, which posits prosodic boundaries based only on acoustic cues. The experiments were conducted on read speech from two corpora, one of Standard German, the Kiel Corpus of Spoken German, and the other of Austrian German, the Graz Corpus of Read and Spontaneous Speech. Averaging across all speakers in the dataset,the tool attained an area under the precision-recall curve of0.308 and 0.215, for the Kiel corpus and the GRASS corpus,respectively. The significant differences obtained in detection across the two varieties were accompanied by large differences between speakers, as well. This was confirmed by a subsequent analysis of the acoustic cues employed in the process, which showed important differences in the way speakers make use of those cues for marking prosodic structure. We discuss these findings with respect to the current literature and their implication for variant-independent automatic annotation.
Originalspracheenglisch
TitelProceedings of Speech Prosody 2020
Untertitel10th International Conference on Speech Prosody 2020
ErscheinungsortTokyo, Japan
Seiten990- 994
Seitenumfang5
Band2020-May
DOIs
PublikationsstatusVeröffentlicht - 1 Jan. 2020

Publikationsreihe

NameProceedings of the International Conference on Speech Prosody
ISSN (Print)2333-2042

ASJC Scopus subject areas

  • Sprache und Linguistik
  • Linguistik und Sprache

Fingerprint

Untersuchen Sie die Forschungsthemen von „An analysis of prosodic boundary detection in German and Austrian German read speech“. Zusammen bilden sie einen einzigartigen Fingerprint.

Dieses zitieren