Dataset of directional room impulse responses for realistic speech data

Stefan Fragner*, Lukas Pfeifenberger, Martin Hagmüller, Franz Pernkopf

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Obtaining real-world multi-channel speech recordings is expensive and time-consuming. Therefore, multi-channel recordings are often artificially generated by convolving existing monaural speech recordings with simulated Room Impulse Responses (RIRs) from a so-called shoebox room [1] for stationary (not moving) speakers. Far-field speech processing for home automation or smart assistants have to cope with moving speakers in reverberant environments. With this dataset, we aim to support the generation of realistic speech data by providing multiple directional RIRs along a fine grid of locations in a real room. We provide directional RIR recordings for a classroom and a large corridor. These RIRs can be used to simulate moving speakers by generating random trajectories on that grid, and quantize the trajectories along the grid points. For each matching grid point, the monaural speech recording can be convolved with the RIR at this grid point. Then, the spatialized recording can be compiled using the overlap-add method for each grid point [2]. An example is provided with the data.

Original languageEnglish
Article number110229
JournalData in Brief
Volume53
DOIs
Publication statusPublished - Apr 2024

Keywords

  • Artificial intelligence
  • Deep learning
  • Reverberant speech data
  • Room impulse response
  • Speech processing

ASJC Scopus subject areas

  • General

Fingerprint

Dive into the research topics of 'Dataset of directional room impulse responses for realistic speech data'. Together they form a unique fingerprint.

Cite this