Skip to main navigation Skip to search Skip to main content

Comparing neural network architectures for non-intrusive speech quality prediction

  • Leif Førland Schill
  • , Tobias Piechowiak
  • , Clément Laroche
  • , Pejman Mowlaee
  • GN Store Nord

Research output: Contribution to journalJournal articleResearchpeer-review

116 Downloads (Orbit)

Abstract

Non-intrusive speech quality predictors evaluate speech quality without the use of a reference signal, making them useful in many practical applications. Recently, neural networks have shown the best performance for this task. Two such models in the literature are the convolutional neural network based DNSMOS and the bi-directional long short-term memory based Quality-Net, which were originally trained to predict subjective targets and intrusive PESQ scores, respectively. In this paper, these two architectures are trained on a single dataset, and used to predict the intrusive ViSQOL score. The evaluation is done on a number of test sets with a variety of mismatch conditions, including unseen speech and noise corpora, and common voice over IP distortions. The experiments show that the models achieve similar predictive ability on the training distribution, and overall good generalization to new noise and speech corpora. Unseen distortions are identified as an area where both models generalize poorly, especially DNSMOS. Our results also suggest that a pervasiveness of ambient noise in the training set can cause problems when generalizing to certain types of noise. Finally, we detail how the ViSQOL score can have undesirable dependencies on the reference pressure level and the voice activity level.
Original languageEnglish
Article number103123
JournalSpeech Communication
Volume165
Number of pages12
ISSN0167-6393
DOIs
Publication statusPublished - 2024

Keywords

  • Generalization
  • Neural network
  • Non-intrusive
  • Speech quality prediction

Fingerprint

Dive into the research topics of 'Comparing neural network architectures for non-intrusive speech quality prediction'. Together they form a unique fingerprint.

Cite this