Unsupervised speech enhancement with deep dynamical generative speech and noise models - Institut d'Electronique et de Télécommunications de Rennes Accéder directement au contenu
Communication Dans Un Congrès Année : 2023

Unsupervised speech enhancement with deep dynamical generative speech and noise models

Résumé

This work builds on a previous work on unsupervised speech enhancement using a dynamical variational autoencoder (DVAE) as the clean speech model and non-negative matrix factorization (NMF) as the noise model. We propose to replace the NMF noise model with a deep dynamical generative model (DDGM) depending either on the DVAE latent variables, or on the noisy observations, or on both. This DDGM can be trained in three configurations: noise-agnostic, noise-dependent and noise adaptation after noise-dependent training. Experimental results show that the proposed method achieves competitive performance compared to state-of-the-art unsupervised speech enhancement methods, while the noise-dependent training configuration yields a much more time-efficient inference process.

Dates et versions

hal-04132312 , version 1 (19-06-2023)

Licence

Paternité

Identifiants

Citer

Xiaoyu Lin, Simon Leglaive, Laurent Girin, Xavier Alameda-Pineda. Unsupervised speech enhancement with deep dynamical generative speech and noise models. Interspeech 2023 - 24th Annual Conference of the International Speech Communication Association, ISCA, Aug 2023, Dublin, Ireland. pp.1-5. ⟨hal-04132312⟩
91 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More