Jump to Content

GENERATIVE SPEECH CODING WITH PREDICTIVE VARIANCE REGULARIZATION

Alejandro Luebs
Andrew Storus
Bastiaan Kleijn
Michael Chinen
Tom Denton
Yero Yeh
ICASSP 2021 (2021)
Google Scholar

Abstract

The recent emergence of machine-learning based generative models for speech suggests a significant reduction in bit rate for speech codecs is possible. However, the performance of generative models deteriorates significantly with the distortions present in real-world input signals. We argue that this deterioration is due to the sensitivity of the maximum likelihood criterion to outliers and the ineffectiveness of modeling a sum of independent signals with a single autoregressive model. We introduce predictive-variance regularization to reduce the sensitivity to outliers, resulting in a significant increase in performance. We show that noise reduction to remove unwanted signals can significantly increase performance. We provide extensive subjective performance evaluations that show that our system based on generative modeling provides state-of-the-art coding performance at 3 kb/s for real-world speech signals at reasonable computational complexity.

Research Areas