Title :
Directly modeling speech waveforms by neural networks for statistical parametric speech synthesis
Author :
Tokuday, Keiichi ; Zen, Heiga
Abstract :
This paper proposes a novel approach for directly-modeling speech at the waveform level using a neural network. This approach uses the neural network-based statistical parametric speech synthesis framework with a specially designed output layer. As acoustic feature extraction is integrated to acoustic model training, it can overcome the limitations of conventional approaches, such as two-step (feature extraction and acoustic modeling) optimization, use of spectra rather than waveforms as targets, use of overlapping and shifting frames as unit, and fixed decision tree structure. Experimental results show that the proposed approach can directly maximize the likelihood defined at the waveform domain.
Keywords :
acoustic signal processing; feature extraction; learning (artificial intelligence); maximum likelihood estimation; neural nets; speech synthesis; acoustic feature extraction; acoustic model training; likelihood maximization; neural network; speech waveform direct model; statistical parametric speech synthesis; waveform domain; Algorithm design and analysis; Cepstral analysis; Distortion; Hidden Markov models; Speech; Training; Statistical parametric speech synthesis; adaptive cepstral analysis; neural network;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
DOI :
10.1109/ICASSP.2015.7178765