DocumentCode :
730658
Title :
Directly modeling speech waveforms by neural networks for statistical parametric speech synthesis
Author :
Tokuday, Keiichi ; Zen, Heiga
fYear :
2015
fDate :
19-24 April 2015
Firstpage :
4215
Lastpage :
4219
Abstract :
This paper proposes a novel approach for directly-modeling speech at the waveform level using a neural network. This approach uses the neural network-based statistical parametric speech synthesis framework with a specially designed output layer. As acoustic feature extraction is integrated to acoustic model training, it can overcome the limitations of conventional approaches, such as two-step (feature extraction and acoustic modeling) optimization, use of spectra rather than waveforms as targets, use of overlapping and shifting frames as unit, and fixed decision tree structure. Experimental results show that the proposed approach can directly maximize the likelihood defined at the waveform domain.
Keywords :
acoustic signal processing; feature extraction; learning (artificial intelligence); maximum likelihood estimation; neural nets; speech synthesis; acoustic feature extraction; acoustic model training; likelihood maximization; neural network; speech waveform direct model; statistical parametric speech synthesis; waveform domain; Algorithm design and analysis; Cepstral analysis; Distortion; Hidden Markov models; Speech; Training; Statistical parametric speech synthesis; adaptive cepstral analysis; neural network;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
Type :
conf
DOI :
10.1109/ICASSP.2015.7178765
Filename :
7178765
Link To Document :
بازگشت