DocumentCode :
1797118
Title :
Recurrent Deep-Stacking Networks for sequence classification
Author :
Palangi, Hamid ; Li Deng ; Ward, Rabab K.
Author_Institution :
Univ. of British Columbia, Vancouver, BC, Canada
fYear :
2014
fDate :
9-13 July 2014
Firstpage :
510
Lastpage :
514
Abstract :
Deep Stacking Networks (DSNs) are constructed by stacking shallow feed-forward neural networks on top of each other using concatenated features derived from the lower modules of the DSN and the raw input data. DSNs do not have recurrent connections, making them less effective to model and classify input data with temporal dependencies. In this paper, we embed recurrent connections into the DSN, giving rise to Recurrent Deep Stacking Networks (R-DSNs). Each module of the R-DSN consists of a special form of recurrent neural networks. Generalizing from the earlier DSN, the use of linearity in the output units of the R-DSN enables us to derive a closed form for computing the gradient of the cost function with respect to all network matrices without backpropagating errors. Each module in the R-DSN is initialized with an echo state network, where the input and recurrent weights are fixed to have the echo state property. Then all connection weights within the module are fine tuned using batch-mode gradient descent where the gradient takes an analytical form. Experiments are performed on the TIMIT dataset for frame-level phone state classification with 183 classes. The results show that the R-DSN gives higher classification accuracy over a single recurrent neural network without stacking.
Keywords :
feedforward neural nets; gradient methods; matrix algebra; pattern classification; recurrent neural nets; R-DSNs; TIMIT dataset; analytical form; batch-mode gradient descent; concatenated features; cost function; echo state network; frame-level phone state classification; input data classification; network matrices; raw input data; recurrent deep-stacking networks; recurrent weights; sequence classification; shallow feedforward neural network stacking; single recurrent neural network; temporal dependencies; Abstracts; Cost function; Equations; Field-flow fractionation; Mathematical model; Silicon; Training; Deep Learning; Recurrent Neural Network; Sequence Classification; Stacking;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal and Information Processing (ChinaSIP), 2014 IEEE China Summit & International Conference on
Conference_Location :
Xi´an
Print_ISBN :
978-1-4799-5401-8
Type :
conf
DOI :
10.1109/ChinaSIP.2014.6889295
Filename :
6889295
Link To Document :
بازگشت