DocumentCode :
3673894
Title :
Convolutional recurrent neural networks: Learning spatial dependencies for image representation
Author :
Zhen Zuo;Bing Shuai;Gang Wang;Xiao Liu;Xingxing Wang;Bing Wang;Yushi Chen
Author_Institution :
Nanyang Technological University, Singapore
fYear :
2015
fDate :
6/1/2015 12:00:00 AM
Firstpage :
18
Lastpage :
26
Abstract :
In existing convolutional neural networks (CNNs), both convolution and pooling are locally performed for image regions separately, no contextual dependencies between different image regions have been taken into consideration. Such dependencies represent useful spatial structure information in images. Whereas recurrent neural networks (RNNs) are designed for learning contextual dependencies among sequential data by using the recurrent (feedback) connections. In this work, we propose the convolutional recurrent neural network (C-RNN), which learns the spatial dependencies between image regions to enhance the discriminative power of image representation. The C-RNN is trained in an end-to-end manner from raw pixel images. CNN layers are firstly processed to generate middle level features. RNN layer is then learned to encode spatial dependencies. The C-RNN can learn better image representation, especially for images with obvious spatial contextual dependencies. Our method achieves competitive performance on ILSVRC 2012, SUN 397, and MIT indoor.
Keywords :
"Recurrent neural networks","Context","Image representation","Mathematical model","Computational modeling","Visualization"
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition Workshops (CVPRW), 2015 IEEE Conference on
Electronic_ISBN :
2160-7516
Type :
conf
DOI :
10.1109/CVPRW.2015.7301268
Filename :
7301268
Link To Document :
بازگشت