DocumentCode :
3739344
Title :
Understanding Deep Networks with Gradients
Author :
Henry Z. Lo;Wei Ding
Author_Institution :
Dept. of Comput. Sci., Univ. of Massachusetts Boston, Boston, MA, USA
fYear :
2015
Firstpage :
1548
Lastpage :
1555
Abstract :
Existing methods for understanding the inner workings of convolutional neural networks have relied on visualizations, which do not describe the connections between the layers and units of the network. We introduce the prediction gradient as a measure of a neuron´s relevance to prediction. Using this quantity, we study a relatively small convolutional neural network and make three observations. First, there exists a small number of high prediction-gradient units, which upon removal, severely impact the ability of the network to classify correctly. Second, this performance loss generalizes spans multiple classes, and is not mirrored by removing low-gradient units. Third, the distributed representation of the neural network prevents performance from being impacted until a critical number of units are destroyed, the number depending highly on the prediction gradient of the units removed. These three observations validate the utility of the prediction gradient in identifying important units in a neural network. We finally use the prediction gradient in order to generate and study adversarial examples.
Keywords :
"Biological neural networks","Visualization","Training","Conferences","Computational modeling","Robustness"
Publisher :
ieee
Conference_Titel :
Data Mining Workshop (ICDMW), 2015 IEEE International Conference on
Electronic_ISBN :
2375-9259
Type :
conf
DOI :
10.1109/ICDMW.2015.227
Filename :
7395858
Link To Document :
بازگشت