DocumentCode :
3734282
Title :
On the weight sparsity of multilayer perceptrons
Author :
Georgios Drakopoulos;Vasileios Megalooikonomou
Author_Institution :
Multidimensional Data Analysis and Knowledge Management Lab, Computer Engineering and Informatics Department, University of Patras, Patras 26500, Hellas
fYear :
2015
fDate :
7/1/2015 12:00:00 AM
Firstpage :
1
Lastpage :
6
Abstract :
Approximating and representing a process, a function, or a system with an adaptive parametric model constitutes a major part of current machine learning research. An important characteristic of these models is parameter sparsity, an indicator of how succintly a model can codify fundamental properties of the approximated function. This paper investigates the sparsity patterns of a multilayer perceptron netwrok trained to mount a man-on-the-middle attack on the DES symmetric cryptosystem. The notions of absolute and effective synaptic weight sparsity are introduced and their importance to network learning procedure is explained. Finally, the results from the training of the actual multilayer perceptron are outlined and discussed. In order to promote reproducible research, the MATLAB network implementation has been posted in GitHub.
Keywords :
"Neurons","Biological neural networks","Multilayer perceptrons","Encryption","Training","Standards"
Publisher :
ieee
Conference_Titel :
Information, Intelligence, Systems and Applications (IISA), 2015 6th International Conference on
Type :
conf
DOI :
10.1109/IISA.2015.7388096
Filename :
7388096
Link To Document :
بازگشت