DocumentCode :
3672350
Title :
Web-scale training for face identification
Author :
Yaniv Taigman;Ming Yang;Marc´Aurelio Ranzato;Lior Wolf
Author_Institution :
Facebook AI Research, Menlo Park, CA 94025, USA
fYear :
2015
fDate :
6/1/2015 12:00:00 AM
Firstpage :
2746
Lastpage :
2754
Abstract :
Scaling machine learning methods to very large datasets has attracted considerable attention in recent years, thanks to easy access to ubiquitous sensing and data from the web. We study face recognition and show that three distinct properties have surprising effects on the transferability of deep convolutional networks (CNN): (1) The bottleneck of the network serves as an important transfer learning regularizer, and (2) in contrast to the common wisdom, performance saturation may exist in CNN´s (as the number of training samples grows); we propose a solution for alleviating this by replacing the naive random subsampling of the training set with a bootstrapping process. Moreover, (3) we find a link between the representation norm and the ability to discriminate in a target domain, which sheds lights on how such networks represent faces. Based on these discoveries, we are able to improve face recognition accuracy on the widely used LFW benchmark, both in the verification (1:1) and identification (1:N) protocols, and directly compare, for the first time, with the state of the art Commercially-Off-The-Shelf system and show a sizable leap in performance.
Keywords :
"Face","Training","Benchmark testing","Entropy","Face recognition","Accuracy","Protocols"
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2015 IEEE Conference on
Electronic_ISBN :
1063-6919
Type :
conf
DOI :
10.1109/CVPR.2015.7298891
Filename :
7298891
Link To Document :
بازگشت