Analysis of Multilayer Neural Networks for Object Recognition
Pulkit Agrawal
EECS Department, University of California, Berkeley
Technical Report No. UCB/EECS-2014-202
December 1, 2014
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2014/EECS-2014-202.pdf
In 2012 Krizhevsky et al. demonstrated that a Convolutional Neural Network (CNN) trained on large amount of data as part of the ImageNet challenge significantly outperformed traditional computer vision approaches on image classification. Subsequently, Girshick et al. (2013) exploited these features to establish the new state of the art on PASCAL object detection. This suggests that computer vision maybe in the process of a feature revolution akin to that following SIFT and HOG nearly a decade ago. It is therefore important to get more insights into features learned by these networks. Our work provides answer to the following four questions: (a) What happens during finetuning of a discriminatively pretrained network? (b) How much information is in the location and how much of it is in the magnitude of filter activation? (c) Does a multilayer CNN contain Grand-Mother Cells? How distributed is the code? (d) How does training of CNN progress over time? Does too much pre-training hurt generalization performance?
Advisors: Jitendra Malik
BibTeX citation:
@mastersthesis{Agrawal:EECS-2014-202, Author= {Agrawal, Pulkit}, Title= {Analysis of Multilayer Neural Networks for Object Recognition}, School= {EECS Department, University of California, Berkeley}, Year= {2014}, Month= {Dec}, Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2014/EECS-2014-202.html}, Number= {UCB/EECS-2014-202}, Abstract= {In 2012 Krizhevsky et al. demonstrated that a Convolutional Neural Network (CNN) trained on large amount of data as part of the ImageNet challenge significantly outperformed traditional computer vision approaches on image classification. Subsequently, Girshick et al. (2013) exploited these features to establish the new state of the art on PASCAL object detection. This suggests that computer vision maybe in the process of a feature revolution akin to that following SIFT and HOG nearly a decade ago. It is therefore important to get more insights into features learned by these networks. Our work provides answer to the following four questions: (a) What happens during finetuning of a discriminatively pretrained network? (b) How much information is in the location and how much of it is in the magnitude of filter activation? (c) Does a multilayer CNN contain Grand-Mother Cells? How distributed is the code? (d) How does training of CNN progress over time? Does too much pre-training hurt generalization performance?}, }
EndNote citation:
%0 Thesis %A Agrawal, Pulkit %T Analysis of Multilayer Neural Networks for Object Recognition %I EECS Department, University of California, Berkeley %D 2014 %8 December 1 %@ UCB/EECS-2014-202 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2014/EECS-2014-202.html %F Agrawal:EECS-2014-202