learning deep features for discriminative …cnnlocalization.csail.mit.edu/supp.pdflearning deep...

9
Learning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, Antonio Torralba Computer Science and Artificial Intelligence Laboratory, MIT {bzhou,khosla,agata,oliva,torralba}@csail.mit.edu 1. Experiment Setups and More Results 1.1. Class Activation Maps Class Activation Maps. More examples of the class activation maps generated by GoogLeNet-GAP using the CAM technique are shown in Figure 1. The comparison results from different CNN-GAPs along with backpropaga- tion results are shown in Figure 2. Localization results. More localization results done by GoogLeNet-GAP using the CAM technique are shown in Figure 3 Comparison of CNN-GAP and CNN-GMP. The class activation maps generated from GoogLeNet-GAP and GoogLeNet-GMP are shown in Figure 4. The proposed Class Activation Mapping technique could be applied to CNNs both trained with global average pooling and CNNs trained with global maximum pooling. But from the com- parison of the heatmaps from two networks, we can see that network with global average pooling could generate better heatmaps, highlighting larger object regions with less back- ground noise. 1.2. Classification using deep features In SUN397 experiment [8], the training size is 50 im- ages per category. Experiments are ran on 10 Splits of train set and test set given in the dataset. In MIT Indoor67 ex- periment [6], the training size is 100 images per category. Experiment is ran on 1 split of train set and test set given in the dataset. In the Scene15 experiment [3], the training size is 50 images per category. Experiments are ran on 10 random splits of train set and test set. In the SUN Attribute experiment [5], the training size is 150 images per attribute. The report result is average precision. The splits of train set and test set are given in the paper. In Caltech101 and Cal- tech256 experiment [1, 2], the training size is 30 images per category. The experiments are ran on 10 random splits of train set and test set. In Stanford Action40 experiment [9], the training size is 100 images per category. Experiments are ran on 10 random splits of train set and test set. The reported result is classification accuracy. In UIUC Event8 experiment [4], training size is 70 per category and the test- ing size is 60 images per category. The experiments are ran on 10 random splits of train set and test set. We plot more class activation maps from Stanford Action 40 dataset and SUN397 in Figure 6 and Figure 7. 1.3. Pattern Discovery Discovering the informative objects in the scenes. The list of 10 scene categories with fully annotations from SUN397 are bathroom, bedroom, building facade, dining room, highway, kitchen, living room, mountain snowy, skyscraper, street, totally 4675 images. We train a one-vs- all linear SVM for each class. Concept localization in weakly labeled images. The detail of the concept discovery algorithm is in [10]. Ba- sically for each concept in the text which have enough im- ages, we learn a binary classifier, then we use the weights of the classifier to generate the class activation maps for those top ranked images detected with this concept. Weakly supervised text detector. There are 350 Google street images in the SVT dataset [7], which are taken as positive training set. We randomly select 1750 im- ages from SUN attribute dataset [5] as negative training set, then train a linear SVM. 2. Visualizing Class-Specific Units More class-specific units from the AlexNet*-GAP trained on ImageNet and Places are plotted in Figure 8. References [1] L. Fei-Fei, R. Fergus, and P. Perona. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. Computer Vision and Image Understanding, 2007. 1 [2] G. Griffin, A. Holub, and P. Perona. Caltech-256 object cat- egory dataset. 2007. 1 [3] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. Proc. CVPR, 2006. 1 [4] L.-J. Li and L. Fei-Fei. What, where and who? classifying events by scene and object recognition. Proc. ICCV, 2007. 1 1

Upload: others

Post on 09-Jun-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Learning Deep Features for Discriminative Localization

Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, Antonio TorralbaComputer Science and Artificial Intelligence Laboratory, MIT

{bzhou,khosla,agata,oliva,torralba}@csail.mit.edu

1. Experiment Setups and More Results

1.1. Class Activation Maps

Class Activation Maps. More examples of the classactivation maps generated by GoogLeNet-GAP using theCAM technique are shown in Figure 1. The comparisonresults from different CNN-GAPs along with backpropaga-tion results are shown in Figure 2.

Localization results. More localization results done byGoogLeNet-GAP using the CAM technique are shown inFigure 3

Comparison of CNN-GAP and CNN-GMP. The classactivation maps generated from GoogLeNet-GAP andGoogLeNet-GMP are shown in Figure 4. The proposedClass Activation Mapping technique could be applied toCNNs both trained with global average pooling and CNNstrained with global maximum pooling. But from the com-parison of the heatmaps from two networks, we can see thatnetwork with global average pooling could generate betterheatmaps, highlighting larger object regions with less back-ground noise.

1.2. Classification using deep features

In SUN397 experiment [8], the training size is 50 im-ages per category. Experiments are ran on 10 Splits of trainset and test set given in the dataset. In MIT Indoor67 ex-periment [6], the training size is 100 images per category.Experiment is ran on 1 split of train set and test set givenin the dataset. In the Scene15 experiment [3], the trainingsize is 50 images per category. Experiments are ran on 10random splits of train set and test set. In the SUN Attributeexperiment [5], the training size is 150 images per attribute.The report result is average precision. The splits of train setand test set are given in the paper. In Caltech101 and Cal-tech256 experiment [1, 2], the training size is 30 images percategory. The experiments are ran on 10 random splits oftrain set and test set. In Stanford Action40 experiment [9],the training size is 100 images per category. Experimentsare ran on 10 random splits of train set and test set. Thereported result is classification accuracy. In UIUC Event8experiment [4], training size is 70 per category and the test-

ing size is 60 images per category. The experiments are ranon 10 random splits of train set and test set.

We plot more class activation maps from Stanford Action40 dataset and SUN397 in Figure 6 and Figure 7.

1.3. Pattern Discovery

Discovering the informative objects in the scenes.The list of 10 scene categories with fully annotations fromSUN397 are bathroom, bedroom, building facade, diningroom, highway, kitchen, living room, mountain snowy,skyscraper, street, totally 4675 images. We train a one-vs-all linear SVM for each class.

Concept localization in weakly labeled images. Thedetail of the concept discovery algorithm is in [10]. Ba-sically for each concept in the text which have enough im-ages, we learn a binary classifier, then we use the weights ofthe classifier to generate the class activation maps for thosetop ranked images detected with this concept.

Weakly supervised text detector. There are 350Google street images in the SVT dataset [7], which aretaken as positive training set. We randomly select 1750 im-ages from SUN attribute dataset [5] as negative training set,then train a linear SVM.

2. Visualizing Class-Specific Units

More class-specific units from the AlexNet*-GAPtrained on ImageNet and Places are plotted in Figure 8.

References[1] L. Fei-Fei, R. Fergus, and P. Perona. Learning generative

visual models from few training examples: An incrementalbayesian approach tested on 101 object categories. ComputerVision and Image Understanding, 2007. 1

[2] G. Griffin, A. Holub, and P. Perona. Caltech-256 object cat-egory dataset. 2007. 1

[3] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags offeatures: Spatial pyramid matching for recognizing naturalscene categories. Proc. CVPR, 2006. 1

[4] L.-J. Li and L. Fei-Fei. What, where and who? classifyingevents by scene and object recognition. Proc. ICCV, 2007. 1

1

Page 2: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Figure 1. Class activation maps generated by GoogLeNet+GAP using the CAM technique. The class categories selected from ILSVRCdataset are goldfish, ostrich, spotted salamander, Norwegian elkhound, Sealyham terrier, fireboat, and football helmet.

Page 3: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

CAM-GoogLeNet CAM-VGG CAM-AlexNet GoogLeNet NIN Backpro AlexNet Backpro GoogLeNet

Figure 2. More examples of the class activation maps generated by several CNN+GAPs from CAM technique, along with the class-specificsaliency maps generated by backpropagation on AlexNet and GoogLeNet.

[5] G. Patterson and J. Hays. Sun attribute database: Discov-ering, annotating, and recognizing scene attributes. Proc.CVPR, 2012. 1

[6] A. Quattoni and A. Torralba. Recognizing indoor scenes.Proc. CVPR, 2009. 1

[7] K. Wang, B. Babenko, and S. Belongie. End-to-end scenetext recognition. Proc. ICCV, 2011. 1

[8] J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torralba.Sun database: Large-scale scene recognition from abbey tozoo. Proc. CVPR, 2010. 1

[9] B. Yao, X. Jiang, A. Khosla, A. L. Lin, L. Guibas, and L. Fei-

Fei. Human action recognition by learning bases of actionattributes and parts. Proc. ICCV, 2011. 1

[10] B. Zhou, V. Jagadeesh, and R. Piramuthu. Conceptlearner:Discovering visual concepts from weakly labeled image col-lections. Proc. CVPR, 2015. 1

Page 4: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Figure 3. Localization results by GoogLeNet+GAP using the CAM technique. Groundtruth bbox and its class annotation is in green, andthe predicted bbox and class label from our method is in red.

Page 5: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Class activation maps from GoogLeNet-GAP

Class activation maps from GoogLeNet-GMP

Figure 4. Class activation maps generated the GoogLeNet-GAP and GoogLeNet-GMP. Class activation map from GoogLeNet-GAP high-lights more complete object regions and less background noise.

Page 6: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Laysan_Albatross

Groove_billed_Ani Rhinoceros_Auklet

Indigo_Bunting

Scissor_tailed_FlycatcherOrchard_Oriole

Sage_ThrasherWhite_Pelican

Figure 5. Class activatiom maps for 8 bird classes from CUB200.

Page 7: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Stanford Action 40applauding

cooking

fishing

gardening

playing_guitar

Figure 6. Examples of the class activation maps for 5 action classes from Stanford Action 40 dataset.

Page 8: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

abbey

bedroom

fountain

SUN 397

landing_deck

nursery

Figure 7. Examples of the class activation maps for 5 scene categoris from SUN397 dataset.

Page 9: Learning Deep Features for Discriminative …cnnlocalization.csail.mit.edu/supp.pdfLearning Deep Features for Discriminative Localization Bolei Zhou, Aditya Khosla, Agata Lapedriza,

Kitchen Restaurant

Ski resort TowerConstruction site

Closet

Car wheel Coffee mug Missile

Chihuahua Fly Rugby ball

Class-specific Units of CNN trained on ImageNet

Class-specific Units of CNN trained on Places Database

Figure 8. Examples of class-specific units from the CNN trained on ImageNet and the CNN trained on Places Database.