Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup,...

4
Supplementary Materials: Towards Global Explanations of Convolutional Neural Networks with Concept Attribution Weibin Wu 1 , Yuxin Su 1* , Xixian Chen 2 , Shenglin Zhao 2 , Irwin King 1 , Michael R. Lyu 1 , Yu-Wing Tai 2 1 Department of Computer Science and Engineering, The Chinese University of Hong Kong, 2 Tencent {wbwu, yxsu, king, lyu}@cse.cuhk.edu.hk, {xixianchen, henryslzhao, yuwingtai}@tencent.com 1. More Attacking Results We report the intermediate attacking results to confirm the desired property of global feature occluders: intra-class generalization and specificity. That is, a global feature oc- cluder should disrupt decisive feature filters common to the specified category of images. Since directly examining zil- lions of internal neurons is prohibitively expensive, we re- sort to the solutions as follows. 1.1. Intra-class Generalization This property means that a global feature occluder can fool model predictions on samples of the target class, in- cluding those that the global feature occluder has never seen during training. Intra-class generalization reflects that global feature occluders can disturb feature filters common to the target class. To examine this property, for a target class, we first occlude clean ImageNet images with the cor- responding global feature occluder. Then we investigate the model performance on the resultant samples. Table S1 reports the average results over the 100 ran- dom classes covered in the main paper. We view the Ima- geNet validation set as the test set. Although global feature occluders have never seen the test set of ImageNet during training, there is a significant degradation of model accu- racy on both the training and test set after the employment of matched global feature occluders. It corroborates that the learned global feature occluders possess a strong capacity for intra-class generalization. 1.2. Specificity This property requires that a global feature occluder can- not severely mislead model decisions on samples from the other classes that it does not target. Specificity confirms that global feature occluders can undermine feature filters specific to the target class. To investigate this property, we still focus on the same 100 classes we explore before. Con- * Corresponding author. cretely, for a global feature occluder, we apply it to the in- stances from the remaining 99 categories that it does not initially target. Then we record the model accuracy on the resultant images. Table S2 summarizes the average results over the 100 classes. Global feature occluders witness a much-limited attack success rate under this setup, which satisfies our expectation of learning a class-specific global feature oc- cluder. Besides, there is an observable degree of vulner- ability of models to unmatched global feature occluders. It may be because similar image categories are likely to share some critical feature detectors. Worse still, ImageNet often requires fine-grained classification between similar subspecies of an animal [S3], such as great white sharks and tiger sharks, which can aggravate the susceptibility of trained classifiers to global feature occluders. To further illustrate the specificity of derived global fea- ture occluders, we exhibit examples of learned global fea- ture occluders, along with correctly classified legitimate samples and their distorted counterparts in Figure S1. Re- markably, global feature occluders are not identical for dif- ferent classes. Besides, the misclassifications incurred by global feature occluders do not converge to the same in- correct label. Therefore, it again shows that global feature occluders are class-specific. From Figure S1, we can also discover that model deci- sions often deviate a lot from the original ones when en- countering global feature occluders, while the perturbation can hardly harm human perception. It confirms that feature occluders do not need to degenerate images in a human- notable fashion, or align with meaningful image regions to humans. This observation is consistent with the finding in the literature that in contrast to human visual systems, the building blocks of model reasoning, namely, the feature de- tectors, are susceptible to structured noises [S5, S4, S2, S1].

Transcript of Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup,...

Page 1: Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup, which satisfies our expectation of learning a class-specific global feature oc-cluder.

Supplementary Materials:Towards Global Explanations of Convolutional Neural Networks with Concept

Attribution

Weibin Wu1, Yuxin Su1∗, Xixian Chen2, Shenglin Zhao2, Irwin King1, Michael R. Lyu1, Yu-Wing Tai21Department of Computer Science and Engineering, The Chinese University of Hong Kong, 2Tencent{wbwu, yxsu, king, lyu}@cse.cuhk.edu.hk, {xixianchen, henryslzhao, yuwingtai}@tencent.com

1. More Attacking ResultsWe report the intermediate attacking results to confirm

the desired property of global feature occluders: intra-classgeneralization and specificity. That is, a global feature oc-cluder should disrupt decisive feature filters common to thespecified category of images. Since directly examining zil-lions of internal neurons is prohibitively expensive, we re-sort to the solutions as follows.

1.1. Intra-class Generalization

This property means that a global feature occluder canfool model predictions on samples of the target class, in-cluding those that the global feature occluder has neverseen during training. Intra-class generalization reflects thatglobal feature occluders can disturb feature filters commonto the target class. To examine this property, for a targetclass, we first occlude clean ImageNet images with the cor-responding global feature occluder. Then we investigate themodel performance on the resultant samples.

Table S1 reports the average results over the 100 ran-dom classes covered in the main paper. We view the Ima-geNet validation set as the test set. Although global featureoccluders have never seen the test set of ImageNet duringtraining, there is a significant degradation of model accu-racy on both the training and test set after the employmentof matched global feature occluders. It corroborates that thelearned global feature occluders possess a strong capacityfor intra-class generalization.

1.2. Specificity

This property requires that a global feature occluder can-not severely mislead model decisions on samples from theother classes that it does not target. Specificity confirmsthat global feature occluders can undermine feature filtersspecific to the target class. To investigate this property, westill focus on the same 100 classes we explore before. Con-

∗Corresponding author.

cretely, for a global feature occluder, we apply it to the in-stances from the remaining 99 categories that it does notinitially target. Then we record the model accuracy on theresultant images.

Table S2 summarizes the average results over the 100classes. Global feature occluders witness a much-limitedattack success rate under this setup, which satisfies ourexpectation of learning a class-specific global feature oc-cluder. Besides, there is an observable degree of vulner-ability of models to unmatched global feature occluders.It may be because similar image categories are likely toshare some critical feature detectors. Worse still, ImageNetoften requires fine-grained classification between similarsubspecies of an animal [S3], such as great white sharksand tiger sharks, which can aggravate the susceptibility oftrained classifiers to global feature occluders.

To further illustrate the specificity of derived global fea-ture occluders, we exhibit examples of learned global fea-ture occluders, along with correctly classified legitimatesamples and their distorted counterparts in Figure S1. Re-markably, global feature occluders are not identical for dif-ferent classes. Besides, the misclassifications incurred byglobal feature occluders do not converge to the same in-correct label. Therefore, it again shows that global featureoccluders are class-specific.

From Figure S1, we can also discover that model deci-sions often deviate a lot from the original ones when en-countering global feature occluders, while the perturbationcan hardly harm human perception. It confirms that featureoccluders do not need to degenerate images in a human-notable fashion, or align with meaningful image regions tohumans. This observation is consistent with the finding inthe literature that in contrast to human visual systems, thebuilding blocks of model reasoning, namely, the feature de-tectors, are susceptible to structured noises [S5, S4, S2, S1].

Page 2: Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup, which satisfies our expectation of learning a class-specific global feature oc-cluder.

Top-1 Training

Accuracy

Top-1 Test

Accuracy

Top-1 Training

Accuracy

Top-1 Test

Accuracy

ResNet-50 0.8771 0.7756 0.0973 0.1662

GoogLeNet 0.8115 0.7494 0.0907 0.2606

VGG-16 0.8095 0.7358 0.1001 0.1616

Clean Perturbed

Model

Table S1: Average top-1 accuracy of different models on clean images and the counterparts perturbed with matched globalfeature occluders. A global feature occluder can successfully fool model predictions on both the training and unseen test dataof the target class, which confirms its intra-class generalization capacity.

Top-1 Training

Accuracy

Top-1 Test

Accuracy

Top-1 Training

Accuracy

Top-1 Test

Accuracy

ResNet-50 0.8771 0.7756 0.6063 0.5586

GoogLeNet 0.8115 0.7494 0.5893 0.5662

VGG-16 0.8095 0.7358 0.5392 0.5035

Model

Clean Perturbed

Table S2: Average top-1 accuracy of different models on clean images and the counterparts perturbed with unmatched globalfeature occluders. Global feature occluders exhibit much-limited attack success rates under this scenario, which reflects thatthey are class-specific.

References[S1] Nicholas Carlini and David Wagner. Towards evaluating the

robustness of neural networks. In 2017 IEEE Symposium onSecurity and Privacy (SP), pages 39–57. IEEE, 2017. 1

[S2] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, OmarFawzi, and Pascal Frossard. Universal adversarial perturba-tions. In Conference on Computer Vision and Pattern Recog-nition (CVPR), pages 1765–1773, 2017. 1

[S3] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, San-jeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy,Aditya Khosla, Michael Bernstein, et al. ImageNet Large

Scale Visual Recognition Challenge. International Journalof Computer Vision, 115(3):211–252, 2015. 1, 3, 4

[S4] Sara Sabour, Yanshuai Cao, Fartash Faghri, and David JFleet. Adversarial manipulation of deep representations.In International Conference on Learning Representations(ICLR), 2016. 1

[S5] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, JoanBruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. In-triguing properties of neural networks. In International Con-ference on Learning Representations (ICLR), 2014. 1

Page 3: Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup, which satisfies our expectation of learning a class-specific global feature oc-cluder.

Fig

Bonnet Paper towel Crab Shark

Grey whale

Bull mastiff

Isopod

Curtain

Global Feature

Occluder

Clean

Perturbed

(a) ResNet-50

Fig

Jak Husky Turtle Iron

Grey whale

Coat

Isopod

Bulldog

Global Feature

Occluder

Clean

Perturbed

(b) GoogLeNet

Figure S1: Examples of learned global feature occluders and their effects on model decisions. In every two columns ofone model, the images from the top row to the bottom one are the learned global feature occluder for a target class (withpixel values magnified to natural image ranges for visibility), the clean images of this class from the ILSVRC 2012 trainingset [S3], and the resultant images distorted by the global feature occluder of this class, respectively. The labels above theseimages report the model predictions. All model decisions on clean images are correct. We see that: (1) global featureoccluders are not identical for different classes, and (2) the misclassifications on perturbed images do not converge to thesame label. These observations further validate that global feature occluders are class-specific.

Page 4: Supplementary Materials: Towards Global Explanations of ......attack success rate under this setup, which satisfies our expectation of learning a class-specific global feature oc-cluder.

Fig

Chain mail Monkey Hippo Stingray

Grey whale

Sock

Isopod

Sombrero

Global Feature

Occluder

Clean

Perturbed

(c) VGG-16

Figure S1: (Continued) Examples of learned global feature occluders and their effects on model decisions. In every twocolumns of one model, the images from the top row to the bottom one are the learned global feature occluder for a targetclass (with pixel values magnified to natural image ranges for visibility), the clean images of this class from the ILSVRC2012 training set [S3], and the resultant images distorted by the global feature occluder of this class, respectively. The labelsabove these images report the model predictions. All model decisions on clean images are correct. We see that: (1) globalfeature occluders are not identical for different classes, and (2) the misclassifications on perturbed images do not converge tothe same label. These observations further validate that global feature occluders are class-specific.