+ All Categories
Home > Documents > arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne...

arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne...

Date post: 07-Aug-2020
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
8
Out of the Black Box: Properties of deep neural networks and their applications Nizar Ouarti 1 , David Carmona 1 , 1 IPAL, Sorbonne Universite, CNRS, NUS [email protected], [email protected] Abstract Deep neural networks are powerful machine learn- ing approaches that have exhibited excellent results on many classification tasks. However, they are considered as black boxes and some of their prop- erties remain to be formalized. In the context of image recognition, it is still an arduous task to un- derstand why an image is recognized or not. In this study, we formalize some properties shared by eight state-of-the-art deep neural networks in or- der to grasp the principles allowing a given deep neural network to classify an image. Our results, tested on these eight networks, show that an image can be sub-divided into several regions (patches) responding at different degrees of probability (lo- cal property). With the same patch, some loca- tions in the image can answer two (or three) orders of magnitude higher than other locations (spatial property). Some locations are activators and others inhibitors (activation-inhibition property). The rep- etition of the same patch can increase (or decrease) the probability of recognition of an object (cumula- tive property). Furthermore, we propose a new ap- proach called Deepception that exploits these prop- erties to deceive a deep neural network. We ob- tain for the VGG-VDD-19 neural network a fool- ing ratio of 88%. Thanks to our ”Psychophysics” approach, no prior knowledge on the networks ar- chitectures is required. 1 Introduction A Multilayer Perceptron is a multidimensional universal function approximator [Cybenko, 1989] that learns a map- ping between an input and an output. However, the black box problem stresses that it is ordinarily a challenge to un- derstand what the network exactly learns. One advantage of classical handcrafted approaches is to be able to guarantee a perfectly predictable behavior. This is critical for some areas like medicine or aviation. Deep Neural Networks (DNN) inherit some issues of Mul- tilayer Perceptron and particularly the black box problem. Figure 1: Depiction of the Deepception’s pipeline. It uses all the four properties reported in this paper: the local, spatial, cumulative and activation-inhibition properties. We extract a sub-part of the oppo- nent image called the patch. Then, we insert the patch into the target at positions where it is imperceptible to human eyes. The resulting perturbed image is misclassified by the Deep Neural Network. However, is it possible to obtain a more explainable AI based on DNN? Our motivation is to propose some methods and properties that help users and designers to understand more formally the behavior of DNNs. In this article our main contributions are the followings: We propose a new methodology that is the Psy- chophysics on Deep Neural Networks. We formalize and quantify some important properties of the Deep Neural Networks. We show an extensive experimental comparison illus- trating that the mentioned properties are followed by eight state-of-the-art deep neural networks. We propose a new strategy, called Deepception, based on the formalized properties and capable of fooling a deep neural network, even if its architecture is unknown. 2 Related works Deep neural networks exhibit the best performances for im- age classification with different datasets like ImageNet[Deng et al., 2009; Russakovsky et al., 2015; Krizhevsky et al., 2012; Simonyan and Zisserman, 2014; He et al., 2016] or Pascal VOC for recognition and detection [Ren et al., 2015]. arXiv:1808.04433v1 [cs.CV] 10 Aug 2018
Transcript
Page 1: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

Out of the Black Box: Properties of deep neural networks and their applications

Nizar Ouarti1, David Carmona1,1 IPAL, Sorbonne Universite, CNRS, NUS

[email protected], [email protected]

AbstractDeep neural networks are powerful machine learn-ing approaches that have exhibited excellent resultson many classification tasks. However, they areconsidered as black boxes and some of their prop-erties remain to be formalized. In the context ofimage recognition, it is still an arduous task to un-derstand why an image is recognized or not. Inthis study, we formalize some properties shared byeight state-of-the-art deep neural networks in or-der to grasp the principles allowing a given deepneural network to classify an image. Our results,tested on these eight networks, show that an imagecan be sub-divided into several regions (patches)responding at different degrees of probability (lo-cal property). With the same patch, some loca-tions in the image can answer two (or three) ordersof magnitude higher than other locations (spatialproperty). Some locations are activators and othersinhibitors (activation-inhibition property). The rep-etition of the same patch can increase (or decrease)the probability of recognition of an object (cumula-tive property). Furthermore, we propose a new ap-proach called Deepception that exploits these prop-erties to deceive a deep neural network. We ob-tain for the VGG-VDD-19 neural network a fool-ing ratio of 88%. Thanks to our ”Psychophysics”approach, no prior knowledge on the networks ar-chitectures is required.

1 IntroductionA Multilayer Perceptron is a multidimensional universalfunction approximator [Cybenko, 1989] that learns a map-ping between an input and an output. However, the blackbox problem stresses that it is ordinarily a challenge to un-derstand what the network exactly learns. One advantage ofclassical handcrafted approaches is to be able to guarantee aperfectly predictable behavior. This is critical for some areaslike medicine or aviation.

Deep Neural Networks (DNN) inherit some issues of Mul-tilayer Perceptron and particularly the black box problem.

Figure 1: Depiction of the Deepception’s pipeline. It uses all the fourproperties reported in this paper: the local, spatial, cumulative andactivation-inhibition properties. We extract a sub-part of the oppo-nent image called the patch. Then, we insert the patch into the targetat positions where it is imperceptible to human eyes. The resultingperturbed image is misclassified by the Deep Neural Network.

However, is it possible to obtain a more explainable AI basedon DNN? Our motivation is to propose some methods andproperties that help users and designers to understand moreformally the behavior of DNNs.

In this article our main contributions are the followings:

• We propose a new methodology that is the Psy-chophysics on Deep Neural Networks.

• We formalize and quantify some important properties ofthe Deep Neural Networks.

• We show an extensive experimental comparison illus-trating that the mentioned properties are followed byeight state-of-the-art deep neural networks.

• We propose a new strategy, called Deepception, basedon the formalized properties and capable of fooling adeep neural network, even if its architecture is unknown.

2 Related worksDeep neural networks exhibit the best performances for im-age classification with different datasets like ImageNet[Denget al., 2009; Russakovsky et al., 2015; Krizhevsky et al.,2012; Simonyan and Zisserman, 2014; He et al., 2016] orPascal VOC for recognition and detection [Ren et al., 2015].

arX

iv:1

808.

0443

3v1

[cs

.CV

] 1

0 A

ug 2

018

Page 2: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

However, one of the principal limitations is the lack of un-derstanding of what the networks are really learning. Thisis usually referred as the black box problem. This problemleads to some limitations. One is related to the reliability ofthe probability. As pointed out by Gal and Gharamani (2016)[Gal and Ghahramani, 2016] the standard deep learning toolsdesigned for regression and classification do not effectivelycapture model uncertainty. They argue that the probability,that is the output of the network, cannot be considered as theconfidence of the model. However, for some critical softwarelike autopilot or medical software, it is a requirement to havean accurate estimation of this uncertainty.

Another limitation is related to predictability. It is indeedvery difficult to understand why some objects are recognizedand others are not. To illustrate this issue, Nguyen et al.(2015) [Nguyen et al., 2015] show that some abstract geo-metric patterns could lead the network to recognize an objectnot present in the image. They also show that a periodic pat-tern is sufficient to mislead a deep neural network.

The fooling is also a critical issue. It is shown that witha small magnitude adversarial noise, it is possible to changethe ranking of the classes [Moosavi-Dezfooli et al., 2016b;Moosavi-Dezfooli et al., 2016a; Lin et al., 2017]. Even a ran-dom noise can provoke the same effect [Fawzi et al., 2016].

In the classical scientific approach, scientists are able toquantify and understand some parameters or properties thathave an important impact on the results. Deep neural net-works are more complex and the non-linearity of the pro-cesses seems to prevent any thorough analysis.

Some studies address this question of black box by vi-sualizing the filters of the networks [Zeiler and Fergus,2014]. Other ones visualize and analyze the different lay-ers of the network, to discover the pixel that have a morerelevant impact on the classification [Yosinski et al., 2015;Binder et al., 2016; Nguyen et al., 2016; Samek et al., 2017].These approaches are not using patches and can be consid-ered as neurophysiological approaches for DNN in our for-malism. This paper uses a totally different approach based onmethodological reductionism. The goal is to identify somefundamental properties of deep neural networks. This novelapproach can be compared to Psychophysics [Fechner, 1860;Stevens, 1960; Henn and Young, 1975]. The observation ofthresholds of perception related to physical stimuli is replacedby the monitoring of the probability score of a class related tothe manipulation of some visual inputs. One consequence isthat no prior knowledge about the architecture of the networkis required and the method can be applied to any deep neuralnetwork.

3 Our approachIn this paper, we formalize some properties that have an im-portant impact on DNNs:

• An image can be sub-divided into a group of regionsresponding at different levels of probability. The sizeof the patch is related to the probability. We called it:the Local property.

• When an identical patch is positioned at different loca-tions of an image, its probability varies. We called it: the

Spatial property.

• The repetition of the same patch can increase (or de-crease) the probability of recognition of an object. Thisrepetition can be added at different location of the im-age. We called it: the Cumulative property.

• Cumulative property can have an activator or inhibitorbehavior related to the location of the next patch. Thislocation of activation and inhibition is not related to theinitial spatial probability with one patch. We called it:the Activation-Inhibition property.

4 MethodologyThe psychophysical-driven approach proposed in this paperhighlights the existence of the four properties. Our strategyis to modify a physical aspect of the image inserted into thedeep neural network and observe how its probability variesaccordingly to the modification we made. The main concernof this study is the relationship between the network’s inputand its output. Consequently, the approach presented in thispaper does not require a prior knowledge of the targeted deepneural network architecture.

4.1 Deep neural networks modelsIn order to investigate the generalization of the four proper-ties, we decide to run the experiments on eight state-of-the-art deep neural networks applied to Pascal VOC 2007 andILSVRC 2012 ImageNet datasets.

We consider the following architectures on Pascal VOC2007 three MatConvnet pre-trained networks: Caffenet,VGG16 and VGGM121. We refer to this ensemble ofmodels as ”Fast-RCNN networks”. On ILSVRC 2012ImageNet, we consider five MatConvnet pre-trained net-works: GoogleLeNet, ResNet152, Vgg-f, Vgg-Verydeep-16and Vgg-Verydeep-192. We call this set of models ”ImageNetnetworks” throughout the paper.

4.2 Object categoriesWe choose from Internet the images of objects used in theexperiments. They are all grayscale images. This paper isdealing with two different architectures of deep neural net-works: Pascal VOC 2007 and ILSVRC 2012 ImageNet. Con-sequently, it is important to take objects belonging to two dis-tinct sets of categories. The categories chosen for Pascal are:bird, person, cat, cow, dog, horse, sheep, airplane, bicycle,bus, car, motorbike, train, bottle, chair, dining table, pottedplant, sofa, tv monitor, boat. For ImageNet, we pick a subsetof the 1000 categories of ImageNet. These 20 categories aredisplayed in the Figure 2.

We take a subset because processing the entire set wouldbe computationally expensive. However, we respect the sameobject-type proportions than the Pascal categories: eight cat-egories correspond to biological objects (i.e. bison, daisy orbald eagle) and twelve are manufactured (i.e. airliner, bullet

1Their mean Average Precisions are 57.3%, 67.3% and 59.4%respectively.

2Their top-1 error rates are 34.2%, 23%, 41.1%, 28.5% and28.7% respectively.

Page 3: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

Figure 2: Patches extracted by our system using the VGG-VDD-19 network. The labels underneath correspond to the classes theybelong to. Most of them are not directly recognizable by a human.

train or cab). The patches are extracted with the followingmethod. For the 8 deep neural networks and the 20 selectedcategories, we choose four different grayscale images. Tominimize the effects related to the size of the images, we startby resizing the input images 600x600 pixels if the DNN isa Fast-RCNN network and 224x224 pixels if it is an Ima-geNet networks. We create several sliding cropping windowswhich sizes depend on the tested DNN and do not overlapeach other: 50x50,100x100,150x150, 200x200 pixels If theDNN is a Fast-RCNN network and 37x37, 56x56, 75x75,112x112 pixels if the DNN is an ImageNet one. The deepneural network assigns a probability and a label each croppedpatch. Then, the algorithm selects among all the patches ex-tracted from the four images, the one having the best proba-bility for the category of the object. In the end of the process,we will have in total 20 patches for each DNN. Figure 2 givesan example of extracted patches for the VGG-VDD-19 net-work and their probability.

5 Results

5.1 Local property

A patch is a sub-part of the image. It has two principal char-acteristics: a size and a probability belonging to a specific cat-egory. Different patches have different probabilities. This isillustrated in Figure 2 were we display after an exhaustive re-search the patches that answered with the highest probability.A second observation is that the probability for a given patchis not scale invariant (see Figure 3). Moreover this probabil-ity is very different if the patch is resized before being sent tothe network (Figure 3) with a range between 5 and 99 percentor if the patch is not resized and sent with all the other pixelput at zero (Figure 4) with a range between 0.6 and 1.5 10−4.

Figure 3 shows how the average probability of the patchesvary according to their sizes. Bigger patches have higherprobabilities, however it is still possible with a relativelysmall patch to obtain an important probability. For the rest ofthe experiments, we decided to take patches of size 150x150and 56x56 pixels for the Fast-RCNN networks and ImageNetrespectively. We chose these sizes because they are challeng-ing to recognize by humans, and at the same time, correctlydetected by the DNNs.

100x100 150x150 200x200 300x3000

10

20

30

40

50

60

70

80

90

Patch size (pixels)

Prob

abili

ty(%

)

frcnn-caffenetfrcnn-vgg16

frcnn-vggm12

(a) Fast-RCNN CNNs

37x37 56x56 75x75 112x1120

10

20

30

40

50

60

70

80

90

100

Patch size (pixels)

Prob

abili

ty(%

)

vgg-fgooglelenetvgg-vdd-19vgg-vdd-16resnet-152

(b) ImageNet CNNs

Figure 3: Average probabilities of the patches for each deep neuralnetwork for different scales.

5.2 Spatial propertyTo demonstrate whether DNNs answer differently for dif-ferent location of the patch inside an image, we create ablack image (i.e. all the pixels have values equal to zero)of size 600x600 pixels (if we use a Faster-RCNN network) or224x224 pixels (if we are using an ImageNet network). Then,we add in a pixel-wise fashion the patch by the correspondingregion of the black image where it should be positioned. Weiterate until we reach the end of the image. We decided toperform on one image a dense mapping to observe the effectof location on the probability. The patch is a 150X150 patchof a cat detected initially at 99% by the Fast-RCNN-VGG16network. This procedure takes a long time and cannot be per-formed on many images. Figure 4 gives an interesting eval-uation of how extreme the differences of probability can be.With the same patch, the range of probability was from 0.6and 1.5 10−4. The ratio between the maximum and minimumis 3769 (i.e. the maximum probability is 3769 times higherthan the minimum probability).

Figure 4: Evolution of the probability according to a (150X150)patch position on the 600x600 black image. The high ratio(Probamax/Probamin),3769,indicates that Fast-RCNN-VGG16 issensitive to the spatial property.

To be more exhaustive and to test more networks, we de-cided to run this experiment on many images and with manypatches but less densely (no overlap of the patch), i.e. 16 po-sitions for a 150X150 patch on a 600X600 image. The max-min ratio for different networks can be observed in Table 1.

For some categories, deep neural networks are very sen-sitive to the spatial property. Notice that a state-of-the-artneural network as ResNet-152 can be extremely sensitive tothe spatial property. For the ’moped’ category, the maximumprobability is 10000 higher than the minimum.

Page 4: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

Deep neural network Avg. Ratio Avg. Maximum Avg. Minimum

Fast-RCNN-CaffeNet 156.98 0.0047 0.00003

Fast-RCNN-VGG16 572.35 0.0007 0.000001

Fast-RCNN-VGGM12 1237.35 0.0005 0.0000004

Vgg-f 20.15 0.0003 0.00002

GoogleLeNet 60.11 0.0017 0.00003

Vgg-VDD-19 208.54 0.0023 0.00001

Vgg-VDD-16 161.57 0.0034 0.00002

ResNet-152 2262.53 0.0115 0.000005

Table 1: Averages of the ratios and gains for the spatial, activationand inhibition experiments.

Using this procedure, we observe that all the ratios betweenthe region that maximize and minimize the patch probabilityare always higher than 1.

We report in Table 1 the average ratios for all the deep neu-ral networks. We can notice that there are different degrees ofsensitivity to the spatial property. However, the most sensitiveDNN to the spatial property is ResNet-152. In that case, themaximum probability is 2262 times higher than the lowest.

5.3 Cumulative Property andActivation-Inhibition Property

In this section, the aim is to show that the addition of thesame patch many times will change the probability of detec-tion. We decided to work with a black image of size 600x600or 224x224 pixels accordingly to the used deep neural net-work. Then, the image is divided into 16 non-overlappingareas of dimensions 150x150 or 56x56 pixels each. We be-gin by placing the patch inside the area that maximizes theprobability. Afterwards, we look for an additional patch po-sition that increases the probability and add the patch again.We keep performing this operation while the probability is in-creasing after a new placement. Because we wanted to showthe Activation-Inhibition effect, we apply the same algorithmfor inhibition. Obviously, instead of looking for the areas thatincrease the overall probability, we search for the areas mak-ing the overall probability decrease.

The gain is defined as the relation between the first bestpositioning and the final probability: Gain.Probinit =Probfinal. It is superior to 1 for activation and inferior to1 for inhibition. Figure 5 reports the gains for all the testedcategories and DNNs. For both types of neural networks,the gains are always higher than 1. This means that the finalprobability will always be increased after a sequence of patchplacement compared to one single position. Furthermore, it isalso possible to decrease the probability of the class. Figure5 proves the probability of placing multiple patches can de-crease the probability. We show with these experiments that itis possible to change the probability of recognition by addingnew patches (cumulative property). But we also show that anoriented strategy of placement of the patch can increase ordecrease the probability (activation-inhibition property).

5.4 Deceiving a deep neural network: DeepceptionTo show a practical usage of the properties we highlighted,we decided to design a new type of fooling algorithm called

Deepception. This algorithm has the particularity to be in-dependent from the architecture, thanks to our psychophysicapproach. We need simply an access to the input (image)and the output of the network (probability of recognition) todeceive the DNN. It means that our approach can even de-ceive an unknown DNN that is on a server. The idea is thatwe can pick some local patches (local property) that haveby themselves a high probability of recognition for a givenclass. And we insert these patches in a targeted image (thatbelongs to another class) to fool the DNN (see Figure 1). Wedesign a specific cost function (equation 1) that encouragesthe fooling by estimating the probability for different spa-tial insertions. Here, we take advantage of the cumulativeand activation-inhibition property. The inserted normalizedpatches are made transparent to not be perceptible by humans.In order to do so, the patch is multiplied by transparency coef-ficient τ . This resulting patch, called the decoy, is multipliedwith the weaker RGB channel. The cost function is:

ArgminN,L∈N

(Pt), stopping criterion Pt > Pi (1)

Where N represents the number of patches and L the locationof the patches, Pt represents the probability of the targetedclass. This targeted class is the initial class of the image. Pi

is the probability of each of the other classes.We report the results of Deepception on a subset of Ima-

geNet consisting of 100 randomly selected images3. We limitthe experiment to 100 randomly selected images to obtainfair results that can be computed in a reasonable amount oftime. This was inspired by recent approaches [Metzen et al.,2017]. We take the patches from figure 2 and generate the de-coys displayed in Figure 7. These decoys are not perceptibleby humans. We decided to observe the capacity of foolingof Deepception with 20 decoys (Figure 6). We observed alinear relation between standard deviation of the decoys andtheir fooling abilities. For this reason we designed 2 differentgaussian noises (std=100 and 150). And we observed that agaussian noise alone cannot provide the same fooling as ourdecoys.

20 40 60 80 100 120 140 160

Standard deviation

12

14

16

18

20

22

24

Numb

er of

foole

d ima

ges

Images fooled by one opponentLinear regression curveImages fooled by random noise

Figure 6: Total number of fooled images versus the standard devi-ations of the decoys. The transparency is equal to 4. Red cross arethe gaussian noise.

3Images were selected by randomly drawing ILSVRC2012 im-ages (i.e. integers from [1, 100]), using the randperm function ofthe scientific computing environment Matlab after initializing Mat-lab random number generator seed with 0.

Page 5: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

aeroplane

bicycle

bird

boat

bottle

bus

car

cat

chair

cow

diningtable

dog

horse

motorbike

person

pottedplant

sheep

sofa

train

tvmonito

r0

200

400

600

800

1,000

1,200

1,400

Category

Gain

frcnn-caffenetfrcnn-vgg16

frcnn-vggm12

(a) Activation Fast-RCNN

aeroplane

bicycle

bird

boat

bottle

bus

car

cat

chair

cow

diningtable

dog

horse

motorbike

person

pottedplant

sheep

sofa

train

tvmonito

r0

0.2

0.4

0.6

0.8

1

CategoryG

ain

frcnn-caffenetfrcnn-vgg16

frcnn-vggm12

(b) Inhibition Fast-RCNN

airliner

bald

eagle

ballp

layer

beer

bottle

bison

bulle

t train ca

bdaisy

dining

table

foldingchair

German

shepherd

llama

mon

itor

mop

ed

mou

ntainbike

sorre

l

speedb

oat

studiocouchtabb

y

trolle

ybus

0

100

200

300

400

500

Category

Gain

vgg-fgooglelenetvgg-vdd-19vgg-vdd-16resnet-152

(c) Activation ImageNet

airliner

bald

eagle

ballp

layer

beer

bottle

bison

bulle

t train ca

bdaisy

dining

table

foldingchair

German

shepherd

llama

mon

itor

mop

ed

mou

ntainbike

sorre

l

speedb

oat

studiocouchtabb

y

trolle

ybus

0

0.2

0.4

0.6

0.8

1

Category

Gain

vgg-fgooglelenetvgg-vdd-19vgg-vdd-16resnet-152

(d) Inhibition ImageNet

Figure 5: Gains for the activation and inhibition experiments. All the gains are higher than one in the case of activation and lower than one inthe case of inhibition. Consequently, it means that all the deep neural networks follow both properties independently of the patch.

Figure 7: Decoys generated from the patches exposed in Figure 2

1 2 4 8 160

10

20

30

40

50

60

70

80

Transparency (τ)

Fool

ing

ratio

(%)

(a)

4x4 8x8 16x160

5

10

15

20

25

Grid size

Num

bero

focc

urre

nces

(b)

Figure 8: (a) Fooling ratio on the validation set versus the decoytransparency. (b) Number of images fooled by positioning a singlepatch for different grid sizes.

We decided to apply our Deepception approach to a pre-trained VGG-VDD-19 DNN. The mountain-bike patch ischosen as a decoy because its exhibit high performance (highprobability and high standard deviation).

Influence of different parameters of DeepceptionFirstly, we fix the grid size to 4x4 and test 5 transparencylevels to see how the fooling ratio varies. Figure 8.a demon-strates the higher the transparency gets, the lower the foolingratio will be.

Transparency coefficient τ = 4 seems to be the optimalvalue for the 100 validation images. At this level of trans-parency, the decoy is unrecognizable to a human eye. Fur-thermore, the fooling ratio of 30% for a 4x4 grid is accept-

0 10 20 30 40 50 600

20

40

60

80

Number of decoys

Cum

ulat

ive

num

bero

ffoo

lings

4x48x8

16x16

(a)

0 1 2 3 4 5

10

20

30

40

50

60

70

80

90

100

Number of foolings

Aver

agep

roba

bilit

y(%

)

(b)

Figure 9: (a)Total number of fooled target images versus the numberof inserted decoys. (b) Average of the targets’ initial probabilitiesversus the number of foolings.

able. Consequently, we decide to fix τ = 4.We want to study how many images are fooled with the

first decoy placement. Figure 8.b shows the number of im-ages which have been fooled only by one decoy. Then, we in-vestigate whether there is a relation between the initial proba-bility of a target image and the number of times it gets fooledapplying τ = 4 . Figure 9.b shows this relation actually ex-ists between both variables. Another important factor is howthe number of decoys inserted inside the target, affects theperformances. Figure 9.a reports these results for differentgrid sizes. The higher the number of inserted decoys gets, thehigher the number of fooled image will be.

5.5 Comparison between Deepception andUniversal Adversarial Perturbation

With a better understanding of the important parametersfor Deepception, we decided to compare our results with a

Page 6: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

Figure 10: Examples of perturbed images and the categories theybelong to. Top row: Universal Adversarial Perturbation. Bottomrow: our approach, Deepception. Zoom for a better visualization.

Universal Adversarial Perturbation [Moosavi-Dezfooli et al.,2017].

Some examples of fooled images by Deepception are illus-trated in Figure 10. The Top row illustrates Universal Adver-sarial Perturbation [Moosavi-Dezfooli et al., 2017] and theBottom row is Deepception, our approach. It can be seen thatour approach is even less visible than the Universal Adversar-ial Perturbation.

We choose VGG-VDD-19 as the targeted deep neural net-work for the comparison. Therefore, we download from theUniversal Adversarial Perturbations authors’ online reposi-tory the pre-computed perturbation for this network ( 10,000images of the ILSVRC 2012 [Russakovsky et al., 2015] train-ing set). Afterwards, we randomly select 100 images from theILSVRC 2012 validation dataset and apply Deepception andthe Universal Perturbation.

For the 100 randomly selected images, the fooling ratiosof Deepception and Universal Adversarial Perturbations are88% and 75% respectively. Figure 10 allows a visual compar-ison between the outputs of both methods. It is worthy notingthat contrary to Universal Adversarial Perturbation, our ap-proach does not need to be trained on a specific network.

6 DiscussionThis paper, formalized and analyzed some properties ofDNNs. One of these properties can be considered as logi-cal inferences of former studies. For instance, it is possible topartially infer the local property, based on related studies ondeconvolutional networks [Noh et al., 2015], or visualization[Durand et al., 2017], showing that some specific locationsof an image respond better than others. But these works aremainly pixel based and did not study the effect of the patchsize as we did. Other properties are more challenging to in-tuit. For instance, it is important to note that the spatial prop-erty reported here, is not related to an interaction between thepatch and the content of the image because we show the effectin an image filled by zeros, excluding the patch. This propertyhighlights that DNNs are not totally translation invariant.

However, the cumulative property allows to reinterpret thework of Nguyen et al. (2014) [Nguyen et al., 2015] and thereason why repeating a pattern from an object inside an im-age, increases its probability of being detected. Moreover,this paper exposes that repeating the pattern is not a sufficientcondition to increase its probability of detection. In fact, itmust be repeated at very specific locations of the image: theactivating positions. Otherwise, the probability of detectingthe object will not increase. Consequently, we highlighteda phenomenon that, at the best of our knowledge, no paperhas reported yet: the activating-inhibitory property. The val-idation of these properties on Pascal VOC07 and ImageNetwith different types of network proves that they can be gen-eralized. We also showed that contrary to a repetitive pattern,activation and inhibition are not mandatorily contiguous. Thepatch can benefit of the cumulative effect even with a sparsespatial distribution.

This work also provides a new way of measuring the per-formance of a DNN: the ratio between minimum and max-imum probability (for the same patch). The ratio shouldbe equal to 1 to have a perfect translation-invariant net-works. This measure can be optimized in the future to ob-tain more robust networks. We can observe that DNNs likeFast-RCNN-VGG16 or Resnet, which have proved to havestate-of-the-art performances on the Pascal VOC07 and Ima-geNet datasets, are sensitive to patch translations. Their highratio is meaning there is a big gap of probability between twodifferent positions inside the image.

Some possible interpretation of the spatial property can bedone based on ”DNN neurophysiologist” studies of Bau etal. [Bau et al., 2017] and Zhang et al. [Zhang et al., 2017].The authors studied the activation of some high-level seman-tic units based on their receptive field. Our local property canbe interpreted as the tendency of a given network to not beable to obtain many receptive fields that will cover uniformlythe image. A conjoint framework combining psychophysicsand neurophysiologist approach remains to be developed buthas a great potential.

The second part of this study took advantage of the exposedproperties. We proposed a new fooling approach called Deep-ception. Unlike Universal Adversarial Perturbation (UAP),our approach uses structured patches from another class thatare hidden in the image and are able to fool the network. Weshowed a fooling ratio of 88% on the VGG-VDD-19 networkcompared to the 85% obtained by UAP on the same network[Moosavi-Dezfooli et al., 2017].

An fundamental difference of our approach compared toclassical approaches is that we do not need to have a priorknowledge of the network’s architecture. We simply need tobe able to send an input and an access to the probability. An-other difference is also the sparsity. In many cases, we canfool the network only with 1 patch (it represents 1/16 th ofthe image). The advantage is that very localized patch couldbe more difficult to perceive by a human than a widespreadfooling noise. We also tested that our effect is stronger witha structured patch rather than using a simple Gaussian noise.And even if we show that variance impacts the capacity offooling the target, a Gaussian noise with high variance is notbetter for fooling compared with a decoy selected with our

Page 7: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

technique. We observed also that the same decoy can be uti-lized to fool different images. Our approach exhibits a goodgenerality. Indeed, when a decoy is selected with a given net-work, it can be employed to fool many other images of thesame network.

7 ConclusionIn this article we used a ”Psychophysic” approach applied toArtificial Intelligence in the realm of ”Neurophysiologists”.We did not study the internal architectures of the networks,but we made some deductions by modifying the input and an-alyzing the resulting probability. We think this approach canhave some benefit and we propose some properties allowingto rank the networks and explaining their high performances.With a practical application, a software called Deepception,we demonstrated that the properties analyzed in this work canhelp designing new methods for deep neural networks’ fool-ing.

References[Bau et al., 2017] David Bau, Bolei Zhou, Aditya Khosla, Aude

Oliva, and Antonio Torralba. Network dissection: Quantifyinginterpretability of deep visual representations. arXiv preprintarXiv:1704.05796, 2017.

[Binder et al., 2016] Alexander Binder, W Samek, G Montavon,S Bach, and KR Muller. Analyzing and validating neural net-works predictions. In Proceedings of the Workshop on Visualiza-tion for Deep Learning at International Conference on MachineLearning (ICML), 2016.

[Cybenko, 1989] George Cybenko. Approximation by superposi-tions of a sigmoidal function. Mathematics of Control, Signals,and Systems (MCSS), 2(4):303–314, 1989.

[Deng et al., 2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li,Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical im-age database. In Computer Vision and Pattern Recognition, 2009.CVPR 2009. IEEE Conference on, pages 248–255. IEEE, 2009.

[Durand et al., 2017] Thibaut Durand, Taylor Mordan, NicolasThome, and Matthieu Cord. Wildcat: Weakly supervised learningof deep convnets for image classification, pointwise localizationand segmentation. In IEEE Conference on Computer Vision andPattern Recognition (CVPR 2017), 2017.

[Fawzi et al., 2016] Alhussein Fawzi, Seyed-Mohsen Moosavi-Dezfooli, and Pascal Frossard. Robustness of classifiers: fromadversarial to random noise. In Advances in Neural InformationProcessing Systems, pages 1632–1640, 2016.

[Fechner, 1860] GT Fechner. Elemente der psychophysik (bre-itkopf und hartel, leipzig). Translation:(1966) Elements of Psy-chophysics, 1, 1860.

[Gal and Ghahramani, 2016] Yarin Gal and Zoubin Ghahramani.Dropout as a bayesian approximation: Representing model un-certainty in deep learning. In international conference on ma-chine learning, pages 1050–1059, 2016.

[He et al., 2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, andJian Sun. Deep residual learning for image recognition. In Pro-ceedings of the IEEE conference on computer vision and patternrecognition, pages 770–778, 2016.

[Henn and Young, 1975] Volker Henn and Laurence R Young.Ernst mach on the vestibular organ 100 years ago. ORL,37(3):138–148, 1975.

[Krizhevsky et al., 2012] Alex Krizhevsky, Ilya Sutskever, and Ge-offrey E Hinton. Imagenet classification with deep convolutionalneural networks. In Advances in neural information processingsystems, pages 1097–1105, 2012.

[Lin et al., 2017] Yen-Chen Lin, Zhang-Wei Hong, Yuan-HongLiao, Meng-Li Shih, Ming-Yu Liu, and Min Sun. Tactics ofadversarial attack on deep reinforcement learning agents. arXivpreprint arXiv:1703.06748, 2017.

[Metzen et al., 2017] Jan Hendrik Metzen, Tim Genewein, VolkerFischer, and Bastian Bischoff. On detecting adversarial perturba-tions. arXiv preprint arXiv:1702.04267, 2017.

[Moosavi-Dezfooli et al., 2016a] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, Omar Fawzi, and PascalFrossard. Universal adversarial perturbations. arXiv preprintarXiv:1610.08401, 2016.

[Moosavi-Dezfooli et al., 2016b] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, and Pascal Frossard. Deepfool:a simple and accurate method to fool deep neural networks. InProceedings of the IEEE Conference on Computer Vision andPattern Recognition, pages 2574–2582, 2016.

[Moosavi-Dezfooli et al., 2017] Seyed-Mohsen Moosavi-Dezfooli,Alhussein Fawzi, Omar Fawzi, and Pascal Frossard. Universaladversarial perturbations. In Computer Vision and Pattern Recog-nition (CVPR), 2017 IEEE Conference on, pages 86–94. IEEE,2017.

[Nguyen et al., 2015] Anh Nguyen, Jason Yosinski, and Jeff Clune.Deep neural networks are easily fooled: High confidence pre-dictions for unrecognizable images. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition, pages427–436, 2015.

[Nguyen et al., 2016] Anh Nguyen, Jason Yosinski, and Jeff Clune.Multifaceted feature visualization: Uncovering the differenttypes of features learned by each neuron in deep neural networks.arXiv preprint arXiv:1602.03616, 2016.

[Noh et al., 2015] Hyeonwoo Noh, Seunghoon Hong, and BohyungHan. Learning deconvolution network for semantic segmenta-tion. In Proceedings of the IEEE International Conference onComputer Vision, pages 1520–1528, 2015.

[Ren et al., 2015] Shaoqing Ren, Kaiming He, Ross Girshick, andJian Sun. Faster r-cnn: Towards real-time object detection withregion proposal networks. In Advances in neural informationprocessing systems, pages 91–99, 2015.

[Russakovsky et al., 2015] Olga Russakovsky, Jia Deng, Hao Su,Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang,Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexan-der C. Berg, and Li Fei-Fei. ImageNet Large Scale VisualRecognition Challenge. International Journal of Computer Vi-sion (IJCV), 115(3):211–252, 2015.

[Samek et al., 2017] Wojciech Samek, Alexander Binder, GregoireMontavon, Sebastian Lapuschkin, and Klaus-Robert Muller.Evaluating the visualization of what a deep neural network haslearned. IEEE transactions on neural networks and learning sys-tems, 2017.

[Simonyan and Zisserman, 2014] Karen Simonyan and AndrewZisserman. Very deep convolutional networks for large-scale im-age recognition. arXiv preprint arXiv:1409.1556, 2014.

[Stevens, 1960] Stanley S Stevens. The psychophysics of sensoryfunction. American scientist, 48(2):226–253, 1960.

Page 8: arXiv:1808.04433v1 [cs.CV] 10 Aug 2018 · Nizar Ouarti 1, David Carmona , 1 IPAL, Sorbonne Universite, CNRS, NUS nizar.ouarti@ipal.cnrs.fr, david.carmona93@gmail.com Abstract ...

[Yosinski et al., 2015] Jason Yosinski, Jeff Clune, Anh Nguyen,Thomas Fuchs, and Hod Lipson. Understanding neural networksthrough deep visualization. arXiv preprint arXiv:1506.06579,2015.

[Zeiler and Fergus, 2014] Matthew D Zeiler and Rob Fergus. Visu-alizing and understanding convolutional networks. In Europeanconference on computer vision, pages 818–833. Springer, 2014.

[Zhang et al., 2017] Quanshi Zhang, Ying Nian Wu, and Song-Chun Zhu. Interpretable convolutional neural networks. arXivpreprint arXiv:1710.00935, 2(3):5, 2017.


Recommended