Notice
Recent Posts
Recent Comments
Link
«   2024/05   »
1 2 3 4
5 6 7 8 9 10 11
12 13 14 15 16 17 18
19 20 21 22 23 24 25
26 27 28 29 30 31
Tags more
Archives
Today
Total
관리 메뉴

러닝머신 하는 K-공대생

[XAI Reading Log] The All Convolutional Net #1 본문

Machine Learning

[XAI Reading Log] The All Convolutional Net #1

prgmti1 2021. 9. 9. 21:57

Date: 2021/09/09

Title: Striving for Simplicity: The All Convolutional Net #1


- Reasons why I read this paper: 

Reading a paper explains Layer-wise Relevance Propagation last time, I had have curiosities about backward processs in neural networks explain or visuallize black boxes of neural networks. Also I have experience on visuallizing Convolutional Neural Network using Guided Backpropagation without deep understanding about it. Visuallizing AI is not XAI but they're inter-related. So I choosed this paper to understand Guided Backpropagation and releated theory because I was judged having basic knowledge to read this. 

 

- Today's Reading Purpose

My purpose is understanding the processes of replacement pooling layer with convolution layer. 

- Summary:

This paper says that CNN model which has without complex activation function, max-pooling operations or response normalization can have enough good perforamance. Furthur, they suggest Guided-Backpropagation as a new visuallization method of CNN and compare the before(ex. deconvnet). 

p-norm subsampling(pooling)
convolution with stride 2

- About pooling layer :

Through pooling layers, we can make features more invariant, easy for optimization and accept large sizes of image as a input. But there is a disadvantage that some feature informations can be vanished after pooling layer which reduces spatial dimensionality.

 

- Replacement pooling layer with convolution layer :

  • We can remove each pooling layer and increase the stride of the convolutional layer that preceded it accordingly.
  • We can replace the pooling layer by a normal convolution with stride larger than one (i.e. for a pooling layer with k = 3 and r = 2 we replace it with a convolution layer with corresponding stride and kernel size and number of output channels equal to the number of input channels)

The three base networks used for classification on CIFAR-10 and CIFAR-100.
Three diffrent ways of replacement pooling layers with convolution layers derived from Model C above.

- Classification Results :


My thought: 

I was suprised by pretty good results of the alternative model (pool->conv) which has simple architectures. I thought if it could imporve the performance but it doesn't improve performance of CNNs. Interesting points of this paper is also showing us advantages using Guided-Backpropagation which is a way of visuallizing CNN. So I will read this part intensively next reading time. Still writing reading log takes long time and is depending on English dictionary, but it seems to have improved a little bit. 

 

Reference:

Comments