Convolutional Neural Networks Prof Adriana Kovashka University of Pittsburgh January 26 2017 Biological analog A biological neuron An artificial neuron Jia bin Huang Hubel and Weisels ID: 708287
Download Presentation The PPT/PDF document "CS 2770: Computer Vision" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.
Slide1
CS 2770: Computer VisionConvolutional Neural Networks
Prof. Adriana
Kovashka
University of Pittsburgh
January 26, 2017Slide2
Biological analog
A biological neuron
An artificial neuron
Jia
-bin HuangSlide3
Hubel and
Weisel’s
architecture
Multi-layer neural network
Adapted from
Jia
-bin Huang
Biological analogSlide4
Convolutional Neural Networks (CNN)Neural network with specialized connectivity structureStack multiple stages of feature extractors
Higher stages compute more global, more invariant,
more abstract
features
Classification layer at the end
Y.
LeCun, L. Bottou, Y. Bengio, and P. Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE 86(11): 2278–2324, 1998.Adapted from Rob FergusSlide5
Feed-forward feature extraction: Convolve input with learned filtersApply non-linearity Spatial pooling (downsample)
Supervised training of convolutional
filters by back-propagating
classification error
Adapted from Lana
Lazebnik
Convolutional Neural Networks (CNN)Input ImageConvolution (Learned)Non-linearitySpatial pooling
Output (class
probs
)
…Slide6
1. Convolution
Apply learned filter weights
One feature map per filter
Stride can be greater than 1 (faster, less memory)
Input
Feature Map
...
Adapted from Rob FergusSlide7
5
2
5
4
4
5
2003200415544551122001352001
2002002001F.06.12.06.12.25.12.06.12.06
H
u = -1,
v = -1
(0, 0)
(
i
, j)
1. Convolution
output
filter
i
nput (image)Slide8
5
2
5
4
4
5
2003200415544551122001352001
2002002001F.06.12.06.12.25.12.06.12.06
H
u = -1,
v = -1
v = 0
(0, 0)
1. Convolution
(
i
, j)
output
filter
i
nput (image)Slide9
5
2
5
4
4
5
2003200415544551122001352001
2002002001F.06.12.06.12.25.12.06.12.06
H
u = -1,
v = -1
v = 0
v = +1
(0, 0)
1. Convolution
(
i
, j)
output
filter
i
nput (image)Slide10
5
2
5
4
4
5
2003200415544551122001352001
2002002001F.06.12.06.12.25.12.06.12.06
H
u = -1,
v = -1
v = 0
v = +1
u = 0,
v = -1
(0, 0)
1. Convolution
(
i
, j)
output
filter
i
nput (image)Slide11
2. Non-LinearityPer-element (independent)Options:Tanh
Sigmoid
: 1/(1+exp(-x))
Rectified linear unit (
ReLU
)Avoids saturation issues
Adapted from Rob FergusSlide12
3. Spatial PoolingSum or max over non-overlapping / overlapping regionsRole of pooling:Invariance to small transformations
Larger receptive fields (neurons see more of input)
Max
Sum
Adapted from Rob FergusSlide13
3. Spatial PoolingSum or max over non-overlapping / overlapping regionsRole of pooling:Invariance to small transformations
Larger receptive fields (neurons see more of input)
Rob Fergus, figure from Andrej
KarpathySlide14
32
3
32x32x3
image
width
height32
depthConvolutions: More detailAndrej KarpathySlide15
32
32
3
5x5x3
filter
32x32x3
imageConvolve the filter with the imagei.e. “slide over the image spatially, computing dot products”Convolutions: More detailAndrej KarpathySlide16
32
32
3
Convolution
Layer
32x32x3
image
5x5x3
filter
1
number:
the result of taking a dot product between the filter and a small 5x5x3 chunk of the
image
(i.e. 5*5*3 = 75-dimensional dot product +
bias)
Convolutions: More detail
Andrej
KarpathySlide17
32
32
3
Convolution
Layer
activation
map
32x32x3
image
5x5x3
filter
1
28
28
convolve (slide) over all spatial
locations
Convolutions: More detail
Andrej
KarpathySlide18
32
32
3
Convolution
Layer
32x32x3
image
5x5x3
filter
activation
maps
1
28
28
convolve (slide) over all spatial
locations
consider a second,
green
filter
Convolutions: More detail
Andrej
KarpathySlide19
32
3
6
28
activation
maps
32
28
Convolution
Layer
For example, if we had 6 5x5 filters, we’ll get 6 separate activation
maps:
We stack these up to get a “new image” of size
28x28x6!
Convolutions: More detail
Andrej
KarpathySlide20
Preview: ConvNet is a sequence of Convolution Layers, interspersed with activation functions
32
32
3
28
28
6
CONV, ReLU
e.g. 6 5x5x3 filters
Convolutions: More detail
Andrej
KarpathySlide21
Preview: ConvNet is a sequence of Convolutional Layers, interspersed with activation functions
32
32
3
CONV, ReLU
e.g. 6 5x5x3 filters
28
28
6
CONV, ReLU
e.g.
10 5x5x
6
filters
CONV, ReLU
….
10
24
24
Convolutions: More detail
Andrej
KarpathySlide22
Preview
[From recent
Yann LeCun
slides]
Convolutions: More detailAndrej KarpathySlide23
example 5x5
filters
(32
total)
We call the layer convolutional because it is related to convolution of two
signals:elementwise multiplication and sum of a filter and the signal (image)one filter =>one activation map
Convolutions: More detail
Adapted from Andrej
Karpathy
, Kristen
GraumanSlide24
A closer look at spatial dimensions:
32
32
3
activation
map
32x32x3
image
5x5x3
filter
1
28
28
convolve (slide) over all spatial
locations
Convolutions: More detail
Andrej
KarpathySlide25
7
7x7 input
(spatially) assume 3x3
filter
7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide26
7
7x7 input
(spatially) assume 3x3
filter
7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide27
77x7 input (spatially) assume 3x3 filter7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide28
77x7 input (spatially) assume 3x3 filter7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide29
=> 5x5 output
7
7x7 input
(spatially) assume 3x3
filter
7A closer look at spatial dimensions:
Convolutions: More detail
Andrej
KarpathySlide30
7x7 input (spatially) assume 3x3 filter applied with stride 2
7
7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide31
7x7 input (spatially) assume 3x3 filter applied with stride 2
7
7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide32
7x7 input (spatially) assume 3x3 filter applied with stride 2
=> 3x3
output!
7
7
A closer look at spatial dimensions:
Convolutions: More detail
Andrej
KarpathySlide33
7x7 input (spatially) assume 3x3 filter applied with stride 3?
7
7
A closer look at spatial
dimensions:
Convolutions: More detail
Andrej
KarpathySlide34
7x7 input (spatially) assume 3x3 filter applied with stride 3?
7
7
A closer look at spatial
dimensions:
doesn’t
fit!
cannot apply 3x3 filter on 7x7 input with stride
3.
Convolutions: More detail
Andrej
KarpathySlide35
N
F
F
N
Output
size:
(N
-
F) / stride +
1
e.g. N = 7, F =
3:
stride 1 => (7
-
3)/1 + 1 =
5
stride 2 => (7
-
3)/2 + 1 =
3
stride 3 => (7
-
3)/3 + 1 = 2.33
:\
Convolutions: More detail
Andrej
KarpathySlide36
In practice: Common to zero pad the border
0
0
0
0
0
0
0
0
0
0
e.g. input
7x7
3x3
filter, applied with
stride
1
pad with 1 pixel
border => what is the
output?
(recall:)
(N
-
F) / stride +
1
Convolutions: More detail
Andrej
KarpathySlide37
In practice: Common to zero pad the bordere.g. input
7x7
3x3
filter, applied with
stride
1pad with 1 pixel border => what is the output?7x7 output!0
00000
0
0
0
0
Convolutions: More detail
Andrej
KarpathySlide38
In practice: Common to zero pad the bordere.g. input
7x7
3x3
filter, applied with
stride
1pad with 1 pixel border => what is the output?7x7 output!in general, common to see CONV layers with stride 1, filters of size FxF, and zero-padding with (F-1)/2. (will preserve size spatially)e.g. F = 3 => zero pad with 1 F = 5 => zero pad with 2 F = 7 => zero pad with 3
000000
0
0
0
0
Convolutions: More detail
Andrej
Karpathy
(N +
2*padding
-
F) / stride +
1Slide39
Examples time:Input volume:
32x32x3
10 5x5 filters with stride 1, pad
2
Output volume size: ?
Convolutions: More detailAndrej KarpathySlide40
Examples time:Input volume:
32x32
x3
10
5x5
filters with stride 1, pad 2Output volume size:(32+2*2-5)/1+1 = 32 spatially, so32x32x10
Convolutions: More detail
Andrej
KarpathySlide41
Examples time:Input volume:
32x32x3
10 5x5 filters with stride 1, pad
2
Number of parameters in this layer?
Convolutions: More detailAndrej KarpathySlide42
Examples time:Input volume:
32x32
x
3
10
5x5
filters with stride 1, pad 2(+1 for bias)Number of parameters in this layer? each filter has 5*5*3 + 1 = 76 params=> 76*10 = 760
Convolutions: More detail
Andrej
KarpathySlide43
Convolutions: More detail
Andrej
KarpathySlide44
preview:
Convolutions: More detail
Andrej
KarpathySlide45
Figure from http://www.mdpi.com/2072-4292/7/11/14680/htm A Common Architecture: AlexNetSlide46
[Zeiler and Fergus, 2013]
Case Study:
ZFNet
AlexNet
but:CONV1: change from (11x11 stride 4) to (7x7 stride 2) CONV3,4,5: instead of 384, 384, 256 filters use 512, 1024, 512ImageNet top 5 error: 15.4% -> 14.8%Andrej KarpathySlide47
Case Study: VGGNet
Only 3x3 CONV stride 1, pad 1 and 2x2 MAX POOL stride
2
best
model11.2% top 5 error in ILSVRC 2013->7.3% top 5 error
[Simonyan and Zisserman, 2014]Andrej KarpathySlide48
[Szegedy et al., 2014]
Inception
module
ILSVRC 2014 winner (6.7% top 5
error)Case Study: GoogLeNetAndrej KarpathySlide49
Slide from Kaiming He’s recent presentation https://www.youtube.com/watch?v=1PGLj-uKT1w
[He et al.,
2015]
ILSVRC 2015 winner (3.6% top 5
error)Case Study: ResNetAndrej KarpathySlide50
(slide from Kaiming He’s recent
presentation)
Case Study:
ResNet
Andrej
KarpathySlide51
[He et al., 2015]ILSVRC 2015 winner (3.6% top 5
error)
(slide from Kaiming He’s recent
presentation)
2-3 weeks of
training on 8 GPU machineat runtime: faster than a VGGNet! (even though it has 8x more layers)Case Study: ResNetAndrej
KarpathySlide52
Practical mattersSlide53
Training: Best practicesUse mini-batch Use regularizationUse gradient checksUse cross-validation for your parameters
Use RELU or leaky RELU or ELU, don’t use sigmoid
Center (subtract mean from) your data
To initialize, use “Xavier initialization”
Learning rate: too high? Too low? Slide54
Regularization: Dropout
Dropout: A simple way to prevent neural networks from
overfitting
[
Srivastava JMLR 2014
]
Randomly turn off some neurons Allows individual neurons to independently be responsible for performanceAdapted from Jia-bin HuangSlide55
Data Augmentation (Jittering)Create virtual training samplesHorizontal flipRandom cropColor castingGeometric distortion
Deep Image [
Wu et al. 2015
]
Jia
-bin HuangSlide56
Transfer Learning“You need a lot of a data if you want to train/use
CNNs”
BUSTED
Andrej
KarpathySlide57
1. Train
on
Image
N
et
2. Small
dataset:
Freeze theseTrain this3. Medium dataset:finetuningmore data = retrain more of the network (or
all of it)
Freeze
these
Lecture 11
-
29
Train
this
Transfer Learning with
CNNs
Adapted from Andrej
Karpathy
Another option: use network as
feature
extractor
,
train SVM on extracted features for target task
Source: classification on ImageNet Target: some other task/dataSlide58
more
generic
more
specific
Lecture 11 - 34
very similar datasetvery different datasetvery little dataUse linear classifier on top layer
You’re in trouble… Try linear
classifier from different stages
quite a lot
of data
Finetune a
few layers
Finetune a larger number
of layers
Transfer Learning with
CNNs
Andrej
KarpathySlide59
Simplest Way to Use CNNsTake model trained on, e.g., ImageNet 2012 training setEasiest: Take outputs of e.g. 6th or 7th fully-connected layer, and
p
lug features from each layer into linear SVM
Features are neuron activations at that level
Can train linear SVM for different tasks, not just one used to learn the deep net
Better: fine-tune features and/or classifier on new dataset
Classify test set of new datasetAdapted from Lana LazebnikSlide60
PackagesCaffe and Caffe Model ZooTorchTheano with Keras/Lasagne
MatConvNet
TensorFlowSlide61
Learning Resourceshttp://deeplearning.net/http://cs231n.stanford.eduSlide62
Things to rememberOverviewNeuroscience, perceptron, multi-layer neural networksConvolutional neural network (CNN)Convolution, nonlinearity, max poolingTraining CNNDropout; data augmentation; transfer learning
Using CNNs for your own task
Basic first step: try the pre-trained
CaffeNet
fc6-fc8 layers as features
Adapted from
Jia-bin Huang