+ All Categories
Home > Documents > Convolutional neural networks...

Convolutional neural networks...

Date post: 01-Oct-2020
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
51
Convolutional neural networks III October 3 rd , 2018 Yong Jae Lee UC Davis Many slides from Rob Fergus, Svetlana Lazebnik, Jia-Bin Huang, Derek Hoiem, Adriana Kovashka, Andrej Karpathy
Transcript
Page 1: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Convolutional neural networks III

October 3rd, 2018

Yong Jae LeeUC Davis

Many slides from Rob Fergus, Svetlana Lazebnik, Jia-Bin Huang, Derek Hoiem, Adriana Kovashka, Andrej Karpathy

Page 2: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Announcements• Sign-up for paper presentations• Paper review due Thurs 11:59 PM

2

Page 3: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Gradient descent• We’ll update weights iteratively• Move in direction opposite to gradient:

LLearning rate

Time

Figure from Andrej Karpathy

original Wnegative gradient direction

W_1

W_2

loss function landscape

Page 4: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Gradient descent in multi-layer nets• We’ll update weights• Move in direction opposite to gradient:

• How to update the weights at all layers?• Answer: backpropagation of loss from higher

layers to lower layers

Page 5: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Backpropagation: Graphic example

• First calculate error of output units and use this to change the top layer of weights.

output

hidden

input

Update weights into j

Adapted from Ray Mooney

k

j

i

w(2)

w(1)

Page 6: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Backpropagation: Graphic example

• Next calculate error for hidden units based on errors on the output units it feeds into.

output

hidden

input

k

j

i

Adapted from Ray Mooney

Page 7: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Backpropagation: Graphic example

• Finally update bottom layer of weights based on errors calculated for hidden units.

output

hidden

input

Update weights into i

k

j

i

Adapted from Ray Mooney

Page 8: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Backpropagation• Easier if we use computational graphs,

especially when we have complicated functions typical in deep neural networks

Figure from Karpathy

Page 9: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Lecture 4 - 10

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 10: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 11

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 11: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 12

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 12: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 13

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 13: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 14

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 14: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 15

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 15: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 16

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 16: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 17

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 17: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 18

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 18: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Chain rule:

Want:

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Upstream gradient Local gradient

Page 19: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Want:

Lecture 4 - 20

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 20: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

e.g. x = -2, y = 5, z = -4

Chain rule:

Want:

Lecture 4 - 21

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Backpropagation: a simple example

Page 21: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

f

activations

Lecture 4 - 22

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

Page 22: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

activations

Lecture 4 - 23

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

“local gradient”

f

gradients

Page 23: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

activations

Lecture 4 - 24

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

“local gradient”

f

gradients

Page 24: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

activations

Lecture 4 - 25

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

“local gradient”

f

gradients

Page 25: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

activations

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

“local gradient”

f

gradients

Page 26: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 4 - 13 Jan 2016

activations

Lecture 4 - 27

13 Jan 2016Fei-Fei Li & Andrej Karpathy & Justin Johnson

Andrej Karpathy

“local gradient”

f

gradients

Page 27: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Backpropagation: another example

Andrej Karpathy

Page 28: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Convolutional Neural Networks (CNN)• Neural network with specialized

connectivity structure• Stack multiple stages of feature extractors• Higher stages compute more global, more

invariant, more abstract features• Classification layer at the end

Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE 86(11): 2278–2324, 1998.

Adapted from Rob Fergus

Page 29: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

• Feed-forward feature extraction:

1. Convolve input with learned filters2. Apply non-linearity 3. Spatial pooling (downsample)

• Supervised training of convolutional filters by back-propagating classification error

Adapted from Lana Lazebnik

Convolutional Neural Networks (CNN)

Input Image

Convolution (Learned)

Non-linearity

Spatial pooling

Output (class probs)

Page 30: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

3

32x32x3 image

width

height

32depth

Convolutions: More detail

Andrej Karpathy

Page 31: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

32

3

5x5x3 filter

32x32x3 image

Convolve the filter with the imagei.e. “slide over the image spatially, computing dot products”

Convolutions: More detail

Andrej Karpathy

Page 32: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

32

3

Convolution Layer32x32x3 image 5x5x3 filter

1 number:the result of taking a dot product between the filter and a small 5x5x3 chunk of the image(i.e. 5*5*3 = 75-dimensional dot product + bias)

Convolutions: More detail

Andrej Karpathy

Page 33: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

32

3

Convolution Layeractivation map

32x32x3 image5x5x3 filter

1

28

28

convolve (slide) over all spatial locations

Convolutions: More detail

Andrej Karpathy

Page 34: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

32

3

Convolution Layer32x32x3 image 5x5x3 filter

activation maps

1

28

28

convolve (slide) over all spatial locations

consider a second, green filter

Convolutions: More detail

Andrej Karpathy

Page 35: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

32

3 6

28

activation maps

32

28

Convolution Layer

For example, if we had 6 5x5 filters, we’ll get 6 separate activation maps:

We stack these up to get a “new image” of size 28x28x6!

Convolutions: More detail

Andrej Karpathy

Page 36: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Preview: ConvNet is a sequence of Convolution Layers, interspersed with activation functions

32

32

3

28

28

6

CONV, ReLUe.g. 65x5x3filters

Convolutions: More detail

Andrej Karpathy

Page 37: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Preview: ConvNet is a sequence of Convolutional Layers, interspersed with activationfunctions

32

32

3

CONV, ReLUe.g. 65x5x3filters 28

28

6

CONV, ReLUe.g. 10 5x5x6 filters

CONV, ReLU

….

10

24

24

Convolutions: More detail

Andrej Karpathy

Page 38: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

preview:

Convolutions: More detail

Andrej Karpathy

Page 39: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Figure from http://www.mdpi.com/2072-4292/7/11/14680/htm

A Common Architecture: AlexNet

Page 40: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Case Study: VGGNet

Only 3x3 CONV stride 1, pad 1 and 2x2 MAX POOL stride 2

best model

11.2% top 5 error in ILSVRC 2013->7.3% top 5 error

[Simonyan and Zisserman, 2014]

Andrej Karpathy

Page 41: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Practical matters

Page 42: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Comments on training algorithm• Not guaranteed to converge to zero training error, may

converge to local optima or oscillate indefinitely.• However, in practice, does converge to low error for many

large networks on real data.• Thousands of epochs (epoch = network sees all training data

once) may be required, hours or days to train.• To avoid local-minima problems, run several trials starting with

different random weights (random restarts), and take results of trial with lowest training set error.

• May be hard to set learning rate and to select number of hidden units and layers.

• Neural networks had fallen out of fashion in 90s, early 2000s; back with a new name and significantly improved performance (deep networks trained with dropout and lots of data).

Ray Mooney, Carlos Guestrin, Dhruv Batra

Page 43: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Over-training prevention• Running too many epochs can result in over-fitting.

• Keep a hold-out validation set and test accuracy on it after every epoch. Stop training when additional epochs actually increase validation error.

0 # training epochs

erro

r

on training data

on test data

Adapted from Ray Mooney

Page 44: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Training: Best practices• Use mini-batch • Use regularization• Use cross-validation for your parameters• Use RELU or leaky RELU, don’t use sigmoid• Center (subtract mean from) your data• Learning rate: too high? too low?

Page 45: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Data Augmentation (Jittering)• Create virtual training samples

– Horizontal flip– Random crop– Color casting– Geometric distortion

Deep Image [Wu et al. 2015]Jia-bin Huang

Page 46: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Regularization: Dropout

Dropout: A simple way to prevent neural networks from overfitting [Srivastava JMLR 2014]

• Randomly turn off some neurons• Allows individual neurons to independently be responsible for performance

Adapted from Jia-bin Huang

Page 47: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Transfer Learning

“You need a lot of a data if you want to train/use CNNs”

Andrej Karpathy

Page 48: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Transfer Learning with CNNs

• The more weights you need to learn, the more data you need

• That’s why with a deeper network, you need more data for training than for a shallower network

• One possible solution:

Set these to the already learned weights from another network

Learn these on your own task

Page 49: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

1. Train on ImageNet

2. Small dataset:

Freeze these

Train this

3. Medium dataset:finetuning

more data = retrain more of the network (or all of it)

Freeze these

Lecture 11 - 29

Train this

Transfer Learning with CNNs

Adapted from Andrej Karpathy

Source: classification on ImageNet Target: some other task/data

Page 50: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Summary• We use deep neural networks because of their

strong performance in practice• Convolutional neural networks (CNN)

• Convolution, nonlinearity, max pooling• Training deep neural nets

• We need an objective function that measures and guides us towards good performance

• We need a way to minimize the loss function: stochastic gradient descent

• We need backpropagation to propagate error through all layers and change their weights

• Practices for preventing overfitting• Dropout; data augmentation; transfer learning

Page 51: Convolutional neural networks IIIweb.cs.ucdavis.edu/~yjlee/teaching/ecs269-fall2018/cnn_basics3.pdf · Convolutional neural networks III October 3rd, 2018. Yong Jae Lee. UC Davis

Questions?

See you Friday!

51


Recommended