Face processing and detection using Artificial Neural...

Post on 03-Aug-2020

0 views 0 download

transcript

1

Face processing and detection using Artificial Neural Networks

and Image Processing

Michel PAINDAVOINE, Fan YANG

paindav@u-bourgogne.fr

fanyang@u-bourgogne.fr

2

Outline1. - Introduction to neural networks

Neural networks learningPerceptron and Adaline networks

5. - Autoassociative memory Principal Components AnalysisWavelet Transform and Face recognition

9. - Multi-Layer Perceptron Back-Propagation learning rule

Face identification

12. - Radial Basis Function neural networkUnsupervised training techniqueFace detection and identification in video sequences

3

1. Introduction to neural networks

Schematic of a biological neuron and an artificial neuron

4

Hebbian learning rule : If two neurons i and j are active simultaneously, their interconnection must be strengthened.

aaw jiij η=∆

1.1 Perceptron

5

1.2 Adaptive linear element : Adaline network

∑=

=I

iiijj xwo

0

Widrow-Hoff learning (delta) rule : Minimize the summed square error function by gradient descent method in order to adjust the weight.

xotw ijjij )( −=∆ η

tj : target response of the jth neuron

6

1.3 An example of pattern recognition

We want to recognize the follow numbers patterns 0 - 9 :

The purpose is to associate each pattern with its class corresponding to :

7

1.3 An example of pattern recognition

The matrix of weights is initiated with values chosen randomly. The number pattern 0,1, …9 are presented to input units and the responses of the output units are calculated with :

)()(14

0xwfafo i

iijjj ∑==

=

Initially, the total error E of incorrect classification is : Error = 54%

The Widrow-Hoff learning rule is applied to modify the weights.

8

1.3 An example of pattern recognition

After 51 iteratios, the final matrix of weight is obtained.

9

1.3 An example of pattern recognition

A perfect pattern of number 8 is correctly categorized

A nearly perfect pattern of number 8 is correctly categorized

A imperfect pattern of number 8 is not correctly categorized

10

2. Auto-associative memory Single layer network made of fully interconnected linear units,

Content addressable memory,

Able to retrieve a whole pattern of information given one part of this information.

An auto-associative memory composed of 5 fully interconnected units.

11

2.1 Auto-associative memory and Face recognition

The linear auto-associator is applied to images : transform each digitized image into a vector by concatening the column of the matrix of the pixel values of the image.

12

2.1 Auto-associative memory and Face recognition

Illustration of a content addressable memory for faces : the auto-associative memory is able to reconstruct a face from an incomplete input (top).

Illustration of a content addressable memory for faces : the auto-associative memory is able to reconstruct a face from an noise degraded input (left).

13

2.2 Wavelet transform and Face recognition

14

2.2 Wavelet transform and Face recognition

15

2.2 Wavelet transform and Face recognition

16

2.2 Wavelet transform and Face recognition

17

2.3 Auto-associative memory and PCA

18

2.3 Auto-associative memory and PCA

XWWW TTt

Tt

Tt XT )( )()()1( −+=+ η

UW Tt

Tt U Φ ++ = )1()1( ][ )( )1(

)1( Λ−Φ+

+ −= ηI t

t I

H.ABDI developed a simple method of implementing the Widrow-Hoff algorithm by using the eigen-decomposition of the synapse matrix W

This decomposition give rise to Principal Components Analysis (PCA).

X is the stimuli matrix of auto-associator, U is the matrix of eigenvectors (Principal components) of XXT,

Λ is the matrix of eigenvalues of XXT.

UW TU=∞)(

19

2.4 Non linear separable problems

20

3. Multi-layer neural networks

Can overcome many restrictions : solve non linear separable problems

Universal approximator

21

3.1 Generalized Widrow-Hoff rule : Back-propagation learning rule

hz ljlj δη=∆ )(')( aot jjjj f−=δ

xw ilil δη=∆

Principle : Back-propagate the error signal from the output layer to hidden layers

)('1

az llj

J

jjl f∑

=

= δδ

From input layer to hidden layer : connection matrix W :

From hidden layer to output layer : connection matrix Z :

22

3.1 Generalized Widrow-Hoff rule : Back-propagation learning rule

e xxf −+=

11)(The sigmoid (logistic) function :

)](1)[()(' xfxfxf −=Interesting property :

23

3.2 Face identification using the MLP

PCA

Datacompress

MLP Identity

24

3.2 Face identification using the MLP

??? Non recognized

Guylaine, Christine, …

Daniel, Bernard, Elbay, …

Patterns for network training

25

Input x Hidden nodes

Output)(xf i

js

Iii ......1=Jjj ......1=

jip

J classesInput vector of N

dimension

4. Radial Basis Function (RBF) network

26

1x

2x

3x

Nx

nx),( iic σ

2

2

2)(

)( i

xd

i exf σ−

=

4.1 Principle of RBF neural network

Hidden nodes :

iC≡0 1 2 3 41−2−3−4−

2.0

4.0

6.0

8.0

1

2.1

( )xfi

x

27

4.1 Principle of RBF neural network

)(1 xf

)(2 xf

)(3 xf

)(xf I

)(xf i

jp1

jp2

jp3

jip

jIp

∑=

⋅=I

iiij xfpS

j1

)(

2.0

4.0

6.0

8.0

0

5 10 15 205−10−15−20− 0

Output units :

28

4.1 Principle of RBF neural networkUn-supervised training technique : Clustering

kc

lcic

iσkc

kσ kσiclciσ

kckσ

lckσ

kc

lckσ

ic kcic

lc

ickσ

kclc kσ

ic

29

Face detection (or localization) :

Faces identification :

Where ?!

Who ?!

1x 2x

1y

2y

Mr. Y

~~~~~

~~~~~

4.2 Faces detection and identification using RBF

30

face

or

non face?

non face !

Face !Mr. X

?person recognized

or non recognized ?

Classic process

Face images Processing

31

Used process

person recognized or

non recognized ?

Mr. X

?

?

Proposed method(based on mapping Algo./Archi. Approach)

32

Extraction of input vectors

Detection and identification by

RBF network

1=e3

2=e9

4=e27

8=e

Pre-processing

Face detection and identification process :

4.2 Faces detection and identification using RBF

33

4.3 Experiments and Results

Variation of the input vector lengths

87,5%551701796801 pix./8 on row1 pix./2 on col.

87,1%601701796801 pix./16 on each row

93,3%012017961601 pix./8 on each row

95,8%07617963201 pix./4 on each row

92,2%014017961280Original window (40x32)

Correct results

Incorrect reco.Non reco. Nb. facesNumber

componentsSub sampling

34

Variation of used mesured distances

96,4%0641796320d1(x)

95,8%0761796320d2(x)

Correctresults

Incorrectreco.

Non reco.Nb. facesNumber componentsDistance

∑=

−=N

nnn cxxd

11 )(( )∑

=−=

N

nnn cxxd

1

22 )(

4.3 Experiments and Results

35

Variation of RBF kernel activation functions

92%01181796320Heaviside

96,4%0321796320Gaussian

Correct results

Incorrectreco.

Non reco.Nb. facesNumber

componentsFonction

4.3 Experiments and Results

36

4.3 Faces detection and identification using RBF

Experimental results for one scale :

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

6. Conclusion/Perspectives

Neural networks :

are efficient tools for pattern recognition

facilitate hardware implementations by massive parallel

Features extraction to few neural networks

Image processing