/
CS 485/685 CS 485/685

CS 485/685 - PowerPoint Presentation

tawny-fly
tawny-fly . @tawny-fly
Follow
352 views
Uploaded On 2015-09-19

CS 485/685 - PPT Presentation

Computer Vision Face Recognition Using Principal Components Analysis PCA M Turk A Pentland Eigenfaces for Recognition Journal of Cognitive Neuroscience 31 pp 7186 1991 ID: 133235

pca face space eigenfaces face pca eigenfaces space principal recognition data analysis component dimensional dimensionality faces distance information case

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "CS 485/685" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

CS 485/685Computer Vision

Face Recognition Using Principal Components

Analysis (PCA)

M. Turk, A.

Pentland

, "

Eigenfaces

for Recognition

", Journal of Cognitive Neuroscience, 3(1), pp. 71-86, 1991. Slide2

2Principal Component Analysis (PCA)

Pattern recognition in high-dimensional spaces

Problems arise when performing recognition in a high-dimensional space (curse of dimensionality).

Significant improvements can be achieved by first mapping the data into a

lower-dimensional sub-

space.

The goal of PCA is to reduce the dimensionality of the data while retaining as much information as possible in the original dataset.Slide3

3Principal Component Analysis (PCA)

Dimensionality reduction

PCA allows us to compute a

linear transformation

that maps data from a high dimensional space to a lower dimensional sub-space.

K

x NSlide4

4Principal Component Analysis (PCA)

Lower dimensionality basis

Approximate vectors by finding a basis in an appropriate lower dimensional space.

(1)

Higher-dimensional

space representation:

(2)

Lower-dimensional

space representation:Slide5

5Principal Component Analysis (PCA)

Information loss

Dimensionality reduction implies information

loss!

PCA preserves

as much information as possible, that

is, it minimizes the error:

How

should we

determine

the best lower dimensional sub-space?Slide6

6Principal Component Analysis (PCA)

Methodology

Suppose

x

1

, x2

, ..., xM are N x 1 vectors

(i.e., center at zero)Slide7

7Principal Component Analysis (PCA)

Methodology – cont.

 Slide8

8Principal Component Analysis (PCA)

Linear transformation implied by PCA

The linear transformation

R

N

 R

K that performs the dimensionality reduction is:

(

i.e., simply computing coefficients of linear expansion)Slide9

9Principal Component Analysis (PCA)

Geometric interpretation

PCA projects the data along the directions where the data varies the

most

.

These directions are determined by the eigenvectors of the covariance matrix corresponding to the

largest eigenvalues.The magnitude of the eigenvalues corresponds to the variance of the data along the eigenvector directions.Slide10

10Principal Component Analysis (PCA)

How to choose

K (i.e., number

of principal

components) ?

To choose

K, use the following criterion:

In this case, we say that we “preserve” 90% or 95% of the

information in our data.

If K=N, then we “preserve” 100% of the information in our data.Slide11

11Principal Component Analysis (PCA)

What

is

the error

due to dimensionality

reduction?

The original vector x can be reconstructed using its principal components:

PCA

minimizes the reconstruction error:

It can be shown that the error is equal to:Slide12

12Principal Component Analysis (PCA)

Standardization

The principal components are dependent on the

units

used to measure the original variables as well as on the range

of values they assume.You should always standardize the data prior to using PCA.

A common standardization method is to transform all the data to have zero mean and unit standard deviation:Slide13

13Application to Faces

Computation

of low-dimensional

basis (i.e.,

eigenfaces

):Slide14

14Application to Faces

Computation of the eigenfaces – cont.Slide15

15Application to Faces

Computation of the eigenfaces – cont.

u

i

Slide16

16Application to Faces

Computation of the eigenfaces – cont.Slide17

17Eigenfaces example

Training imagesSlide18

18Eigenfaces example

Top eigenvectors:

u

1

,…

ukMean: μSlide19

19Application to Faces

Representing faces onto this basis

Face reconstruction:Slide20

20Eigenfaces

Case Study

: Eigenfaces for Face Detection/Recognition

M. Turk, A. Pentland, "Eigenfaces for Recognition",

Journal of Cognitive Neuroscience

, vol. 3, no. 1, pp. 71-86, 1991.

Face Recognition

The simplest approach is to think of it as a template matching problem

Problems arise when performing recognition in a high-dimensional space.

Significant improvements can be achieved by first mapping the data into a

lower dimensionality

space.Slide21

21Eigenfaces

Face Recognition

Using Eigenfaces

whereSlide22

22Eigenfaces

Face Recognition

Using Eigenfaces – cont.

The distance

e

r is called

distance within face space (difs)The Euclidean distance can be used to compute er

, however, the

Mahalanobis

distance

has shown

to

work better:

Mahalanobis

distance

Euclidean distance

Slide23

23Face detection and recognition

Detection

Recognition

“Sally”Slide24

24Eigenfaces

Face Detection

Using Eigenfaces

The distance

e

d

is called

distance from face space

(

dffs

)Slide25

25Eigenfaces

Reconstruction of faces and non-faces

Reconstructed face looks

like a face.

Reconstructed non-face

looks like a fac again!

Input ReconstructedSlide26

26Eigenfaces

Face Detection

Using Eigenfaces – cont.

Case 1:

in face space AND close to a given face

Case 2:

in face space but NOT close to any given faceCase 3: not in face space AND close to a given faceCase 4: not in face space and NOT close to any given faceSlide27

27Reconstruction using partial information

Robust to partial face occlusion.

Input ReconstructedSlide28

28Eigenfaces

Face detection, tracking, and recognition

Visualize

dffs

:Slide29

29Limitations

Background

changes

cause problems

De-emphasize the outside of the face (e.g., by multiplying the input image by a 2D Gaussian window centered on the face).Light changes degrade performanceLight normalization helps.Performance decreases quickly with changes to face size

Multi-scale eigenspaces.Scale input image to multiple sizes.Performance decreases with changes to face orientation (but not as fast as with scale changes)Plane rotations are easier to handle.Out-of-plane rotations are more difficult to handle.Slide30

30Limitations

Not robust to misalignmentSlide31

31Limitations

PCA assumes that the data follows a Gaussian distribution (mean

µ, covariance matrix

Σ

)

The shape of this dataset is not well described by its principal componentsSlide32

32Limitations

PCA is

not

always an optimal dimensionality-reduction procedure for classification purposes:

Related Contents


Next Show more