TAMU CSCE 689 - he2005laplacianfacesSLIDES

Unformatted text preview:

Face Recognition Using LaplacianfacesHe et al. (IEEE Trans PAMI, 2005)presented by Hassan A. KingraviOverview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold Learning LPP and its Connections to LDA and PCA Laplacianfaces for Face Analysis Experiments ConclusionsIntroduction Face Recognition* Many methods developed* An example of the Appearance-Based Method, i.e. templates extracted from given information Current Approach (Laplacianfaces)* Builds on the dimensionality reduction approach* Takes into account the possibility of the distribution of the data on a non-linear subspace by preserving local structure (Locality Preserving Projections)* Generalizes to unseen points (problem in most nonlinear methods)Dimensionality Reduction Images represented as vectors in extremely high-dimensional space Idea is that we can create decision boundaries between different classes of objects Dimensionality reduction is required because:a) Inherent structure of data may not be apparent in high dimensional space; vector can have irrelevant attributes that contain little useful informationb) A projection makes the data more tractable to manage Reduction can be linear (PCA, LDA) or nonlinear (ISOMAP)DR Methods : PCA Projects high-dimensional data to low dimensional subspace Objective function maximizes variance globally:  Solution by finding the eigenvectors of the (reduced) covariance matrix .21max ( )niwiyy=−∑12, ,...,kww wDR Methods : PCA Since PCA maximizes global variance, it can be thought of re-representing as much of the original signal as possible. This does NOT mean that the data is projected with an aim for classification; rather, it is compressed. Other methods exist which aim to project the vector based on multiclass discrimination.DR Methods : LDA LDA seeks directions of projection that are efficient for discrimination.  The objective function maximizes between-class scatter over within-class scatter:maxTBTwWwSwwS wDR Methods : LDA LDA seeks directions of projection that are efficient for discrimination.  LDA is defined in terms of the number of classes of projection; hence, it is a supervised learning algorithm (need to know class labels.) In the face recognition problem, one unique face represents a class. Good performance in general (better than PCA for faces), but still linear.Nonlinear Methods When the linear subspace assumption is violated, for e.g. the figure on the right.  The data lies on a nonlinear manifold, a mathematical space where the local area around a point may be Euclidean, but the overall structure may be more complicatedNonlinear Methods Idea: manifolds arise naturally whenever there’s a smooth variation of parameters.  Hypothesis; face recognition problems are non-linear in nature.  Specifically, images that change smoothly over time (video).Nonlinear Methods: ISOMAP An example of a nonlinear embedding method: ISOMAP. Need to see geodesic structure; solution: graph embeddings. Steps of ISOMAP:a) Find nearest neighbors to each sample (either use a k rule or a radius based on Euclidean distance) and construct a graph of the geodesic.b) Use Dijkstra’s shortest path algorithm to find distances between all points.c) Apply multidimensional scaling.Locality Preserving Projections Problems: ISOMAP is computationally intensive and the embedding’s only defined on actual data points. Solution: Locality Preserving Projections, the method of Laplacianfaces.  LPP is a linear method that approximates nonlinear methods (specifically, the Laplacian Eigenmap.) LPP minimizes the following objective function:where 2min ( )ijijijyyS−∑22exp( || || / ), || ||ij ij ijSxxtxx=−− − <εLocality Preserving Projections The resulting mapping amounts to the following eigenvalue problem: L is the Laplacian matrix, i.e. D – S, where S corresponds to the similarity values defined, and D is a column matrix which reflects how important a certain projection is. The more data points that surround a given point, the more “important” it is. Thus, the mapping preserves locality.  The given equation corresponds to the Laplace-Beltrami operator on differential manifolds. TTXLX w XDX wλ=LPP and other methods LPP is a linear approximation to nonlinear methods, which takes locality into account. If one aims to preserve global structure only, let the neighborhood grow to infinity; data points are projected in directions of maximal variance, i.e. LPP becomes similar to PCA. LDA preserves discriminating information and global geometric structure; through manipulation, LPP can induce LDA. LDA is supervised, while LPP is unsupervised.Laplacianfaces Method:a) Because can be sparse, the image is first projected onto a PCA subspace.b) The nearest neighbor graph is constructed, like ISOMAP. Laplacianfaces use the knn rule. c) Weights are chosen by the equation.d) The eigenmap is calculated. We get a set of k vectors that represent the new subspace. TXDXijSTTXLX w XDX wλ=Laplacianfaces Results from a mapping to a 2-d space.Experimental ResultsDiscussion & Conclusions Method has higher accuracy than both PCA and LDA approaches.  LDA needs more than one sample per class to classify; LPP behaves like PCA.  Method is faster than ISOMAP, and generalizes well (not specifically nonlinear.) LPP can also be applied to other machine learning issues.References He et al. “Face Recognition Using Laplacianfaces.” Ricardo Gutierrez-Osuna, Pattern Recognition Slides, Fall


View Full Document

TAMU CSCE 689 - he2005laplacianfacesSLIDES

Documents in this Course
slides

slides

10 pages

riccardo2

riccardo2

33 pages

ffd

ffd

33 pages

intro

intro

23 pages

slides

slides

19 pages

p888-ju

p888-ju

8 pages

w1

w1

23 pages

vfsd

vfsd

8 pages

subspace

subspace

48 pages

chapter2

chapter2

20 pages

MC

MC

41 pages

w3

w3

8 pages

Tandem

Tandem

11 pages

meanvalue

meanvalue

46 pages

w2

w2

10 pages

CS689-MD

CS689-MD

17 pages

VGL

VGL

8 pages

ssq

ssq

10 pages

Load more
Download he2005laplacianfacesSLIDES
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view he2005laplacianfacesSLIDES and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view he2005laplacianfacesSLIDES 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?