nose 2 summer school statistical pattern recognition...

100
Ricardo Gutierrez-Osuna Wright State University Statistical Pattern Recognition 1 NOSE 2 nd Summer School Lloret de Mar, Spain, October 2-5, 2000

Upload: others

Post on 27-Jun-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition1

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Page 2: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition2

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Outline

g Introduction to Pattern Recognitiong Dimensionality reductiong Classificationg Validationg Clustering

Page 3: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition3

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

SECTION I: Introduction

g Features, patterns and classifiers g Components of a PR systemg An exampleg Probability definitionsg Bayes Theoremg Gaussian densities

Page 4: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition4

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Features, patterns and classifiers

g Featuren Feature is any distinctive aspect, quality or characteristic

g Features may be symbolic (i.e., color) or numeric (i.e., height)

n The combination of d features is represented as a d-dimensional column vector called a feature vectorg The d-dimensional space defined by the feature vector is called

feature spaceg Objects are represented as points in feature space. This

representation is called a scatter plot

Feature vector Feature space (3D)

=

d

2

1

x

x

x

x

x3

x1 x2

xF

eatu

re 2

Feature 1

Class 1

Class 2

Class 3

Scatter plot (2D)

Page 5: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition5

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Features, patterns and classifiers

g Patternn Pattern is a composite of traits or features characteristic of an

individualn In classification, a pattern is a pair of variables {x,ω} where

g x is a collection of observations or features (feature vector)

g ω is the concept behind the observation (label)

g What makes a “good” feature vector?n The quality of a feature vector is related to its ability to

discriminate examples from different classesg Examples from the same class should have similar feature valuesg Examples from different classes have different feature values

“Good” features “Bad” features

Page 6: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition6

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Features, patterns and classifiers

g More feature properties

g Classifiersn The goal of a classifier is to partition feature space into class-

labeled decision regionsn Borders between decision regions are called decision

boundaries

Highly correlated featuresNon-linear separabilityLinear separability Multi-modal

R1R2

R3

R1

R2

R3

R4

Page 7: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition7

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Components of a pattern recognition system

g A typical pattern recognition system containsn A sensorn A preprocessing mechanismn A feature extraction mechanism (manual or automated)n A classification or description algorithmn A set of examples (training set) already classified or described

Sensor / transducer

Preprocessing and

enhancement

Feature

extraction

Classification algorithm

Description algorithm

The “real world”

Class assignment

Description

Feedback / Adaptation

[Duda, Hart and Stork, 2001]

Page 8: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition8

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

An example

g Consider the following scenario*n A fish processing plan wants to automate the process of sorting

incoming fish according to species (salmon or sea bass)n The automation system consists of

g a conveyor belt for incoming productsg two conveyor belts for sorted products g a pick-and-place robotic armg a vision system with an overhead CCD camerag a computer to analyze images and control the robot arm

*Adapted from [Duda, Hart and Stork, 2001]

Conveyor belt

CCD camera

Conveyor belt (bass)

Conveyor belt (salmon)

Robotarm

computer

Page 9: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition9

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

An example

g Sensorn The camera captures an image as a new fish enters the sorting area

g Preprocessingn Adjustments for average intensity levelsn Segmentation to separate fish from background

g Feature Extractionn Suppose we know that, on the average, sea bass is larger than salmon

count

length

Sea bassSalmon

Decision boundary

g Classificationn Collect a set of examples from both species

g Plot a distribution of lengths for both classesn Determine a decision boundary (threshold)

that minimizes the classification errorg We estimate the system’s probability of

error and obtain a discouraging result of 40%

n What is next?

Page 10: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition10

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Avg. scale intensityle

ngth Decision

boundary

Sea bass Salmon

count

Avg. scale intensity

Sea bass Salmon

Decision boundary

An exampleg Improving the performance of our PR system

n Committed to achieve a recognition rate of 95%, we try a number of features

g Width, Area, Position of the eyes w.r.t. mouth...g only to find out that these features contain no discriminatory information

n Finally we find a “good” feature: average intensity of the scales

n We combine “length” and “average intensity of the scales” to improve class separability

n We compute a linear discriminant function to separate the two classes, and obtain a classification rate of 95.7%

Page 11: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition11

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

An example

g Cost Versus Classification raten Is classification rate the best objective function for this problem?

g The cost of misclassifying salmon as sea bass is that the end customer will occasionally find a tasty piece of salmon when he purchases sea bass

g The cost of misclassifying sea bass as salmon is a customer upset when he finds a piece of sea bass purchased at the price of salmon

n We could intuitively shift the decision boundary to minimize an alternative cost function

Avg. scale intensity

leng

th Decision boundary

Sea bass Salmon

Avg. scale intensity

leng

th NewDecision boundary

Sea bass

Salmon

Page 12: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition12

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

An example

g The issue of generalizationn The recognition rate of our linear classifier (95.7%) met the

design specs, but we still think we can improve the performance of the systemg We then design an artificial neural

network with five hidden layers, a combination of logistic and hyperbolic tangent activation functions, train it with the Levenberg-Marquardt algorithm and obtain an impressive classification rate of 99.9975% with the following decision boundary

n Satisfied with our classifier, we integrate the system and deploy it to the fish processing plantg A few days later the plant manager calls to complain that the system

is misclassifying an average of 25% of the fish

g What went wrong?

Avg. scale intensity

leng

th

SalmonSea bass

Page 13: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition13

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theoryg Probability

n Probabilities are numbers assigned to events that indicate “how likely” it is that the event will occur when a random experiment is performed

g Conditional Probabilityn If A and B are two events, the probability of event A when we already

know that event B has occurred P[A|B] is defined by the relation

g P[A|B] is read as the “conditional probability of A conditioned on B”, or simply the “probability of A given B”

A1

A2

A3A4event

prob

abili

ty

A1 A2 A3

Probability Law

A4

Sample space

0P[B]forP[B]

B]P[AB]|P[A >= �

Page 14: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition14

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theory

g Conditional probability: graphical interpretation

g Theorem of Total Probabilityn Let B1, B2, …, BN be mutually exclusive events, then

S S

“B has occurred”

A A∩B B A A∩B B

∑=

=+=N

1kkkNN11 ]]P[BB|P[A]]P[BB|...P[A]]P[BB|P[AP[A]

B1

B2

B3 BN-1

BN

A

Page 15: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition15

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theory

g Bayes Theoremn Given B1, B2, …, BN, a partition of the sample space S. Suppose

that event A occurs; what is the probability of event Bj?n Using the definition of conditional probability and the Theorem of

total probability we obtain

n Bayes Theorem is definitely the fundamental relationship in Statistical Pattern Recognition

∑=

⋅== N

1kkk

jjjj

]P[B]B|P[A

]P[B]B|P[A

P[A]

]BP[AA]|P[B

Rev. Thomas Bayes (1702-1761)

Page 16: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition16

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theory

g For pattern recognition, Bayes Theorem can be expressed as

n where ωj is the ith class and x is the feature vector

g Each term in the Bayes Theorem has a special name, which you should be familiar with

n P(ωi) Prior probability (of class ωi)

n P(ωi|x) Posterior Probability (of class ωi given the observation x)n P(x|ωi) Likelihood (conditional prob. of x given class ωi)n P(x) A normalization constant that does not affect the decision

g Two commonly used decision rules aren Maximum A Posteriori (MAP): choose the class ωi with highest P(ωi|x)n Maximum Likelihood (ML): choose the class ωi with highest P(x|ωi)

g ML and MAP are equivalent for non-informative priors (P(ωi)=constant)

P(x)

)P()|P(x

)P()|P(x

)P()|P(xx)|P( jj

N

1kkk

jjj

ωω

ωω

ωωω

⋅=

⋅=∑

=

Page 17: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition17

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theory

g Characterizing features/vectorsn Complete: Probability mass/density function

n Partial: Statisticsg Expectation

n The expectation represents the center of mass of a density

g Variancen The variance represents the spread about the mean

g Covariance (only for random vectors)n The tendency of each pair of features to vary together, i.e., to co-vary

100 200 300 400 500

pdf

x(lb)pdf for a person’s weight

1

1 2 3 4 5 6

pmf

x

pmf for rolling a (fair) dice

1

5/6

4/6

3/6

2/6

1/6

Page 18: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition18

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theory

g The covariance matrix (cont.)

n The covariance terms can be expressed as

g where ρik is called the correlation coefficient

g Graphical interpretation

=

−−−−

−−−−=

−−=∑=

2N1N

1N21

NNNN11NN

NN111111

T

...c

......

c...

)])(xE[(x...)])(xE[(x

...

)])(xE[(x...)])(xE[(x

]���;E[(XCOV[X]

σ

σO

Xi

Xk

Cik=-σiσkρik=-1

Xi

Xk

Cik=-½σiσkρik=-½

Xi

Xk

Cik=0ρik=0

Xi

Xk

Cik=+½σiσkρik=+½

Xi

Xk

Cik=σiσkρik=+1

kiikik2

iii candc σσρσ ==

Page 19: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition19

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Review of probability theoryg Meet the multivariate Normal or Gaussian

density N(µ,Σ):

n For a single dimension, this formula reduces to the familiar expression

g Gaussian distributions are very popularn The parameters (µ,Σ) are sufficient to uniquely

characterize the normal distributionn If the xi’s are mutually uncorrelated (cik=0), then

they are also independentg The covariance matrix becomes diagonal, with

the individual variances in the main diagonal

n Marginal and conditional densitiesn Linear transformations

−∑−−

∑= − �(X�(X

21

exp)(2

1(x)f 1T

1/2n/2X π

−=

2

X

-X21

exp2

1(x)f

σσπ

Page 20: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition20

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

SECTION II: Dimensionality reduction

g The “curse of dimensionality”g Feature extraction vs. feature selectiong Feature extraction criteriag Principal Components Analysis (briefly)g Linear Discriminant Analysisg Feature Subset Selection

Page 21: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition21

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

g The “curse of dimensionality” [Bellman, 1961]n Refers to the problems associated with multivariate data analysis as the

dimensionality increases

g Consider a 3-class pattern recognition problemn A simple (Maximum Likelihood) procedure would be to

g Divide the feature space into uniform bins

g Compute the ratio of examples for each class at each bin and, g For a new example, find its bin and choose the predominant class in that bin

n We decide to start with one feature and divide the real line into 3 bins

g Notice that there exists a large overlap between classes ⇒ to improve discrimination, we decide to incorporate a second feature

Dimensionality reduction

x1x1

Page 22: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition22

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

g Moving to two dimensions increases the number of bins from 3 to 32=9n QUESTION: Which should we maintain constant?

g The density of examples per bin? This increases the number of examples from 9 to 27

g The total number of examples? This results in a 2D scatter plot that is very sparse

g Moving to three features …n The number of bins grows to 33=27n To maintain the initial density of examples,

the number of required examples grows to 81n For the same number of examples the

3D scatter plot is almost empty

Dimensionality reduction

x1

x2 Constant density Constant # examples

x1

x2

x1

x2

x3

Page 23: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition23

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Dimensionality reductiong Implications of the curse of dimensionality

n Exponential growth with dimensionality in the number of examplesrequired to accurately estimate a function

g How do we beat the curse of dimensionality?n By incorporating prior knowledgen By providing increasing smoothness of the target functionn By reducing the dimensionality

g In practice, the curse of dimensionality means thatn For a given sample size, there is a maximum number of features above

which the performance of our classifier will degrade rather than improveg In most cases, the information

that was lost by discarding some features is compensated by a more accurate mapping in lower-dimensional space

dimensionality

perf

orm

ance

Page 24: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition24

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

The curse of dimensionalityg Implications of the curse of dimensionality

n Exponential growth in the number of examples required to maintain a given sampling density

g For a density of N examples/bin and D dimensions, the total number of examples grows as ND

n Exponential growth in the complexity of the target function (a density) with increasing dimensionality

g “A function defined in high-dimensional space is likely to be much more complex than a function defined in a lower-dimensional space, and those complications are harder to discern” –Jerome Friedman

n This means that a more complex target function requires denser sample points to learn it well!

n What to do if it ain’t Gaussian?g For one dimension, a large number of density functions are availableg For most multivariate problems, only the Gaussian density is employedg For high-dimensional data, the Gaussian density can only be handled in a

simplified form!

n Human performance in high dimensionsg Humans have an extraordinary capacity to discern patterns in 1-3

dimensions, but these capabilities degrade drastically for 4 or higher dimensions

Page 25: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition25

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Dimensionality reduction

g Two approaches to perform dim. reduction ℜN→ℜM (M<N)n Feature selection: choosing a subset of all the features

n Feature extraction: creating new features by combining existing ones

g In either case, the goal is to find a low-dimensional representation of the data that preserves (most of) the information or structure in the data

g Linear feature extractionn The “optimal” mapping y=f(x) is, in general, a non-linear function whose

form is problem-dependentg Hence, feature extraction is commonly limited to linear projections y=Wx

[ ] [ ] [ ]( )M21 iiiM21N21 ...xxxf...yyy...xxx = →extraction

feature

[ ] [ ]M21 iiiN21 ...xxx...xxx →selection

feature

=

N

2

1

MN2M1M

N22221

N11211

M

2

1

extractionfeaturelinear

N

2

1

x

x

x

www

www

www

y

y

y

x

x

x

MMOMM

L

L

M

Page 26: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition26

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

g Two criteria can be used to find the “optimal” feature extraction mapping y=f(x) n Signal representation: The goal of feature extraction is to represent the

samples accurately in a lower-dimensional space

n Classification: The goal of feature extraction is to enhance the class-discriminatory information in the lower-dimensional space

g Within the realm of linear feature extraction, two techniques are commonly usedn Principal Components (PCA)

g Based on signal representation

n Fisher’s Linear Discriminant (LDA)g Based on classification

Signal representation versus classification

Classif

icatio

n

(LDA)

Feature 1

Fea

ture

2

1

2

111 1 1

1111 1

1

111

11 11

11

22

2 22

22

22 222 22 2 2

2

2Signal representation

(PCA)

Page 27: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition27

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Principal Components Analysis

g Summaryn The optimal* approximation of a random vector x∈ℜN by a linear

combination of M (M<N) independent vectors is obtained by projecting the random vector x onto the eigenvectors ϕi corresponding to the largest eigenvalues λi of the covariance matrix Σx

g *optimality is defined as the minimum of the sum-square magnitude of the approximation error

n Since PCA uses the eigenvectors of the covariance matrix Σx, it is able to find the independent axes of the data under the unimodal Gaussian assumption

g However, for non-Gaussian or multi-modal Gaussian data, PCA simply de-correlates the axes

n The main limitation of PCA is that it does not consider class separability since it does not take into account the class label of the feature vector

g PCA simply performs a coordinate rotation that aligns the transformed axes with the directions of maximum variance

g There is no guarantee that the directions of maximum variance will contain good features for discrimination!!

Page 28: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition28

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

PCA example

g 3D Gaussian distributionn Mean and covariance are

g RESULTSn First PC has largest variancen PCA projections are de-correlated

-10 -5 05 10

0510

-8

-6

-4

-2

0

2

4

6

8

10

12

x1x2

x3

-15 -10 -5 0 5 10 15

-5

0

5

10

15

y1

y3

-10 -8 -6 -4 -2 0 2 4 6 8 10

-2

0

2

4

6

8

10

12

y2-15 -10 -5 0 5 10 15

-10

-5

0

5

10

y1

y2

[ ]

−−−

−==

1047

441

7125

and250 T

Page 29: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition29

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Linear Discriminant Analysis, two-classes

g The objective of LDA is to perform dimensionality reduction while preserving as much of the class discriminatory information as possiblen Assume a set of D-dimensional samples {x1, x2, …, xN}, N1 of which

belong to class ω1, and N2 to class ω2

n We seek to obtain a scalar y by projecting the samples x onto a line y=wTx

n Of all possible lines we want the one that maximizes the separability of the scalars

x1

x2

x1

x2

Page 30: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition30

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Linear Discriminant Analysis, two-classes

g In order to find a good projection vector, we need to define a measure of separation between the projectionsn We could choose the distance between the projected means

g Where the mean values of the x and y examples are

n However, the distance between projected means is not a very goodmeasure since it does not take into account the standard deviation within the classes

iT

y

T

iyii

xii wxw

N1

yN1~andx

N1

iii

==== ∑∑∑∈∈∈

( )21T

21 w~~)w(J −=−=

x1

x2

µ1

µ2

This axis yields better class separability

This axis has a larger distance between means

Page 31: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition31

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Linear Discriminant Analysis, two-classesn The solution proposed by Fisher is to normalize the difference

between the means by a measure of the within-class varianceg For each class we define the scatter, an equivalent of the variance, as

g and the quantity is called the within-class scatter of the projected examples

n The Fisher linear discriminant is defined as the linear function wTx that maximizes the criterion function

n In a nutshell: we look for a projection where examples from the same class are projected very close to each other and, at the same time, the projected means are as farther apart as possible

( )∑∈

−=iy

2i

2i

~ys~

( )22

21 s~s~ +

22

21

2

21

ssJ(w) ~~

~~

+−

=

x1

x2

µ1

µ2

Page 32: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition32

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Linear Discriminant Analysis, two-classesn Noticing that

n We can express J(w) in terms of x and w as

n It can be shown that the projection vector w* which maximizes J(w) is

( ) ( ) ( )( ) wSwwxxwwxwys iT

S SCATTER CLASS

x

Tii

T

x

2

iTT

y

2i

2i

i

iii

=−−=−=−= ∑∑∑∈∈∈

444 3444 21

~~

( ) ( ) ( )( ) wSwwwww BT

S SCATTERCLASS-BETWEEN

T2121

T2

2T

1T2

21

B

=−−=−=−444 3444 21

~~

( ) wSwwSSwss WT

S SCATTERCLASS-WITHIN

21T2

221

W

=+=+43421

~~

wSwwSw

)w(JW

TB

T

=

( )211

WW

TB

T

wS

wSwwSw

argmaxw* −=

= −

Page 33: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition33

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Linear Discriminant Analysis, C-classes

g Fisher’s LDA generalizes for C-class problems very gracefullyn For the C class problem we will seek (C-1) projection vectors wi, which can be

arranged by columns into a projection matrix W=[w1|w2|…|wC-1]

g The solution uses a slightly different formulationn The within-class scatter matrix is similar as the two-class problem

n The generalization of the between-class scatter matrix is

g Where µ is the mean of the entire dataset

n The objective function becomes

g Recall that we are looking for a projection that, in some sense, maximizes the ratio of between-class to within-class scatter:

n Since the projection is not scalar (it has C-1 dimensions), we have used the determinant of the scatter matrices to obtain a scalar criterion function

( )( )∑ ∑∑= ∈=

−−==C

1i x

Tii

C

1iiW

i

xxSS

( )( )∑=

−−=C

1i

TiiiB NS

WSW

WSW

S~S~

)W(JW

T

BT

W

B==

Page 34: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition34

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

n It can be shown that the optimal projection matrix W* is the one whose columns are the eigenvectors corresponding to the largest eigenvalues of the following generalized eigenvalue problem

g NOTEn SB is the sum of C matrices of rank one or less and the mean vectors

are constrained by ΣNiµi=Nµg Therefore, SB will be at most of rank (C-1)

g This means that only (C-1) of the eigenvalues λi will be non-zero

n Therefore, LDA produces at most C-1 feature projectionsg If the classification error estimates establish that more features are needed,

some other method must be employed to provide those additional features

Linear Discriminant Analysis, C-classes

[ ] ( ) 0wSSWSWWSW

argmaxw||w|wW* *iWiB

WT

BT

*1C

*2

*1 =−⇒

== −L

Page 35: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition35

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

PCA Versus LDA

-0.02 -0.01 0 0.01 0.02 0.03 0.04 0.05

-5

0

5

10

15

x 10-3

1 1111

1111

1

1111111

11

222

2

2

2222

22

22

2

22

22 22223

3

333333 33

33

3

3

3333

3

4444

444444

44

4

4

44

444

555

55555 55

5

5

55 555

5555

66 6

66

666 666666

666

666

7

7

7

7

77

777

7

7

777777

77

88

8

8

888 888 888888

8

88

9 99999999999

999 9999

10101010101010101010

1010101010

101010

10

axis 1

axis

2LDA

-10 -5 0 5 10

-10

-8

-6

-4

-2

0

2

4

6

11

11

1

1

11

1

11

1

1

1

11 11

1

222

22

2

2

2

2

2

22

2

2

22

2

222

2

2

33

3 3

3333

3

3

33

3

3

333

33

44

44

4

4

4

4

4

44

4

4

4

4

44

4

4

5

55

55

5

5

55

5

5

5

5

5

5

5

55

5

55

66

66

6

66

66

6

66

6 6

6

6

666

6

7

7

77 7

777

777

7

7

7

77

7

77

8

8

8 8

888

88

8

8

8

8

8

88

8

88

9

99

999

99 99

9 99

999

9

99

10

10

10

101010

10101010

10

1010

10

1010 10 10

10

axis 1

axis

2

-0.02

0

0.02

0.04 -50

510

15

x 10-3

-15

-10

-5

0

5

x 10-3

33 33 33333333 333 3333

6

7777777

67

77

8

76

76

6

8

10

77

10

6

10

66

8

1010

7

8

10

61

410

6

1 77

10

1 716

7

10

1

6

1010

8

6

8

1

8 5

1666

11

5

15

6

1010104

16

1161

4

10

8

10

8

6

5

188

10

88

10

589

4 4444

9

1

9

1

4

8 5851

10

4

551

8 5895 58 952

52

922

55

4

25

4

9

295

229

2

4

22 22

59

4

9

4

295

22

995

4

992

4

2

44

922292

9

axis 1axis 2

axis

3

-10-5

05

10 -10

-5

0

5

-5

0

5

10

7

3

10

53

4

101010

33

10

1

104

10

35

4

33

11

7

2

10

6 6663

4

5383

410

6

81

47

81

46

233

2

7

51 816

7

2 82

74

166

1

677

6

23

858

9

10

9

4

9

65

44

51

7

16

88

10

2

10

52619

10

6

107

4

85

6 19 8

10

91

41

77

35 3

10

3

1 2

93

7

10

295

298

4

9

10

3528

46

22 9

474

298

6

92

7

5

7

9 3

7

1

4

8

7

522

4

12

7

95

6

10

59

28885

51 599

55

922

8

6

axis 1axis 2

axis

3

PCA

Page 36: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition36

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Limitations of LDA

g LDA assumes unimodal Gaussian likelihoodsn If the densities are significantly non-Gaussian, LDA may not preserve

any complex structure of the data needed for classification

g LDA will fail when the discriminatory information is not in the mean but rather in the variance of the data

ω1 ω2

ω2 ω1

µ1=µ2=µ

ω1

ω2

µ1

µ2

µ1=µ2=µω1

ω2

x1

x2

LDAPCA

Page 37: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition37

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Limitations of LDA

g LDA has a tendency to overfit training datan To illustrate this problem, we generate an artificial dataset

g Three classes, 50 examples per class, with the exact same likelihood: a multivariate Gaussian with zero mean and identity covariance

g As we arbitrarily increase the number of dimensions, classes appear to separate better, even though they come from the same distribution

11

1

1

1

1

1

11

1

1

1

1

11

1

11

11

1

1

11

1

1

11

1

1

11

1 11

111

1

11

11

1

11

1

11 12

2

2

2

2

2

2

22

2

2

222

2

2

22

2

2

2

2

2

22

22

2

2

2

22

22

22

22

2

2

2

2

22

22

2 222

3

3

3

33

33

3

3

3

3

3

33

3

33

3

3

33

3

3

3

33

3

3

3

3

3

3

3

3

3

3

33

33

3

3

33 33

3

333

10 dimensions

1

1

111

1

1

1

1

11

1

1

1

11

11

1

1

1

1

1

1

1

1

1

1 1

1

1

1

1

1

1

1

11

1

1

11

11

1

1

1

11

12

2222

22

22

2

2

22

22

22

22

2

22

2

2

22

2

2

22 22

2

2

2

2

2

222

2

2

2

2

2

2 2

2

2

2

33

3

3

3

3

3

3 3

33

3

33

3

3

3

3

33 3

3

33

3

3

33

3

3

33

3333

33

3

3

3

3 3

3

3

33

3

3

3

50 dimensions

1

11

11

1

1

1

1

11 1

1

111

1

1

1

11

1 1

1

11

1

11

11 11

111

1

11 1

11

1

11 1 11

1

1

2

22

22222 2

22

2 22

2

2

22

2

22

2

222

22

2

2

22

22

2

22

22

2

2

2

2

22

22

2 222

3

3

33

33 33

3 33

33

333

3

3

3

33

3

3

3

333 33 333

3

3

333

3

3

3

3

3

33

3

3 33

3

3

100 dimensions

11 11 11 1111

11

1

11111

111 11

111

11

1

111

1111

1111 111

1 11

111

1

2222222222 222 2

222

22

22

222222 22222222

2

22222 2

2 2 222222

33333 3333

333333333

333 333 33

33

333

33

333

3333

333 33

33

333

150 dimensions

Page 38: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition38

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Feature Subset Selection

g Why Feature Subset Selection?n Feature Subset Selection is necessary in a number of situations

g Features may be expensive to obtainn You evaluate a large number of features (sensors) in the test bed and select only

a few for the final implementation

g You may want to extract meaningful rules from your classifiern When you transform or project, the measurement units (length, weight, etc.) of

your features are lost

g Features may not be numericn A typical situation in the machine learning domain

g Although FSS can be thought of as a special case of feature extraction (think of an identity projection matrix with a few diagonal zeros), in practice it is a quite different problem

n FSS looks at dimensionality reduction from a different perspective

n FSS has a unique set of methodologies

Page 39: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition39

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Search strategy and objective functiong Feature Subset Selection requires

n A search strategy to select candidate subsets n An objective function to evaluate candidates

g Search Strategyn Exhaustive evaluation involves feature

subsets for a fixed value of M, and 2N subsets if M must be optimized as wellg This number of combinations is unfeasible, even

for moderate values of M and Nn For example, exhaustive evaluation of 10 out of

20 features involves 184,756 feature subsets; exhaustive evaluation of 10 out of 20 involves more than 1013 feature subsets

n A search strategy is needed to explore the space of all possible feature combinations

g Objective Functionn The objective function evaluates candidate

subsets and returns a measure of their “goodness”, a feedback that is used by the search strategy to select new candidates

M

N

Feature Subset Selection

“Goodness”Featuresubset

PRalgorithm

Objectivefunction

Search

Training data

Final feature subset

Complete feature set

Page 40: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition40

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Objective function

g Objective functions are divided in two groupsn Filters: The objective function evaluates feature subsets by their

information content, typically interclass distance, statistical dependence or information-theoretic measures

n Wrappers: The objective function is a pattern classifier, which evaluates feature subsets by their predictive accuracy (recognition rate on test data) by statistical resampling or cross-validation

Filter FSS

Informationcontent

Featuresubset

MLalgorithm

Objectivefunction

Search

Training data

Final feature subset

Complete feature set

Wrapper FSS

Predictiveaccuracy

Featuresubset

PRalgorithm

PRalgorithm

Search

Training data

Final feature subset

Complete feature set

Page 41: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition41

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Filter typesg Distance or separability measures

n These methods use distance metrics to measure class separability:g Distance between classes: Euclidean, Mahalanobis, etc.g Determinant of SW

-1SB (LDA eigenvalues)

g Correlation and information-theoretic measuresn These methods assume that “good” subsets contain features highly

correlated with (predictive of) the class, yet uncorrelated with each otherg Linear relation measures

n Linear relationship between variables can be measured using the correlation coefficient J(YM)

g where ρ is a correlation coefficient

g Non-Linear relation measuresn Correlation is only capable of measuring linear dependence. A more powerful

method is the mutual information I(Yk;C)

g The mutual information between feature vector and class label I(YM;C) measures the amount by which the uncertainty in the class H(C) is decreased by knowledge of the feature vector H(C|YM), where H(·) is the entropy function

( )∑∑

= +=

== M

1i

M

1ijij

M

1iic

MYJ

( ) ( ) ( ) ( ) ( ) ( )( ) ( )∑ ∫

=

=−==C

1c Y cM

cMcMMMM

M

dxPYP,YP

lg,YPY|CHCHC;YIYJ

Page 42: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition42

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Filters Vs. Wrappersg Wrappers

n Advantagesg Accuracy: wrappers generally achieve better recognition rates than filters since they

are tuned to the specific interactions between the classifier and the datasetg Ability to generalize: wrappers have a mechanism to avoid overfitting, since they

typically use cross-validation measures of predictive accuracy

n Disadvantagesg Slow execution: since the wrapper must train a classifier for each feature subset (or

several classifiers if cross-validation is used), the method is unfeasible for computationally intensive classifiers

g Lack of generality: the solution lacks generality since it is tied to the bias of the classifier used in the evaluation function

g Filtersn Advantages

g Fast execution: Filters generally involve a non-iterative computation on the dataset, which can execute much faster than a classifier training session

g Generality: Since filters evaluate the intrinsic properties of the data, rather than their interactions with a particular classifier, their results exhibit more generality: the solution will be “good” for a larger family of classifiers

n Disadvantagesg Tendency to select large subsets: Since the filter objective functions are generally

monotonic, the filter tends to select the full feature set as the optimal solution. This forces the user to select an arbitrary cutoff on the number of features to be selected

Page 43: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition43

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Naïve sequential feature selectiong One may be tempted to evaluate each

individual feature separately and select those M features with the highest scores

n Unfortunately, this strategy will very rarely work since it does not account for feature dependence

g An example will help illustrate the poor performance of this naïve approach

n The scatter plots show a 4-dimensional pattern recognition problem with 5 classes

g The objective is to select the best subset of 2 features using the naïve sequential FSS procedure

n A reasonable objective function will generate the following feature ranking: J(x1)>J(x2)≈J(x3)>J(x4)

g The optimal feature subset turns out to be {x1, x4}, because x4 provides the only information that x1needs: discrimination between classes ω4 and ω5

g If we were to choose features according to the individual scores J(xk), we would choose x1 and either x2 or x3, leaving classes ω4 and ω5 non separable

ω2

ω3

ω4ω5

ω1

x1

x2

ω1

ω2

ω3

ω5

ω4

x4

x3

Page 44: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition44

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Search strategies

g FSS search strategies can be grouped in three categories*n Sequential

g These algorithms add or remove features sequentially, but have a tendency to become trapped in local minima

n Sequential Forward Selectionn Sequential Backward Selection

n Sequential Floating Selection

n Exponentialg These algorithms evaluate a number of subsets that grows exponentially

with the dimensionality of the search spacen Exhaustive Search (already discussed)n Branch and Boundn Beam Search

n Randomizedg These algorithms incorporating randomness into their search procedure to

escape local miniman Simulated Annealingn Genetic Algorithms

*The subsequent description of FSS is borrowed from [Doak, 1992]

Page 45: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition45

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Sequential Forward Selection (SFS)

g Sequential Forward Selection is a simple greedy search

1. Start with the empty set Y={∅}2. Select the next best feature3. Update Yk+1=Yk+x; k=k+14. Go to 2

1. Start with the empty set Y={∅}2. Select the next best feature3. Update Yk+1=Yk+x; k=k+14. Go to 2

( )[ ]xYJmaxargx kYXx k

+=−∈

+

Empty feature set

Full feature set

g Notesn SFS performs best when the optimal subset has a

small number of featuresg When the search is near the empty set, a large number

of states can be potentially evaluated

g Towards the full set, the region examined by SFS is narrower since most of the features have already been selected

g The main disadvantage of SFS is that it is unable to remove features that become obsolete with the addition of new features

Page 46: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition46

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Sequential Backward Selection (SBS)

g Sequential Backward Selection works in the opposite manner as SFS

1. Start with the full set Y=X2. Remove the worst feature3. Update Yk+1=Yk-x; k=k+14. Go to 2

1. Start with the full set Y=X2. Remove the worst feature3. Update Yk+1=Yk-x; k=k+14. Go to 2

( )[ ]xYJmaxargx kYx k

−=∈

g Notesn SBS works best when the optimal feature subset

has a large number of features, since SBS spends most of its time visiting large subsets

n The main limitation of SBS is its inability to reevaluate the usefulness of a feature after it has been discarded

Empty feature set

Full feature set

Page 47: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition47

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Branch and Boundg The Branch and Bound algorithm is

guaranteed to find the optimal feature subset under the monotonicity assumption

n The monotonicity assumption states that the addition of features can only increase the value of the objective function, this is

n Branch and Bound starts from the full set and removes features using a depth-first strategy

g Nodes whose objective function are lower than the current best are not explored since the monotonicity assumption ensures that their children will not contain a better solution

( ) ( ) ( ) ( )N21321211 iiiiiiiii x,, x,xJ x, x,xJ x,xJxJ LL <<<<

Empty feature set

Full feature set

Page 48: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition48

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Branch and Boundg Algorithm

n The algorithm is better explained by considering the subsets of M’=N-M features already discarded, where N is the dimensionality of the state space and M is the desired number of features

n Since the order of the features is irrelevant, we will only consider an increasing ordering i1<i2<...iM’ of the feature indices, this will avoid exploring states that differ only in the ordering of their features

n The Branch and Bound tree for N=6 and M=2 is shown below (numbers indicate features that are being removed)

g Notice that at the level directly below the root we only consider removing features 1, 2 or 3, since a higher number would not allow sequences (i1<i2<i3<i4) with four indices

4 5 6 5 6 6 5 6 6 6 5 6 6

3 4 5

2

4 5

3

5

4

1

4 5

3

6

5

4

2

6

5

4

3

Page 49: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition49

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

1. Initialize: α=-∞, k=02. Generate successors of the current node and store them in LIST(k)3. Select new node

if LIST(k) is empty go to Step 5

else

4. Check boundif

go to 5else if k=M’ (we have the desired number of features)

go to 6else

k=k+1go to 2

5. Backtrack to lower levelset k=k-1if k=0

terminate algorithmelse

go to 36. Last level

go to 5

1. Initialize: α=-∞, k=02. Generate successors of the current node and store them in LIST(k)3. Select new node

if LIST(k) is empty go to Step 5

else

4. Check boundif

go to 5else if k=M’ (we have the desired number of features)

go to 6else

k=k+1go to 2

5. Backtrack to lower levelset k=k-1if k=0

terminate algorithmelse

go to 36. Last level

go to 5

( )[ ])k(LISTfromidelete

j,x,...x,xJmaxargi

k

iii)k(LISTj

k 1k21 −∈

=

Branch and Bound

( )x,...x,xJk21 iii <

( ) { }k211k21 iii

*’Miii x,...x,xYandj,x,...x,xJSet ==

Page 50: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition50

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Beam Search

g Beam Search is a variation of best-first search with a bounded queue to limit the scope of the searchn The queue organizes states from best to worst,

with the best states placed at the head of the queue

n At every iteration, BS evaluates all possible states that result from adding a feature to the feature subset, and the results are inserted into the queue in their proper locationsg It is trivial to notice that BS degenerates to

Exhaustive search if there is no limit on the size of the queue. Similarly, if the queue size is set to one, BS is equivalent to Sequential Forward Selection

Empty feature set

Full feature set

Page 51: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition51

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Beam Search

g This example illustrates BS for a 4D space and a queue of size 3n BS cannot guarantee that the optimal subset is found:

g in the example, the optimal is 2-3-4(9), which is never exploredg however, with the proper queue size, Beam Search can avoid local minimal by

preserving solutions from varying regions in the search space

4(5)

3(7) 4(3) 4(5) 4(9)

4(5)

3(7)

2(6)

4(3) 4(5)

3(5) 4(2)

1(5)

3(8)

2(3)

4(6)

3(4)

4(5)

4(9)

4(1)

root

4(5)

3(7)

2(6)

4(3) 4(5)

3(5) 4(2)

1(5)

3(8)

2(3)

4(6)

3(4)

4(5)

4(9)

4(1)

root

4(5)

3(7)

2(6)

4(3) 4(5)

3(5) 4(2)

1(5)

3(8)

2(3)

4(6)

3(4)

4(5)

4(9)

4(1)

root

2(6) 3(5) 4(2)

1(5)

3(8)

2(3)

4(6)

3(4)

4(5)

4(1)

root

LIST={∅}

LIST={1(5), 3(4), 2(3)}

LIST={1-2(6), 1-3(5), 3(4)}

LIST={1-2-3(7), 1-3(5), 3(4)}

Page 52: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition52

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Genetic Algorithms

g Genetic algorithms are optimization techniques that mimic the evolutionary process of “survival of the fittest”

n Starting with an initial random population of solutions, evolve new populations by mating (crossover) pairs of solutions and mutating solutions according to their fitness (an objective function)

n The better solutions are more likely to be selected for the mating and mutation operators and, therefore, carry their “genetic code” from generation to generation

n In FSS, individual solutions are represented with a binary number (1 if the feature is selected, 0 otherwise)

1. Create an initial random population2. Evaluate initial population3. Repeat until convergence (or a number of generations)

3a. Select the fittest individuals in the population3b. Perform crossover on selected individuals to create offspring3c. Perform mutation on selected individuals3d. Create new population from old population and offspring3e. Evaluate the new population

1. Create an initial random population2. Evaluate initial population3. Repeat until convergence (or a number of generations)

3a. Select the fittest individuals in the population3b. Perform crossover on selected individuals to create offspring3c. Perform mutation on selected individuals3d. Create new population from old population and offspring3e. Evaluate the new population

Empty feature set

Full feature set

Page 53: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition53

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Genetic operators

g Single-point crossovern Select two individuals (parents) according to their fitnessn Select a crossover point

n With probability Pc (0.95 is reasonable) create two offspring by combining the parents

g Binary mutationn Select an individual according to its fitness

n With probability PM (0.01 is reasonable) mutate each one of its bits

01001010110

11010110000

11011010110

01000110000

Parenti

Parentj

Offspringi

Offspringj

Crossover

Crossover point selected randomly

11010110000Individual Mutation 11001010111 Offspringi

Mutated bits

Page 54: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition54

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Selection methods

g The selection of individuals is based on their fitness g We will describe a selection method called Geometric selection

n Several methods are available: Roulette Wheel, Tournament Selection…

0 10 20 30 40 50 60 70 80 90 1000

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Individual rank

Sel

ecti

on p

roba

bilit

y

q=0.08

q=0.02

q=0.01

q=0.04

g Geometric selectionn The probability of selecting the rth best

individual is given by the geometric probability mass function

g q is the probability of selecting the best individual (0.05 is a reasonable value)

n The geometric distribution assigns higher probability to individuals ranked better, but also allows unfit individuals to be selected

g In addition, it is typical to carry the best individual of each population to the next one

n This is called the Elitist Model

( ) ( ) 1-rq-1qrP =

Page 55: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition55

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

GAs, parameter choices for Feature selection

g The choice of crossover rate PC is not criticaln You will want a value close to 1.0 to have a large number of offspring

g The choice of mutation rate PM is very criticaln An optimal choice of PM will allow the GA to explore the more promising

regions while avoiding getting trapped in local minimag A large value (i.e., PM>0.25) will not allow the search to focus on the better

regions, and the GA will perform like random searchg A small value (i.e., close to 0.0) will not allow the search to escape local

minima

g The choice of ‘q’, the probability of selecting the best individual is also critical

n An optimal value of ‘q’ will allow the GA to explore the most promising solution, and at the same time provide sufficient diversity to avoid early convergence of the algorithm

g In general, poorly selected control parameters will result in sub-optimal solutions due to early convergence

Page 56: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition56

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Accuracy Complexity Advantages Disadvantages

Exhaustive Always finds theoptimal solution Exponential High accuracy High complexity

SequentialGood if no

backtrackingneeded

Quadratic O(NEX

2) Simple and fastCannot

backtrack

RandomizedGood with proper

controlparameters

Generally lowDesigned toescape local

minima

Difficult tochoose goodparameters

Search Strategies, summary

g A highly recommended review of this material

Justin Doak“An evaluation of feature selection methods and their application to Computer Security”University of California at Davis, Tech Report CSE-92-18

Page 57: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition57

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

SECTION III: Classification

g Discriminant functionsg The optimal Bayes classifierg Quadratic classifiersg Euclidean and Mahalanobis metricsg K Nearest Neighbor Classifiers

Page 58: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition58

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Discriminant functionsg A convenient way to represent a pattern classifier is in terms of

a family of discriminant functions gi(x) with a simple MAX gate as the classification rule

g How do we choose the discriminant functions gi(x)n Depends on the objective function to minimize

g Probability of errorg Bayes Risk

x2x2 x3

x3 xdxd

g1(x)g1(x)

x1x1

g2(x)g2(x) gC(x)gC(x)

Select maxSelect max

CostsCosts

Class assignment

Discriminant functions

Features

ij(x)jg(x)gifclasstoxAssign ii ≠∀>

Page 59: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition59

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Minimizing probability of error

g Probability of error P[error|x] is “the probability of assigning x to the wrong class”

n For a two-class problem, P[error|x] is simply

g It makes sense that the classification rule be designed to minimize the average probability of error P[error] across all possible values of x

g To ensure P(error) is minimum we minimize P(error|x) by choosing the class with maximum posterior P(ωi|x) at each x

n This is called the MAXIMUM A POSTERIORI (MAP) RULEg And the associated discriminant functions become

=12

21

decideweifx)|P(

decideweifx)|P(x)|P(error

∫∫+∞

∞−

+∞

∞−== x)P(x)dx|P(errorx)dxP(error,P(error)

x)|P((x)g iMAPi =

Page 60: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition60

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Minimizing probability of errorg We “prove” the optimality of

the MAP rule graphicallyn The right plot shows the

posterior for each of the two classes

n The bottom plots shows the P(error) for the MAP rule and another rule

n Which one has lower P(error) (color-filled area) ?

x

P(w

i|x)

ChooseRED

ChooseBLUE

ChooseRED

THE MAP RULE

ChooseRED

ChooseBLUE

ChooseRED

THE “OTHER” RULE

Page 61: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition61

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Quadratic classifiers

g Let us assume that the likelihood densities are Gaussian

g Using Bayes rule, the MAP discriminant functions become

n Eliminating constant terms

n We take natural logs (the logarithm is monotonically increasing)

g This is known as a Quadratic Discriminant Functiong The quadratic term is know as the Mahalanobis distance

−∑−−

∑= − )(x)(x

21

exp�(2

1)|P(x i

1i

Ti1/2

in/2i

P(x)1

)P()(x)(x21

exp�(2

1P(x)

))P(|P(xx)|P((x)g ii

1i

Ti1/2

in/2

iiii

−∑−−

∑=== −

)P()(x)(x21

exp(x)g ii1

iT

i

1/2-

ii

−∑−−∑= −

( ) ( ))P(loglog21

-)(x)(x21

(x)g iii1

iT

ii +∑−∑−−= −

Page 62: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition62

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Mahalanobis distance

g The Mahalanobis distance can be thought of vector distance that uses a ∑i

-1 norm

n ∑-1 can be thought of as a stretching factor on the spacen Note that for an identity covariance matrix (∑i=I), the Mahalanobis

distance becomes the familiar Euclidean distance

g In the following slides we look at special cases of the Quadratic classifier

n For convenience we will assume equiprobable priors so we can drop the term log(P(ωi))

µx2

x1

-x2

i =

K -x2

i 1 =−∑

y)(xy)(xy-x 1i

T21

i−∑−= −

∑ −

DistancesMahalanobi

Page 63: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition63

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Special case I: Σi=σ2Ig In this case, the discriminant

becomes

n This is known as a MINIMUM DISTANCE CLASSIFIER

n Notice the linear decision boundaries

)(x)(x(x)g iT

ii −−−=

Page 64: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition64

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Special case 2: Σi= Σ (Σ diagonal)g In this case, the discriminant

becomes

n This is known as a MAHALANOBIS DISTANCE CLASSIFIER

n Still linear decision boundaries

)(x)(x21

(x)g i1T

ii −∑−−= −

Page 65: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition65

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Special case 3: Σi=Σ (Σ non-diagonal)g In this case, the discriminant

becomes

n This is also known as a MAHALANOBIS DISTANCE CLASSIFIER

n Still linear decision boundaries

)(x)(x21

(x)g i1

iT

ii −∑−−= −

Page 66: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition66

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

g In this case the quadratic expression cannot be simplified any further

g Notice that the decision boundaries are no longer linear but quadratic

Case 4: Σi=σi2I, example

Zoomout

Page 67: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition67

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

g In this case there are no constraints so the quadratic expression cannot be simplified any further

g Notice that the decision boundaries are also quadratic

Case 5: Σi≠Σj general case, example

Zoomout

Page 68: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition68

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Limitations of quadratic classifiers

g The fundamental limitation is the unimodal Gaussian assumptionn For non-Gaussian or multimodal

Gaussian, the results may be significantly sub-optimal

g A practical limitation is associated with the minimum required size for the datasetn If the number of examples per class is less than the number of

dimensions, the covariance matrix becomes singular and, therefore, its inverse cannot be computed

g In this case it is common to assume the same covariance structure for all classes and compute the covariance matrix using all the examples, regardless of class

Page 69: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition69

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Conclusionsg We can extract the following conclusions

n The Bayes classifier for normally distributed classes is quadraticn The Bayes classifier for normally distributed classes with equal

covariance matrices is a linear classifiern The minimum Mahalanobis distance classifier is optimum for

g normally distributed classes and equal covariance matrices and equal priorsn The minimum Euclidean distance classifier is optimum for

g normally distributed classes and equal covariance matrices proportional to the identity matrix and equal priors

n Both Euclidean and Mahalanobis distance classifiers are linear

g The goal of this discussion was to show that some of the most popular classifiers can be derived from decision-theoretic principles and some simplifying assumptions

n It is important to realize that using a specific (Euclidean or Mahalanobis) minimum distance classifier implicitly corresponds to certain statistical assumptions

n The question whether these assumptions hold or don’t can rarely be answered in practice; in most cases we are limited to posting and answering the question “does this classifier solve our problem or not?”

Page 70: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition70

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

K Nearest Neighbor classifier

g The kNN classifier is based on non-parametric density estimation techniques

n Let us assume we seek to estimate the density function P(x) from a dataset of examples

n P(x) can be approximated by the expression

n The volume V is determined by theD-dim distance Rk

D(x) between xand its k nearest neighbor

g Where cD is the volume of the unit sphere in D dimensions

≅Vinsideexamplesofnumbertheisk

examplesofnumbertotaltheisN

xgsurroundinvolumetheisV

whereNVk

P(x)

R

V=πR2

x2

RN

kP(x) =

(x)RcNk

NVk

P(x)DkD ⋅⋅

=≅

Page 71: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition71

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

K Nearest Neighbor classifierg We use the previous result to estimate the posterior probability

n The unconditional density is, again, estimated with

n And the priors can be estimated by

n The posterior probability then becomes

n Yielding discriminant functions

g This is known as the k Nearest Neighbor classifier

VNk

)|P(xi

ii =

NN

)P( ii =

kk

NVk

NN

VNk

P(x)))P(|P(x

x)|P( i

i

i

i

iii =

⋅==

kk

(x)g ii =

Page 72: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition72

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

K Nearest Neighbor classifier

g The kNN classifier is a very intuitive method n Examples are classified based on their similarity with training data

g For a given unlabeled example xu∈ℜD, find the k “closest” labeled examples in the training data set and assign xu to the class that appears most frequently within the k-subset

g The kNN only requiresn An integer kn A set of labeled examples

n A measure of “closeness”

-0.06 -0.05 -0.04 -0.03 -0.02 -0.01 0 0.01 0.02 0.03

-20

-15

-10

-5

0

511

1111111

1111 1111

11222

2222

2222

22

22222

22

223

3

3333 3

33333

33

33

33

3

44 444444444

44

4

4

4

44

4

555

555

555 5555 55555555

6

666 6

6666 666

6

6666666

7

7

77

7777 7

7

7777 777

77

888

88

888888

8888 88

88

99

9

999999999

9999

999

101010101010101010101010

1010

1010

1010

10

axis 1

axis

2

?

-0.06 -0.05 -0.04 -0.03 -0.02 -0.01 0 0.01 0.02 0.03

-20

-15

-10

-5

0

511

1111111

1111 1111

11222

2222

2222

22

22222

22

223

3

3333 3

33333

33

33

33

3

44 444444444

44

4

4

4

44

4

555

555

555 5555 55555555

6

666 6

6666 666

6

6666666

7

7

77

7777 7

7

7777 777

77

888

88

888888

8888 88

88

99

9

999999999

9999

999

101010101010101010101010

1010

1010

1010

10

axis 1

axis

2

??

Page 73: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition73

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

kNN in action: example 1

g We generate data for a 2-dimensional 3-class problem, where the class-conditional densities are multi-modal, and non-linearly separable

g We used kNN withn k = fiven Metric = Euclidean distance

Page 74: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition74

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

kNN in action: example 2

g We generate data for a 2-dim 3-class problem, where the likelihoods are unimodal, and are distributed in rings around a common mean

n These classes are also non-linearly separable

g We used kNN withn k = fiven Metric = Euclidean distance

Page 75: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition75

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

kNN versus 1NN1-NN 5-NN 20-NN

Page 76: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition76

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Characteristics of the kNN classifierg Advantages

n Analytically tractable, simple implementationn Nearly optimal in the large sample limit (N→∞)

g P[error]Bayes<P[error]1-NNR<2P[error]Bayes

n Uses local information, which can yield highly adaptive behaviorn Lends itself very easily to parallel implementations

g Disadvantagesn Large storage requirementsn Computationally intensive recalln Highly susceptible to the curse of dimensionality

g 1NN versus kNNn The use of large values of k has two main advantages

g Yields smoother decision regionsg Provides probabilistic information: The ratio of examples for each class

gives information about the ambiguity of the decision

n However, too large values of k are detrimentalg It destroys the locality of the estimation

Page 77: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition77

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Section IV: Validation

g Motivationg The Holdoutg Re-sampling techniquesg Three-way data splits

Page 78: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition78

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Motivation

g Validation techniques are motivated by two fundamental problems in pattern recognition: model selection and performance estimation

g Model selectionn Almost invariably, all pattern recognition techniques have one or

more free parametersg The number of neighbors in a kNN classification ruleg The network size, learning parameters and weights in MLPs

n How do we select the “optimal” parameter(s) or model for a given classification problem?

g Performance estimationn Once we have chosen a model, how do we estimate its

performance?g Performance is typically measured by the TRUE ERROR RATE,

the classifier’s error rate on the ENTIRE POPULATION

Page 79: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition79

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Motivation

g If we had access to an unlimited number of examples these questions have a straightforward answer

n Choose the model that provides the lowest error rate on the entire population and, of course, that error rate is the true error rate

g In real applications we only have access to a finite set of examples, usually smaller than we wanted

n One approach is to use the entire training data to select our classifier and estimate the error rate

g This naïve approach has two fundamental problemsn The final model will normally overfit the training data

g This problem is more pronounced with models that have a large number of parameters

n The error rate estimate will be overly optimistic (lower than the true error rate)g In fact, it is not uncommon to have 100% correct classification on training

data

n A much better approach is to split the training data into disjoint subsets: the holdout method

Page 80: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition80

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

The holdout methodg Split dataset into two groups

n Training set: used to train the classifiern Test set: used to estimate the error rate of the trained classifier

g A typical application the holdout method is determining a stopping point for the back propagation error

Training Set Test Set

Total number of examples

Epochs

MSE

Training set error

Test set errorStopping point

Page 81: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition81

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

The holdout method

g The holdout method has two basic drawbacksn In problems where we have a sparse dataset we may not be able to

afford the “luxury” of setting aside a portion of the dataset for testingn Since it is a single train-and-test experiment, the holdout estimate of

error rate will be misleading if we happen to get an “unfortunate” split

g The limitations of the holdout can be overcome with a family of resampling methods at the expense of more computations

n Cross Validationg Random Subsamplingg K-Fold Cross-Validation

g Leave-one-out Cross-Validation

n Bootstrap

Page 82: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition82

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Random Subsampling

g Random Subsampling performs K data splits of the datasetn Each split randomly selects a (fixed) no. examples without replacementn For each data split we retrain the classifier from scratch with the training

examples and estimate Ei with the test examples

g The true error estimate is obtained as the average of the separate estimates Ei

n This estimate is significantly better than the holdout estimate

Total number of examples

Experiment 1

Experiment 2

Experiment 3

Test example

∑=

=K

1iiE

K1

E

Page 83: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition83

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

K-Fold Cross-validation

g Create a K-fold partition of the the datasetn For each of K experiments, use K-1 folds for training and the remaining

one for testing

g K-Fold Cross validation is similar to Random Subsampling n The advantage of K-Fold Cross validation is that all the examples in the

dataset are eventually used for both training and testing

g As before, the true error is estimated as the average error rate

Total number of examples

Experiment 1

Experiment 2

Experiment 3Test examples

Experiment 4

∑=

=K

1iiE

K1

E

Page 84: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition84

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Leave-one-out Cross Validation

g Leave-one-out is the degenerate case of K-Fold Cross Validation, where K is chosen as the total number of examplesn For a dataset with N examples, perform N experimentsn For each experiment use N-1 examples for training and the remaining

example for testing

g As usual, the true error is estimated as the average error rate on test examples

∑=

=N

1iiE

N1

E

Total number of examples

Experiment 1

Experiment 2

Experiment 3

Experiment N

Single test example

Page 85: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition85

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

How many folds are needed?

g With a large number of folds + The bias of the true error rate estimator will be small (the estimator will

be very accurate)- The variance of the true error rate estimator will be large- The computational time will be very large as well (many experiments)

g With a small number of folds + The number of experiments and, therefore, computation time are

reduced+ The variance of the estimator will be small- The bias of the estimator will be large (conservative or larger than the

true error rate)

g In practice, the choice of the number of folds depends on the size of the dataset

n For large datasets, even 3-Fold Cross Validation will be quite accuraten For very sparse datasets, we may have to use leave-one-out in order to

train on as many examples as possible

g A common choice for K-Fold Cross Validation is K=10

Page 86: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition86

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Three-way data splits

g If model selection and true error estimates are to be computed simultaneously, the data needs to be divided into three disjointsets

n Training set: a set of examples used for learning the parameters of the model

g In the MLP case, we would use the training set to find the “optimal” weights with the back propagation rule

n Validation set: a set of examples used to tune the meta-parameters of a model

g In the MLP case, we would use the validation set to find the “optimal” number of hidden units or determine a stopping point for the back propagation algorithm

n Test set: a set of examples used only to assess the performance of a fully-trained classifier

g In the MLP case, we would use the test to estimate the error rate after we have chosen the final model (MLP size and actual weights)

g After assessing the final model with the test set, YOU MUST NOT further tune the model

Page 87: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition87

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Three-way data splits

g Why separate test and validation sets?n The error rate estimate of the final model on validation data will

be biased (smaller than the true error rate) since the validation set is used to select the final model

n After assessing the final model with the test set, YOU MUST NOT tune the model any further

g Procedure outline

n This outline assumes a holdout methodg If CV or Bootstrap are used, steps 3 and 4 have to be repeated for

each of the K folds

1. Divide the available data into training, validation and test set2. Select architecture and training parameters3. Train the model using the training set4. Evaluate the model using the validation set5. Repeat steps 2 through 4 using different architectures and training parameters6. Select the best model and train it using data from the training and validation set7. Assess this final model using the test set

1. Divide the available data into training, validation and test set2. Select architecture and training parameters3. Train the model using the training set4. Evaluate the model using the validation set5. Repeat steps 2 through 4 using different architectures and training parameters6. Select the best model and train it using data from the training and validation set7. Assess this final model using the test set

Page 88: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition88

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Three-way data splits

ΣModel1 Error1

ΣModel2 Error2

ΣModel3 Error3

ΣModel4 Error4

Min

Final Model Σ

Final Error

Test set

Training setValidation set

ModelSelection

ErrorRate

Page 89: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition89

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Section V: Clustering

g Unsupervised learning conceptsg K-meansg Agglomerative clusteringg Divisive clustering

Page 90: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition90

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Unsupervised learningg In this section we investigate a family of techniques which use

unlabeled data: the feature vector X without the class label ωn These methods are called unsupervised since they are not provided the

correct answerg Although unsupervised learning methods may appear to have

limited capabilities, there are several reasons that make them extremely useful

n Labeling large data sets can be costly (i.e., speech recognition)n Class labels may not be known beforehand (i.e., data mining)n Very large datasets can be compressed by finding a small set of

prototypes (i.e., kNN)

g Two major approaches for unsupervised learningn Parametric (mixture modeling)

g A functional form for the underlying density is assumed, and we seek to estimate the parameters of the model

n Non-parametric (clustering)g No assumptions are made about the underlying densities, instead we seek

a partition of the data into clustersn These methods are typically referred to as clustering

Page 91: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition91

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Unsupervised learning

g Non-parametric unsupervised learning involves three stepsn Defining a measure of similarity between examplesn Defining a criterion function for clustering

n Defining an algorithm to minimize (or maximize) the criterion function

g Similarity measuresn The most straightforward similarity measure is distance: Euclidean, city

block or Mahalanobisg Euclidean and city block measures are very sensitive to scaling of the axis,

so caution must be exercised

g Criterion functionn The most widely used criterion function for clustering is the sum-of-

square-error criterion

g This criterion measures how well the data set X={x(1, x(2, …, x(N} is represented by the cluster centers µ={µ(1, µ(2, …, µ(N}

∑∑ ∑≈= ≈

=−=ii xi

i

C

1i x

2

iMSE xN1

wherexJ

Page 92: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition92

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Unsupervised learningg Once a criterion function has been defined, we must find a partition of

the data set that minimizes the criterionn Exhaustive enumeration of all partitions, which guarantees the optimal solution,

is unfeasiblen For example, a problem with 5 clusters and 100 examples yields 1067 different

ways to partition the data

g The common approach is to proceed in an iterative fashionn Find some reasonable initial partition and then n Move samples from one cluster to another in order to reduce the criterion

function

g These iterative methods produce sub-optimal solution but are computationally tractable

g We will consider two groups of iterative methodsn Flat clustering algorithms

g These algorithms produce a set of disjoint clusters (a.k.a. a partition)g Two algorithms are widely used: k-means and ISODATA

n Hierarchical clustering algorithms: g The result is a hierarchy of nested clustersg These algorithms can be broadly divided into agglomerative and divisive approaches

Page 93: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition93

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

The k-means algorithmg The k-means algorithm is a simple clustering procedure that

attempts to minimize the criterion function JMSE iteratively

n The k-means algorithm is widely used in the fields of signal processing and communication for Vector Quantization

g Scalar signal values are usually quantized into a number of levels (typically a power of 2 so the signal can be transmitted in binary)

g The same idea can be extended for multiple channelsn However, rather than quantizing each separate channel, we can obtain a more

efficient signal coding if we quantize the overall multidimensional vector by finding a number of multidimensional prototypes (cluster centers)

g The set of cluster centers is called a “codebook”, and the problem of finding this codebook is normally solved using the k-means algorithm

1. Define the number of clusters2. Initialize clusters by

• an arbitrary assignment of examples to clusters or • an arbitrary set of cluster centers (examples assigned to nearest centers)

3. Compute the sample mean of each cluster4. Reassign each example to the cluster with the nearest mean5. If the classification of all samples has not changed, stop, else go to step 3

1. Define the number of clusters2. Initialize clusters by

• an arbitrary assignment of examples to clusters or • an arbitrary set of cluster centers (examples assigned to nearest centers)

3. Compute the sample mean of each cluster4. Reassign each example to the cluster with the nearest mean5. If the classification of all samples has not changed, stop, else go to step 3

Page 94: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition94

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Hierarchical clusteringg k-means creates disjoint clusters, resulting in a “flat” data

representation (a partition)g Sometimes it is desirable to obtain a hierarchical representation

of data, with clusters and sub-clusters arranged in a tree-structured fashionn Hierarchical representations are commonly used in the sciences (i.e.,

biological taxonomy)

g Hierarchical clustering methods can be grouped in two general classesn Agglomerative (bottom-up, merging):

g Starting with N singleton clusters, successively merge clusters until one cluster is left

n Divisive (top-down, splitting)g Starting with a unique cluster, successively split the clusters until N singleton

examples are left

Page 95: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition95

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Hierarchical clustering

g The preferred representation for hierarchical clusters is the dendrogramn The dendrogram is a binary tree that shows the structure of the clusters

g In addition to the binary tree, the dendrogram provides the similarity measure between clusters (the vertical axis)

n An alternative representation is based on sets: {{x1, {x2, x3}}, {{{x4, x5}, {x6, x7}}, x8}} but, unlike the dendrogram, sets cannot express quantitative information

x1 x2 x3 x4 x5 x6 x7 x8

High similarity

Low similarity

Page 96: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition96

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Divisive clusteringg Outline

n Defineg NC: Number of clustersg NEX: Number of examples

g How to choose the “worst” clustern Largest number of examplesn Largest variancen Largest sum-squared-errorn ...

g How to split clustersn Mean-median in one feature directionn Perpendicular to the direction of largest variancen …

g The computations required by divisive clustering are more intensive than for agglomerative clustering methods and, thus, agglomerative approaches are more common

1. Start with one large cluster

2. Find “worst” cluster3. Split it

4. If NC<NEX go to 1

1. Start with one large cluster

2. Find “worst” cluster3. Split it

4. If NC<NEX go to 1

Page 97: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition97

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Agglomerative clustering

g Outlinen Define

g NC: Number of clustersg NEX: Number of examples

g How to find the “nearest” pair of clusters

n Minimum distance

n Maximum distance

n Average distance

n Mean distance

1. Start with NEX singleton clusters2. Find nearest clusters

3. Merge them4. If NC>1 go to 1

1. Start with NEX singleton clusters2. Find nearest clusters

3. Merge them4. If NC>1 go to 1

( ) yxmin,dji

yxjimin −=∈∈

( ) yxmax,dji

yx

jimax −=∈∈

( ) ∑ ∑∈ ∈

−=i jx yji

jiavg yxNN1

,d

( ) jijimean ,d −=

Page 98: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition98

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Agglomerative clusteringg Minimum distance

n When dmin is used to measure distance between clusters, the algorithm is called the nearest-neighbor or single-linkage clustering algorithm

n If the algorithm is allowed to run until only one cluster remains, the result is a minimum spanning tree (MST)

n This algorithm favors elongated classes

g Maximum distancen When dmax is used to measure distance between clusters, the algorithm is called

the farthest-neighbor or complete-linkage clustering algorithmn From a graph-theoretic point of view, each cluster constitutes a complete sub-

graphn This algorithm favors compact classes

g Average and mean distancen The minimum and maximum distance are extremely sensitive to outliers since

their measurement of between-cluster distance involves minima or maxima

n The average and mean distance approaches are more robust to outliersn Of the two, the mean distance is computationally more attractive

g Notice that the average distance approach involves the computation of NiNj distances for each pair of clusters

Page 99: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition99

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Conclusions

g Dimensionality reductionn PCA may not find the most discriminatory axesn LDA may overfit the data

g Classifiersn Quadratic classifiers are efficient for unimodal datan kNN is very versatile but is computationally inefficient

g Validationn An fundamental subject, oftentimes overlooked

g Clusteringn Helps understand the underlying structure of the data

ACKNOWLEDGMENTSThis research was supported by award

from NSF/CAREER 9984426.

Page 100: NOSE 2 Summer School Statistical Pattern Recognition ...research.cs.tamu.edu/prism/lectures/talks/nose00.pdf · Ricardo Gutierrez-Osuna Wright State University Statistical Pattern

Ricardo Gutierrez-OsunaWright State University

Statistical Pattern Recognition100

NOSE 2nd Summer SchoolLloret de Mar, Spain, October 2-5, 2000

Referencesg Fukunaga, K., 1990, Introduction to statistical pattern recognition, 2nd edition,

Academic Pr.g Duda, R. O.; Hart, P. E. and Stork, D. G., 2001, Pattern classification, 2nd Ed.,

Wiley.g Bishop, C. M., 1995, Neural networks for Pattern Recognition, Oxford.g Therrien, C. W., 1989, Decision, estimation, and classification : an introduction

to pattern recognition and related topics, Wiley.g Devijver, P. A. and Kittler, J., 1982, Pattern recognition: a statistical approach,

Prentice-Hall. g Ballard, D., 1997, An introduction to Natural Computation, The MIT Press.g Silverman, B. W., 1986, Density estimation for statistics and data analysis,

Chapman & Hall.g Smith, M., 1994, Neural networks for statistical modeling, Van Nostrand

Reinhold.g Doak, J., 1992, An evaluation of feature selection methods and their application

to Computer Security, University of California at Davis, Tech Report CSE-92-18g Kohavi, R. and Sommerfield, D., 1995, "Feature Subset Selection Using the

Wrapper Model: Overfitting and Dynamic Search Space Topology" in Proc. of the 1st Intl’ Conf. on Knowledge Discovery and Data Mining, pp. 192-197.