michigan reu final presentations, august 10, 2006matt jachowski 1 multivariate analysis, tmva, and...

25
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski [email protected] http:// tmva.sourceforge.net/

Upload: helena-kennedy

Post on 31-Dec-2015

223 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 1

Multivariate Analysis, TMVA, and Artificial Neural Networks

Matt Jachowski

[email protected]

http://tmva.sourceforge.net/

Page 2: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 2

Multivariate Analysis

• Techniques dedicated to analysis of data with multiple variables

• Active field – many recently developed techniques rely on computational ability of modern computers

Page 3: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 3

Multivariate Analysis and HEP

• Goal is to classify events as signal or background

• Single event defined by several variables (energy, transverse momentum, etc.)

• Use all the variables to classify the event

• Multivariate analysis!

Page 4: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 4

Multivariate Analysis and HEP

• Rectangular cuts optimization common

Page 5: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 5

Multivariate Analysis and HEP

• Likelihood Estimator analysis also common

• Use of more complicated methods (Neural Networks, Boosted Decision Trees) not so common (though growing) – why?– Difficult to implement– Physicists are skeptical of new methods

Page 6: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 6

Toolkit for Multivariate Analysis (TMVA)

• ROOT-integrated software package with several MVA techniques

• Automatic training, testing, and evaluation of MVA methods

• Guidelines and documentation to describe methods for users – this isn’t a black box!

Page 7: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 7

Toolkit for Multivariate Analysis (TMVA)

• Easy to configure methods

• Easy to “plug-in” HEP data

• Easy to compare different MVA methods

Page 8: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 8

TMVA in Action

Page 9: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 9

TMVA and Me

• TMVA started in October 2005– Still young– Very active group of developers

• My involvement– Decorrelation for Cuts Method (mini project)– New Artificial Neural Network implementation

(main project)

Page 10: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 10

Decorrelated Cuts Method

• Some MVA methods suffer if data has linear correlations– i.e. Likelihood Estimator, Cuts

• Linear correlations can be easily transformed away

• I implemented this for the Cuts Method

Page 11: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 11

Decorrelated Cuts Method

• Find the square root of the covariance matrix (C=C’C’)

• Decorrelate the data

• Apply cuts to decorrelated data

CSSD T TSDSC'

xC'x'

Page 12: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 12

Artificial Neural Networks (ANNs)

• Robust non-linear MVA technique

Page 13: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 13

Page 14: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 14

Training an ANN

• Challenge is training the network

• Like human brain, network learns from seeing data over and over again

• Technical details:Ask me if you’re really interested

Page 15: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 15

MLP

• MLP (Multi-Layer Perceptron) – my ANN implementation for TMVA

• MLP is TMVA’s main ANN

• MLP serves as base for any future ANN developments in TMVA

Page 16: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 16

MLP – Information & Statistics

• Implemented in C++

• Object-Oriented

• 4,000+ lines of code

• 16 classes

Page 17: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 17

Acknowledgements

• Joerg Stelzer

• Andreas Hoecker

• CERN

• University of Michigan

• Ford

• NSF

Page 18: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 18

Questions?

(I have lots of technical slides in reserve that I would be glad to talk about)

Page 19: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 19

Page 20: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 20

Synapses and Neurons

.

.

.

w0j

w1j

wnj

y0

y1

yn

vj yj

v0

v1

vn

),...,,,...,( 00 njjnj wwyyfv

)( jj vy

Page 21: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 21

Synapses and Neurons

jvjje

vy 1

1)(

i

n

iijnjjnj ywwwyyfv

0

00 ),...,,,...,(

yj

vj

Page 22: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 22

Universal Approximation Theorem

Every continuous function that maps intervals of real numbers to some output interval of real numbers can be approximated arbitrarily closely by a multi-layer perceptron with just one hidden layer (with non-linear activation functions).

j

jj bwf )()( xvx j

inputsweights between input and hidden layer

biasnon-linear activation function

weights between hidden and output layer

output

Page 23: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 23

Training an MLPTraining Event:

Network:

x0

x1

x2

x3

y

dxxxxf ),,,( 3210

yxxxxg ),,,( 3210

ydeerror

Page 24: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 24

Training an MLPAdjust weights to minimize error (or an estimator that is some function of the error)

}_{

2 )(2

1)(

neuronsoutputjj nenε

)()()( nyndne jjj

N

navg n

N 1

)(1

Page 25: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu

Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 25

Back-Propagation AlgorithmMake correction in direction of steepest descent

)()()1( nwnwnw ijijij

)(

)()(

nw

nnw

ij

ijij

Corrections made to output layer first, propagated backwards