michigan reu final presentations, august 10, 2006matt jachowski 1 multivariate analysis, tmva, and...
TRANSCRIPT
![Page 1: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/1.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 1
Multivariate Analysis, TMVA, and Artificial Neural Networks
Matt Jachowski
http://tmva.sourceforge.net/
![Page 2: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/2.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 2
Multivariate Analysis
• Techniques dedicated to analysis of data with multiple variables
• Active field – many recently developed techniques rely on computational ability of modern computers
![Page 3: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/3.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 3
Multivariate Analysis and HEP
• Goal is to classify events as signal or background
• Single event defined by several variables (energy, transverse momentum, etc.)
• Use all the variables to classify the event
• Multivariate analysis!
![Page 4: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/4.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 4
Multivariate Analysis and HEP
• Rectangular cuts optimization common
![Page 5: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/5.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 5
Multivariate Analysis and HEP
• Likelihood Estimator analysis also common
• Use of more complicated methods (Neural Networks, Boosted Decision Trees) not so common (though growing) – why?– Difficult to implement– Physicists are skeptical of new methods
![Page 6: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/6.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 6
Toolkit for Multivariate Analysis (TMVA)
• ROOT-integrated software package with several MVA techniques
• Automatic training, testing, and evaluation of MVA methods
• Guidelines and documentation to describe methods for users – this isn’t a black box!
![Page 7: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/7.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 7
Toolkit for Multivariate Analysis (TMVA)
• Easy to configure methods
• Easy to “plug-in” HEP data
• Easy to compare different MVA methods
![Page 8: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/8.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 8
TMVA in Action
![Page 9: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/9.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 9
TMVA and Me
• TMVA started in October 2005– Still young– Very active group of developers
• My involvement– Decorrelation for Cuts Method (mini project)– New Artificial Neural Network implementation
(main project)
![Page 10: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/10.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 10
Decorrelated Cuts Method
• Some MVA methods suffer if data has linear correlations– i.e. Likelihood Estimator, Cuts
• Linear correlations can be easily transformed away
• I implemented this for the Cuts Method
![Page 11: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/11.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 11
Decorrelated Cuts Method
• Find the square root of the covariance matrix (C=C’C’)
• Decorrelate the data
• Apply cuts to decorrelated data
CSSD T TSDSC'
xC'x'
![Page 12: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/12.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 12
Artificial Neural Networks (ANNs)
• Robust non-linear MVA technique
![Page 13: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/13.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 13
![Page 14: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/14.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 14
Training an ANN
• Challenge is training the network
• Like human brain, network learns from seeing data over and over again
• Technical details:Ask me if you’re really interested
![Page 15: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/15.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 15
MLP
• MLP (Multi-Layer Perceptron) – my ANN implementation for TMVA
• MLP is TMVA’s main ANN
• MLP serves as base for any future ANN developments in TMVA
![Page 16: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/16.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 16
MLP – Information & Statistics
• Implemented in C++
• Object-Oriented
• 4,000+ lines of code
• 16 classes
![Page 17: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/17.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 17
Acknowledgements
• Joerg Stelzer
• Andreas Hoecker
• CERN
• University of Michigan
• Ford
• NSF
![Page 18: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/18.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 18
Questions?
(I have lots of technical slides in reserve that I would be glad to talk about)
![Page 19: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/19.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 19
![Page 20: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/20.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 20
Synapses and Neurons
.
.
.
w0j
w1j
wnj
y0
y1
yn
vj yj
v0
v1
vn
),...,,,...,( 00 njjnj wwyyfv
)( jj vy
![Page 21: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/21.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 21
Synapses and Neurons
jvjje
vy 1
1)(
i
n
iijnjjnj ywwwyyfv
0
00 ),...,,,...,(
yj
vj
![Page 22: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/22.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 22
Universal Approximation Theorem
Every continuous function that maps intervals of real numbers to some output interval of real numbers can be approximated arbitrarily closely by a multi-layer perceptron with just one hidden layer (with non-linear activation functions).
j
jj bwf )()( xvx j
inputsweights between input and hidden layer
biasnon-linear activation function
weights between hidden and output layer
output
![Page 23: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/23.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 23
Training an MLPTraining Event:
Network:
x0
x1
x2
x3
y
dxxxxf ),,,( 3210
yxxxxg ),,,( 3210
ydeerror
![Page 24: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/24.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 24
Training an MLPAdjust weights to minimize error (or an estimator that is some function of the error)
}_{
2 )(2
1)(
neuronsoutputjj nenε
)()()( nyndne jjj
N
navg n
N 1
)(1
![Page 25: Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski jachowski@stanford.edu](https://reader036.vdocuments.us/reader036/viewer/2022062304/56649eb35503460f94bba515/html5/thumbnails/25.jpg)
Michigan REU Final Presentations, August 10, 2006 Matt Jachowski 25
Back-Propagation AlgorithmMake correction in direction of steepest descent
)()()1( nwnwnw ijijij
)(
)()(
nw
nnw
ij
ijij
Corrections made to output layer first, propagated backwards