1 helge voss genvea 28 th march 2007root 2007: tmva toolkit for multivariate analysis tmva a toolkit...

34
1 Helge Voss Genvea 28 th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker (ATLAS), Helge Voss (LHCb), Fredrik Tegenfeld (ATLAS), Kai Voss (ex. ATLAS), Joerg Stelzer (ATLAS) http:// tmva.sourceforge.net/ supply an environment to easily: apply a (large) variety of sophisticated data selection algorithms have them all trained and tested choose the best one for your selection problem

Upload: gerald-norris

Post on 16-Jan-2016

220 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

1Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA A Toolkit for MultiVariate Data Analysis

with ROOT

Andreas Höcker (ATLAS), Helge Voss (LHCb), Fredrik Tegenfeld (ATLAS),

Kai Voss (ex. ATLAS), Joerg Stelzer (ATLAS)

http://tmva.sourceforge.net/

supply an environment to easily:

apply a (large) variety of sophisticated data selection algorithms

have them all trained and tested

choose the best one for your selection problem

Page 2: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

2Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Motivation/Outline

Outline:

Introduction: MVAs, what / where / why

the MVA methods available in TMVA

demonstration with toy examples

Summary/Outlook

Idea: Rather than just implementing new MVA techniques and making them available in ROOT in the way TMulitLayerPerceton does:

have one common interface to different MVA methods

easy to use

easy to compare many different MVA methods

train/test on same data sample

have one common place for possible pre-processing (decorrelation of variables) available for all MVA selection algorithms

Page 3: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

3Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Introduction to MVA

At the beginning of each physics analysis:

select your event sample, discriminating against background

b-tagging

Or even earlier:

e.g. particle identification, pattern recognition (ring finding) in RICH detectors

trigger applications

discriminate “tau-jets” from quark-jets

MVA -- MulitVariate Analysis:

nice name, means nothing else but:

Use several observables from your events to form

ONE combined variable and use this in order to

discriminate between “signal” and “background”

Always one uses several variables in some sort of combination:

Page 4: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

4Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Introduction to MVAssequence of cuts multivariate methods

sequence of cuts is easy to understand offers easy interpretation

sequences of cuts are often inefficient!! e.g. events might be rejected because of just ONE variable, while the others look very “signal like”

MVA: several observables One selection criterium

e.g: Likelihood selection

calculate for each observable in an event a probability that its value belongs to a “signal” or a “background” event using reference distributions (PDFs) for signal and background.

Then cut on the combination of all these probabilities

E.g. Variable Et“expected” signal and

background distributions

Et

Et (event1):

background “like”

Et (event2):

signal “like”

Page 5: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

5Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Introduction: Event Classification

A linear boundary? A nonlinear one?Rectangular cuts?

H1

H0

x1

x2 H1

H0

x1

x2 H1

H0

x1

x2

How to exploit the information present in the discriminating variables:

Often, a lot of information is also given in by the correlation of the variables

Different techniques use different ways trying to exploit (all) features

compare and choose

How to make a selection let the machine learn (training)

Page 6: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

6Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

What is TMVA

TMVA presently includes:

Rectangular cut optimisation Projective and Multi-dimensional likelihood estimator Fisher discriminant and H-Matrix (2 estimator)Artificial Neural Network (3 different implementations) Boosted/bagged Decision Trees Rule Fitting, upcomming: Support Vector Machines, “Committee” methodscommon pre-processing of input data: de-correlation, principal component analysis

Toolkit for Multivariate Analysis (TMVA) with ROOT

‘parallel’ processing of various MVA techniques to discriminate signal from background samples.

easy comparison of different MVA techniques – choose the best one

TMVA package provides training, testing and evaluation of the MVAs

Each MVA method provides a ranking of the input variables

MVAs produce weight files that are read by reader class for MVA application

Page 7: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

7Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

MVAs certainly made it’s way through HEP but “simple “cuts” are also still widely used

MVA Experience

MVAs are tedious to implement. Ready made tools often are just for one method only

few true comparisons between different methods are made

Ease of use will hopefully also help to remove remaining “black box” mystic once one gains more experience in how the methods behave

black boxes ! how to interpret the selection ?

what if the training samples incorrectly describe the data ?

how can one evaluate systematics ?

Page 8: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

8Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA Methods

Page 9: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

9Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Commonly realised for all methods in TMVA (centrally in DataSet class):

Removal of linear correlations by rotating variables

Determine square-root C of correlation matrix C, i.e., C = C C compute C by diagonalising C:

transformation from original (x) in de-correlated variable space (x) by: x = C 1x

Note that this “de-correlation” is only complete, if: input variables are Gaussians

correlations linear only

in practise: gain form de-correlation often rather modest – or even harmful

T TD S SSSC C D

Preprocessing the Input Variables: Decorrelation

Various ways to choose diagonalisation (also implemented: principal component analysis)

originaloriginal SQRT derorr.SQRT derorr. PCA derorr.PCA derorr.

Page 10: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

10Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Simplest method: cut in rectangular volume using Nvar input variables

event eventvariabl

cut, , ,min ,mas

xe

(0,1) ,v vv

ivix x x x

Usually training files in TMVA do not contain realistic signal and background abundance cannot optimize for best significance (S/(S+B) )

scan in signal efficiency [0 1] and maximise background rejection

Technical problem: how to perform optimisation random sampling: robust (if not too many observables used) but suboptimal

new techniques Genetics Algorithm and Simulated Annealing

Huge speed improvement by sorting training events in Binary Search Tree (for 4 variables we gained a factor 41)

Cut Optimisation

do this in normal variable space or de-correlated variable space

Page 11: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

11Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

automatic,unbiased, but suboptimal

easy to automate, can create artefactsTMVA uses: Splines0-5, Kernel estimators

difficult to automate

Combine probability for an event to be signal / background from individual variables to

Assumes uncorrelated input variables

in that case it is the optimal MVA approach, since it contains all the information

usually it is not true development of different methods

Technical problem: how to implement reference PDFs

3 ways: function fitting , counting , parametric fitting (splines, kernel est.)

event

event

event

variables,

signal

speci

P

variabe le

DE,

,ss

( )

( )

v vv

vS

i

iv

i

vS

p x

x

p x

discriminating variables

Species: signal, background types

Likelihood ratio for event ievent

PDFs

Projected Likelihood Estimator (PDE Appr.)

Page 12: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

12Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Carli-Koblitz, NIM A501, 576 (2003)

Generalisation of 1D PDE approach to Nvar dimensions

Optimal method – in theory – if “true N-dim PDF” were known

Practical challenges: derive N-dim PDF from training sample

TMVA implementation:

count number of signal and background events in “vicinity” of a data event fixed size or adaptive

Multidimensional Likelihood Estimator

H1

H0

x1

x2

test event

speed up range search by sorting training events in Binary Trees

use multi-D kernels (Gaussian, triangular, …) to weight events within a volume

volumes can be rectangular or spherical

Page 13: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

13Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Well-known, simple and elegant MVA method:

determine linear boundary between signal and background in transformed variable space where:

linear correlations are removed

mean values of signal and background are “pushed” as far apart as possible

optimal for linearly correlated Gaussians with equal RMS’ and different means

no separation if equal means and different RMS (shapes)

Computation of the trained Fisher MVA couldn’t be simpler:

event eventvariab

Fisheres

,l

, i v vv

ix x F

“Fisher coefficients”

Fisher Discriminant (and H-Matrix)

Page 14: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

14Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Fee

d-fo

rwar

d M

ultil

ayer

Per

cept

ron

1( ) 1 xA x e

1

i

. . .N

1 input layer k hidden layers 1 ouput layer

1

j

M1

. . .

. . . 1

. . .Mk

2 output classes (signal and background)

Nvar discriminating input variables

11w

ijw

1jw. . .. . .

1( ) ( ) ( ) ( 1)

01

kMk k k k

j j ij ii

x w w xA

var

(0)1..i Nx

( 1)1,2kx

(“Activation” function)

with:

Get a non-linear classifier response by “activating” output nodes using non-lieaear weights (activation)

Artificial Neural Network (ANN)

Taining adjust weight of each input variable to node-activation using training events

Nodes are called “neurons” and arranged in series

Feed-Forward Multilayer Perceptrons (3 different implementations in TMVA)

Page 15: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

15Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Decision Trees

sequential application of “cuts” which splits the data into nodes, and the final nodes (leaf) classifies an event as signal or background

Training: growing a decision tree

Start with Root node

Split training sample according to cut on best variable at this node

Splitting criterion: e.g., maximum “Gini-index”: purity (1– purity)

Continue splitting until min. number of events or max. purity reached

Bottom up Pruning:

remove statistically insignificant nodes (avoid overtraining)

Classify leaf node according to majority of events, or give weight; unknown test events are classified accordinglyDecision tree before pruning

Decision tree after pruning

Decision Trees

Page 16: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

16Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Decision Trees: used since a long time in general “data-mining” applications, less known in HEP (although very similar to “simple Cuts”)

Advantages:

easy to interpret: visualisation in a 2D tree

independent of monotone variable transformation: immune against outliers

usless/weak variables are ignored

Disadvatages:

instability: small changes in training sample can give large changes in tree structure

Boosted Decision Trees

Boosted Decision Trees (1996): combining several decision trees (forest) derived from one training sample via the application of event weights into ONE mulitvariate event classifier by performing “majority vote”

e.g. AdaBoost: wrong classified training events are given a larger weight

bagging, random weights re-sampling with replacement

bagging/boosting: means of creating basis functions: final classifier is a linear combination of those

Boosted Decision Trees

Page 17: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

17Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Following RuleFit from Friedman-Popescu:

Model is a linear combination of rules; a rule is a sequence of cuts:

The problem to solve is:1. create the rule ensemble created from a set of decision trees2. fit the coefficients “Gradient directed regularization” (Friedman et al)

Fast, robust and good performance

Rule Fitting(Predictive Learing via Rule Ensembles)

RF 01 1

ˆ ˆR RM n

m m k km k

x x xy a ra b

rules (cut sequence rm=1 if all cuts satisfied, =0 otherwise)

normalised discriminating event variables

RuleFit classifier

Linear Fisher termSum of rules

Friedman-Popescu, Tech Rep, Stat. Dpt, Stanford U., 2003

Page 18: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

18Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Using TMVA in Training and Application

Can be ROOT scripts, C++ executables or python scripts (via PyROOT), or any other high-level language that interfaces with ROOT

Can be ROOT scripts, C++ executables or python scripts (via PyROOT), or any other high-level language that interfaces with ROOT

Page 19: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

19Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

A Complete Example Analysisvoid TMVAnalysis( ) { TFile* outputFile = TFile::Open( "TMVA.root", "RECREATE" );

TMVA::Factory *factory = new TMVA::Factory( "MVAnalysis", outputFile,"!V");

TFile *input = TFile::Open("tmva_example.root"); TTree *signal = (TTree*)input->Get("TreeS"); TTree *background = (TTree*)input->Get("TreeB"); factory->AddSignalTree ( signal, 1. ); factory->AddBackgroundTree( background, 1.);

factory->AddVariable("var1+var2", 'F'); factory->AddVariable("var1-var2", 'F'); factory->AddVariable("var3", 'F'); factory->AddVariable("var4", 'F');

factory->PrepareTrainingAndTestTree("", "NSigTrain=3000:NBkgTrain=3000:SplitMode=Random:!V" );

factory->BookMethod( TMVA::Types::kLikelihood, "Likelihood", "!V:!TransformOutput:Spline=2:NSmooth=5:NAvEvtPerBin=50" );

factory->BookMethod( TMVA::Types::kMLP, "MLP", "!V:NCycles=200:HiddenLayers=N+1,N:TestRate=5" );

factory->TrainAllMethods(); factory->TestAllMethods(); factory->EvaluateAllMethods(); outputFile->Close(); delete factory;}

create Factory

give training/test trees

tell which variables

select the MVA methods

train,test and evaluate

Page 20: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

20Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Example Applicationvoid TMVApplication( ) { TMVA::Reader *reader = new TMVA::Reader("!Color");

Float_t var1, var2, var3, var4; reader->AddVariable( "var1+var2", &var1 ); reader->AddVariable( "var1-var2", &var2 ); reader->AddVariable( "var3", &var3 ); reader->AddVariable( "var4", &var4 );

reader->BookMVA( "MLP method", "weights/MVAnalysis_MLP.weights.txt" );

TFile *input = TFile::Open("tmva_example.root"); TTree* theTree = (TTree*)input->Get("TreeS");

Float_t userVar1, userVar2; theTree->SetBranchAddress( "var1", &userVar1 ); theTree->SetBranchAddress( "var2", &userVar2 ); theTree->SetBranchAddress( "var3", &var3 ); theTree->SetBranchAddress( "var4", &var4 );

for (Long64_t ievt=3000; ievt<theTree->GetEntries();ievt++) { theTree->GetEntry(ievt);

var1 = userVar1 + userVar2; var2 = userVar1 - userVar2; cout << reader->EvaluateMVA( "MLP method" ) <<endl; }

delete reader;}

create Reader

tell it about the variables

selected MVA method

set tree variables

event loop

calculate the MVA output

Page 21: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

21Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

A Toy Example (idealized)

Use data set with 4 linearly correlated Gaussian distributed variables:

--------------------------------------- Rank : Variable  : Separation ---------------------------------------      1 : var3      : 3.834e+02     2 : var2       : 3.062e+02     3 : var1       : 1.097e+02     4 : var0       : 5.818e+01 ---------------------------------------

A purely academic Toy example

Page 22: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

22Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Preprocessing the Input Variables

Decorrelation of variables before the training is usful for THIS example.

Similar distributions for PCA

Note that in cases with non-Gaussian distributions and/or nonlinear correlations decorrelation may do more harm than any good

Preprocessing of input Variables

Page 23: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

23Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Validating the Classifier Training

average no. of nodes before/after pruning: 4193 / 968

Validating the classifiers

TMVA GUIProjective likelihood PDFs, MLP training, BDTs, ....

Page 24: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

24Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA output distributions for Fisher, Likelihood, BDT and MLP

The OutputEvaluation Output

Page 25: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

25Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA output distributions for Fisher, Likelihood, BDT and MLP…

The OutputEvaluation Output

For this case: Fisher discriminant provides the theoretically ‘best’ possible method

Same as de-correlated Likelihood

For this case: Fisher discriminant provides the theoretically ‘best’ possible method

Same as de-correlated Likelihood

Cuts, Decision Trees and Likelihood w/o de-correlation are inferior

Cuts, Decision Trees and Likelihood w/o de-correlation are inferior

Note: About A

ll Realistic

Use Cases are Much More Difficult T

han This One

Page 26: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

26Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Evaluation Output (taken from TMVA printout)

Evaluation results ranked by best signal efficiency and purity (area)------------------------------------------------------------------------------ MVA Signal efficiency at bkg eff. (error): | Sepa- Signifi-

Methods: @B=0.01 @B=0.10 @B=0.30 Area | ration: cance:------------------------------------------------------------------------------

Fisher : 0.268(03) 0.653(03) 0.873(02) 0.882 | 0.444 1.189MLP : 0.266(03) 0.656(03) 0.873(02) 0.882 | 0.444 1.260LikelihoodD : 0.259(03) 0.649(03) 0.871(02) 0.880 | 0.441 1.251PDERS : 0.223(03) 0.628(03) 0.861(02) 0.870 | 0.417 1.192RuleFit : 0.196(03) 0.607(03) 0.845(02) 0.859 | 0.390 1.092HMatrix : 0.058(01) 0.622(03) 0.868(02) 0.855 | 0.410 1.093BDT : 0.154(02) 0.594(04) 0.838(03) 0.852 | 0.380 1.099CutsGA : 0.109(02) 1.000(00) 0.717(03) 0.784 | 0.000 0.000Likelihood : 0.086(02) 0.387(03) 0.677(03) 0.757 | 0.199 0.682

------------------------------------------------------------------------------Testing efficiency compared to training efficiency (overtraining check)

------------------------------------------------------------------------------ MVA Signal efficiency: from test sample (from traing sample)

Methods: @B=0.01 @B=0.10 @B=0.30------------------------------------------------------------------------------

Fisher : 0.268 (0.275) 0.653 (0.658) 0.873 (0.873)MLP : 0.266 (0.278) 0.656 (0.658) 0.873 (0.873)LikelihoodD : 0.259 (0.273) 0.649 (0.657) 0.871 (0.872)PDERS : 0.223 (0.389) 0.628 (0.691) 0.861 (0.881)RuleFit : 0.196 (0.198) 0.607 (0.616) 0.845 (0.848)HMatrix : 0.058 (0.060) 0.622 (0.623) 0.868 (0.868)BDT : 0.154 (0.268) 0.594 (0.736) 0.838 (0.911)CutsGA : 0.109 (0.123) 1.000 (0.424) 0.717 (0.715)Likelihood : 0.086 (0.092) 0.387 (0.379) 0.677 (0.677)

-----------------------------------------------------------------------------

Bet

ter

clas

sifie

r

Check for over-training

Page 27: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

27Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

More Toys: Linear-, Cross-, Circular Correlations

Illustrate the behaviour of linear and nonlinear classifiers

Linear correlations(same for signal and background)

Linear correlations(opposite for signal and background)

Circular correlations(same for signal and background)

More Toys: Linear, Cross, Circular correlations

Page 28: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

28Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Final Classifier Performance

Background rejection versus signal efficiency curve:

LinearExampleCrossExampleCircularExample

Final Classifier Performance

Page 29: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

29Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA releases:

part of the ROOT package (since Development release 5.11/06)

started and still available as open source package on sourceforge home page: http://tmva.sourceforge.net/

more frequent updates than for the ROOT version. (We are still heavily developping )

current release number 3.6.1 from 9th March 2007

new developers are always welcome currently 4 main developers and 24 registered contributors on sourceforge

TMVA Technicalities

Acknowledgments: The fast development of TMVA would not have been possible without the contribution and feedback from many developers and users to whom we are indebted. We thank in particular the CERN Summer students Matt Jachowski (Stan-ford) for the implementation of TMVA's new MLP neural network, and Yair Mahalalel (Tel Aviv) for a significant improvement of PDERS. We are grateful to Doug Applegate, Kregg Arms, Ren\'e Brun and the ROOT team, Tancredi Carli, Elzbieta Richter-Was, Vincent Tisserand and Marcin Wolter for helpful conversations.

Page 30: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

30Helge Voss 19th June 2006 Lausanne Seminar TMVA Toolkit for MultiVariate Analysis

We (finally) have a Users Guide !

Available from tmva.sf.net

a d v e r t i s e m e n t

TMVA Users Guide68pp, incl. code examples

submitted to arXiv:physics

Page 31: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

31Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

TMVA is still a young project!

first release on sourceforge March 8, 2006

now also as part of the ROOT package

TMVA provides the training and evaluation tools, but the decision which method is the best is certainly depends on the use case train several methods in parallel and see what is best for YOUR analysis

also provides set of ROOT macros to visualize the results

Most methods can be improved over default by optimising the training options

Aimed for easy to use tool to give access to many different “complicated” selection algorithms

Already have a number of users, but still need more real-life experience

Concluding Remarks

Page 32: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

32Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Outlook

We will continue to improve

the selection methods already implemented

flexibility of the data interface

New Methods are under development

Support Vector Machines

Bayesian Classifiers

“Committee Method” combination of different MVA techniques

Page 33: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

33Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Weight Variables by Classifier Performance

Linear correlations(same for signal and background)

Linear correlations(opposite for signal and background)

Circular correlations(same for signal and background)

How well do the classifier resolve the various correlation patterns ?

Illustustration: Events weighted by MVA-response:

Page 34: 1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker

34Helge Voss Genvea 28th March 2007 ROOT 2007: TMVA Toolkit for MultiVariate Analysis

Stability with Respect to Irrelevant Variables

Toy example with 2 discriminating and 4 non-discriminating variables ?

use only two discriminant variables in classifiers

use only two discriminant variables in classifiersuse all discriminant variables in classifiers

use all discriminant variables in classifiers

Stability with respect to irrelevant variables