class TMVA::MethodBDT: public TMVA::MethodBase


 Analysis of Boosted Decision Trees

 Boosted decision trees have been successfully used in High Energy
 Physics analysis for example by the MiniBooNE experiment
 (Yang-Roe-Zhu, physics/0508045). In Boosted Decision Trees, the
 selection is done on a majority vote on the result of several decision
 trees, which are all derived from the same training sample by
 supplying different event weights during the training.

 Decision trees:

 successive decision nodes are used to categorize the
 events out of the sample as either signal or background. Each node
 uses only a single discriminating variable to decide if the event is
 signal-like ("goes right") or background-like ("goes left"). This
 forms a tree like structure with "baskets" at the end (leave nodes),
 and an event is classified as either signal or background according to
 whether the basket where it ends up has been classified signal or
 background during the training. Training of a decision tree is the
 process to define the "cut criteria" for each node. The training
 starts with the root node. Here one takes the full training event
 sample and selects the variable and corresponding cut value that gives
 the best separation between signal and background at this stage. Using
 this cut criterion, the sample is then divided into two subsamples, a
 signal-like (right) and a background-like (left) sample. Two new nodes
 are then created for each of the two sub-samples and they are
 constructed using the same mechanism as described for the root
 node. The devision is stopped once a certain node has reached either a
 minimum number of events, or a minimum or maximum signal purity. These
 leave nodes are then called "signal" or "background" if they contain
 more signal respective background events from the training sample.

 Boosting:

 the idea behind the boosting is, that signal events from the training
 sample, that end up in a background node (and vice versa) are given a
 larger weight than events that are in the correct leave node. This
 results in a re-weighed training event sample, with which then a new
 decision tree can be developed. The boosting can be applied several
 times (typically 100-500 times) and one ends up with a set of decision
 trees (a forest).

 Bagging:

 In this particular variant of the Boosted Decision Trees the boosting
 is not done on the basis of previous training results, but by a simple
 stochasitc re-sampling of the initial training event sample.

 Random Trees:
 similar to the "Random Forests" from Leo Breiman and Adele Cutler, it
 uses the bagging algorithm together and bases the determination of the
 best node-split during the training on a random subset of variables only
 which is individually chosen for each split.

 Analysis:

 applying an individual decision tree to a test event results in a
 classification of the event as either signal or background. For the
 boosted decision tree selection, an event is successively subjected to
 the whole set of decision trees and depending on how often it is
 classified as signal, a "likelihood" estimator is constructed for the
 event being signal or background. The value of this estimator is the
 one which is then used to select the events from an event sample, and
 the cut value on this estimator defines the efficiency and purity of
 the selection.


Function Members (Methods)

public:
virtual~MethodBDT()
voidTObject::AbstractMethod(const char* method) const
voidTObject::AbstractMethod(const char* method) const
virtual voidTMVA::MethodBase::AddClassifierToTestTree(TTree* theTestTree)
virtual voidTObject::AppendPad(Option_t* option = "")
virtual voidTObject::AppendPad(Option_t* option = "")
TDirectory*TMVA::MethodBase::BaseDir() const
Double_tBoost(vector<TMVA::Event*,allocator<TMVA::Event*> >, TMVA::DecisionTree* dt, Int_t iTree)
virtual voidTObject::Browse(TBrowser* b)
virtual voidTObject::Browse(TBrowser* b)
voidTMVA::Configurable::CheckForUnusedOptions() const
static TClass*Class()
virtual const char*TObject::ClassName() const
virtual const char*TObject::ClassName() const
virtual voidTObject::Clear(Option_t* = "")
virtual voidTObject::Clear(Option_t* = "")
virtual TObject*TObject::Clone(const char* newname = "") const
virtual TObject*TObject::Clone(const char* newname = "") const
virtual Int_tTObject::Compare(const TObject* obj) const
virtual Int_tTObject::Compare(const TObject* obj) const
TMVA::ConfigurableTMVA::Configurable::Configurable(const TString& theOption = "")
virtual voidTObject::Copy(TObject& object) const
virtual voidTObject::Copy(TObject& object) const
virtual const TMVA::Ranking*CreateRanking()
TMVA::DataSet&TMVA::MethodBase::Data() const
virtual voidDeclareOptions()
virtual voidTObject::Delete(Option_t* option = "")MENU
virtual voidTObject::Delete(Option_t* option = "")MENU
virtual Int_tTObject::DistancetoPrimitive(Int_t px, Int_t py)
virtual Int_tTObject::DistancetoPrimitive(Int_t px, Int_t py)
virtual voidTObject::Draw(Option_t* option = "")
virtual voidTObject::Draw(Option_t* option = "")
virtual voidTObject::DrawClass() constMENU
virtual voidTObject::DrawClass() constMENU
virtual TObject*TObject::DrawClone(Option_t* option = "") constMENU
virtual TObject*TObject::DrawClone(Option_t* option = "") constMENU
virtual voidTObject::Dump() constMENU
virtual voidTObject::Dump() constMENU
virtual voidTObject::Error(const char* method, const char* msgfmt) const
virtual voidTObject::Error(const char* method, const char* msgfmt) const
virtual voidTObject::Execute(const char* method, const char* params, Int_t* error = 0)
virtual voidTObject::Execute(TMethod* method, TObjArray* params, Int_t* error = 0)
virtual voidTObject::Execute(const char* method, const char* params, Int_t* error = 0)
virtual voidTObject::Execute(TMethod* method, TObjArray* params, Int_t* error = 0)
virtual voidTObject::ExecuteEvent(Int_t event, Int_t px, Int_t py)
virtual voidTObject::ExecuteEvent(Int_t event, Int_t px, Int_t py)
virtual voidTObject::Fatal(const char* method, const char* msgfmt) const
virtual voidTObject::Fatal(const char* method, const char* msgfmt) const
virtual TObject*TObject::FindObject(const char* name) const
virtual TObject*TObject::FindObject(const TObject* obj) const
virtual TObject*TObject::FindObject(const char* name) const
virtual TObject*TObject::FindObject(const TObject* obj) const
const vector<double>&GetBoostWeights() const
virtual Option_t*TObject::GetDrawOption() const
virtual Option_t*TObject::GetDrawOption() const
static Long_tTObject::GetDtorOnly()
static Long_tTObject::GetDtorOnly()
virtual Double_tTMVA::MethodBase::GetEfficiency(TString, TTree*, Double_t& err)
TMVA::Event&TMVA::MethodBase::GetEvent() const
Double_tTMVA::MethodBase::GetEventVal(Int_t ivar) const
Double_tTMVA::MethodBase::GetEventValNormalised(Int_t ivar) const
Double_tTMVA::MethodBase::GetEventWeight() const
const vector<TMVA::DecisionTree*>&GetForest() const
virtual voidGetHelpMessage() const
virtual const char*TObject::GetIconName() const
virtual const char*TObject::GetIconName() const
const TString&TMVA::MethodBase::GetInputExp(int i) const
const TString&TMVA::MethodBase::GetInputVar(int i) const
const TString&TMVA::MethodBase::GetJobName() const
virtual Double_tTMVA::MethodBase::GetMaximumSignificance(Double_t SignalEvents, Double_t BackgroundEvents, Double_t& optimal_significance_value) const
const TString&TMVA::MethodBase::GetMethodName() const
const TString&TMVA::MethodBase::GetMethodTitle() const
TMVA::Types::EMVATMVA::MethodBase::GetMethodType() const
virtual Double_tGetMvaValue()
virtual const char*TMVA::MethodBase::GetName() const
Int_tTMVA::MethodBase::GetNvar() const
virtual char*TObject::GetObjectInfo(Int_t px, Int_t py) const
virtual char*TObject::GetObjectInfo(Int_t px, Int_t py) const
static Bool_tTObject::GetObjectStat()
static Bool_tTObject::GetObjectStat()
virtual Option_t*TObject::GetOption() const
virtual Option_t*TObject::GetOption() const
const TString&TMVA::Configurable::GetOptions() const
virtual Double_tTMVA::MethodBase::GetProba(Double_t mvaVal, Double_t ap_sig)
const TStringTMVA::MethodBase::GetProbaName() const
virtual Double_tTMVA::MethodBase::GetRarity(Double_t mvaVal, TMVA::Types::ESBType reftype = Types::kBackground) const
Double_tTMVA::MethodBase::GetRMS(Int_t ivar) const
virtual Double_tTMVA::MethodBase::GetSeparation(TH1*, TH1*) const
virtual Double_tTMVA::MethodBase::GetSeparation(TMVA::PDF* pdfS = 0, TMVA::PDF* pdfB = 0) const
Double_tTMVA::MethodBase::GetSignalReferenceCut() const
virtual Double_tTMVA::MethodBase::GetSignificance() const
const TString&TMVA::MethodBase::GetTestvarName() const
virtual const char*TObject::GetTitle() const
virtual const char*TObject::GetTitle() const
virtual Double_tTMVA::MethodBase::GetTrainingEfficiency(TString)
const vector<TMVA::Event*>&GetTrainingEvents() const
UInt_tTMVA::MethodBase::GetTrainingROOTVersionCode() const
TStringTMVA::MethodBase::GetTrainingROOTVersionString() const
UInt_tTMVA::MethodBase::GetTrainingTMVAVersionCode() const
TStringTMVA::MethodBase::GetTrainingTMVAVersionString() const
virtual UInt_tTObject::GetUniqueID() const
virtual UInt_tTObject::GetUniqueID() const
vector<Double_t>GetVariableImportance()
Double_tGetVariableImportance(UInt_t ivar)
TMVA::VariableTransformBase&TMVA::MethodBase::GetVarTransform() const
Double_tTMVA::MethodBase::GetXmax(Int_t ivar) const
Double_tTMVA::MethodBase::GetXmin(Int_t ivar) const
virtual Bool_tTObject::HandleTimer(TTimer* timer)
virtual Bool_tTObject::HandleTimer(TTimer* timer)
virtual ULong_tTObject::Hash() const
virtual ULong_tTObject::Hash() const
virtual voidTObject::Info(const char* method, const char* msgfmt) const
virtual voidTObject::Info(const char* method, const char* msgfmt) const
virtual Bool_tTObject::InheritsFrom(const char* classname) const
virtual Bool_tTObject::InheritsFrom(const TClass* cl) const
virtual Bool_tTObject::InheritsFrom(const char* classname) const
virtual Bool_tTObject::InheritsFrom(const TClass* cl) const
voidInitEventSample()
virtual voidTObject::Inspect() constMENU
virtual voidTObject::Inspect() constMENU
voidTObject::InvertBit(UInt_t f)
voidTObject::InvertBit(UInt_t f)
virtual TClass*IsA() const
virtual Bool_tTObject::IsEqual(const TObject* obj) const
virtual Bool_tTObject::IsEqual(const TObject* obj) const
virtual Bool_tTObject::IsFolder() const
virtual Bool_tTObject::IsFolder() const
Bool_tTObject::IsOnHeap() const
Bool_tTObject::IsOnHeap() const
Bool_tTMVA::MethodBase::IsSignalEvent() const
virtual Bool_tTMVA::MethodBase::IsSignalLike()
virtual Bool_tTObject::IsSortable() const
virtual Bool_tTObject::IsSortable() const
Bool_tTObject::IsZombie() const
Bool_tTObject::IsZombie() const
virtual voidTObject::ls(Option_t* option = "") const
virtual voidTObject::ls(Option_t* option = "") const
virtual voidTMVA::MethodBase::MakeClass(const TString& classFileName = "") const
voidMakeClassInstantiateNode(TMVA::DecisionTreeNode* n, ostream& fout, const TString& className) const
virtual voidMakeClassSpecific(ostream&, const TString&) const
virtual voidMakeClassSpecificHeader(ostream&, const TString&) const
voidTObject::MayNotUse(const char* method) const
voidTObject::MayNotUse(const char* method) const
TDirectory*TMVA::MethodBase::MethodBaseDir() const
TMVA::MethodBDTMethodBDT(TMVA::DataSet& theData, const TString& theWeightFile, TDirectory* theTargetDir = NULL)
TMVA::MethodBDTMethodBDT(const TString& jobName, const TString& methodTitle, TMVA::DataSet& theData, const TString& theOption = "", TDirectory* theTargetDir = 0)
virtual Bool_tTObject::Notify()
virtual Bool_tTObject::Notify()
static voidTObject::operator delete(void* ptr)
static voidTObject::operator delete(void* ptr)
static voidTObject::operator delete(void* ptr, void* vp)
static voidTObject::operator delete(void* ptr, void* vp)
static voidTObject::operator delete[](void* ptr)
static voidTObject::operator delete[](void* ptr)
static voidTObject::operator delete[](void* ptr, void* vp)
static voidTObject::operator delete[](void* ptr, void* vp)
void*TObject::operator new(size_t sz)
void*TObject::operator new(size_t sz)
void*TObject::operator new(size_t sz, void* vp)
void*TObject::operator new(size_t sz, void* vp)
void*TObject::operator new[](size_t sz)
void*TObject::operator new[](size_t sz)
void*TObject::operator new[](size_t sz, void* vp)
void*TObject::operator new[](size_t sz, void* vp)
TMVA::IMethod&TMVA::IMethod::operator=(const TMVA::IMethod&)
virtual voidTObject::Paint(Option_t* option = "")
virtual voidTObject::Paint(Option_t* option = "")
voidTMVA::Configurable::ParseOptions(Bool_t verbose = kTRUE)
virtual voidTObject::Pop()
virtual voidTObject::Pop()
virtual voidTObject::Print(Option_t* option = "") const
virtual voidTObject::Print(Option_t* option = "") const
virtual voidTMVA::MethodBase::PrintHelpMessage() const
voidTMVA::Configurable::PrintOptions() const
virtual voidProcessOptions()
Double_tPruneTree(TMVA::DecisionTree* dt, Int_t itree)
virtual Int_tTObject::Read(const char* name)
virtual Int_tTObject::Read(const char* name)
Bool_tTMVA::MethodBase::ReadEvent(TTree* tr, UInt_t ievt, TMVA::Types::ESBType type = Types::kMaxSBType) const
voidTMVA::MethodBase::ReadStateFromFile()
voidTMVA::MethodBase::ReadStateFromStream(istream& tf)
voidTMVA::MethodBase::ReadStateFromStream(TFile& rf)
Bool_tTMVA::MethodBase::ReadTestEvent(UInt_t ievt, TMVA::Types::ESBType type = Types::kMaxSBType) const
Bool_tTMVA::MethodBase::ReadTrainingEvent(UInt_t ievt, TMVA::Types::ESBType type = Types::kMaxSBType) const
virtual voidReadWeightsFromStream(istream& istr)
virtual voidTObject::RecursiveRemove(TObject* obj)
virtual voidTObject::RecursiveRemove(TObject* obj)
voidTObject::ResetBit(UInt_t f)
voidTObject::ResetBit(UInt_t f)
virtual voidTObject::SaveAs(const char* filename = "", Option_t* option = "") constMENU
virtual voidTObject::SaveAs(const char* filename = "", Option_t* option = "") constMENU
virtual voidTObject::SavePrimitive(basic_ostream<char,char_traits<char> >& out, Option_t* option = "")
virtual voidTObject::SavePrimitive(basic_ostream<char,char_traits<char> >& out, Option_t* option = "")
voidTObject::SetBit(UInt_t f)
voidTObject::SetBit(UInt_t f)
voidTObject::SetBit(UInt_t f, Bool_t set)
voidTObject::SetBit(UInt_t f, Bool_t set)
virtual voidTObject::SetDrawOption(Option_t* option = "")MENU
virtual voidTObject::SetDrawOption(Option_t* option = "")MENU
static voidTObject::SetDtorOnly(void* obj)
static voidTObject::SetDtorOnly(void* obj)
voidTMVA::MethodBase::SetMethodName(TString methodName)
voidTMVA::MethodBase::SetMethodTitle(TString methodTitle)
voidTMVA::MethodBase::SetMethodType(TMVA::Types::EMVA methodType)
voidTMVA::Configurable::SetName(const char* n)
static voidTObject::SetObjectStat(Bool_t stat)
static voidTObject::SetObjectStat(Bool_t stat)
voidTMVA::Configurable::SetOptions(const TString& s)
voidTMVA::MethodBase::SetTestvarName(const TString& v = "")
voidTMVA::MethodBase::SetTestvarPrefix(TString prefix)
virtual voidTObject::SetUniqueID(UInt_t uid)
virtual voidTObject::SetUniqueID(UInt_t uid)
virtual voidShowMembers(TMemberInspector& insp, char* parent)
virtual voidStreamer(TBuffer& b)
voidStreamerNVirtual(TBuffer& b)
virtual voidTObject::SysError(const char* method, const char* msgfmt) const
virtual voidTObject::SysError(const char* method, const char* msgfmt) const
virtual voidTMVA::MethodBase::Test(TTree* theTestTree = 0)
Bool_tTObject::TestBit(UInt_t f) const
Bool_tTObject::TestBit(UInt_t f) const
Int_tTObject::TestBits(UInt_t f) const
Int_tTObject::TestBits(UInt_t f) const
Double_tTestTreeQuality(TMVA::DecisionTree* dt)
virtual voidTrain()
voidTMVA::MethodBase::TrainMethod()
virtual voidTObject::UseCurrentStyle()
virtual voidTObject::UseCurrentStyle()
virtual voidTObject::Warning(const char* method, const char* msgfmt) const
virtual voidTObject::Warning(const char* method, const char* msgfmt) const
virtual Int_tTObject::Write(const char* name = 0, Int_t option = 0, Int_t bufsize = 0)
virtual Int_tTObject::Write(const char* name = 0, Int_t option = 0, Int_t bufsize = 0) const
virtual Int_tTObject::Write(const char* name = 0, Int_t option = 0, Int_t bufsize = 0)
virtual Int_tTObject::Write(const char* name = 0, Int_t option = 0, Int_t bufsize = 0) const
virtual voidTMVA::MethodBase::WriteEvaluationHistosToFile()
virtual voidWriteMonitoringHistosToFile() const
voidTMVA::MethodBase::WriteStateToFile() const
voidTMVA::MethodBase::WriteStateToStream(TFile& rf) const
voidTMVA::MethodBase::WriteStateToStream(ostream& tf, Bool_t isClass = kFALSE) const
virtual voidWriteWeightsToStream(ostream& o) const
protected:
Bool_tTMVA::MethodBase::CheckSanity(TTree* theTree = 0)
virtual voidTObject::DoError(int level, const char* location, const char* fmt, va_list va) const
virtual voidTObject::DoError(int level, const char* location, const char* fmt, va_list va) const
voidTMVA::Configurable::EnableLooseOptions(Bool_t b = kTRUE)
const TString&TMVA::MethodBase::GetInternalVarName(Int_t ivar) const
const TString&TMVA::MethodBase::GetOriginalVarName(Int_t ivar) const
TTree*TMVA::MethodBase::GetTestTree() const
static TMVA::MethodBase*TMVA::MethodBase::GetThisBase()
TTree*TMVA::MethodBase::GetTrainingTree() const
TMVA::Types::EVariableTransformTMVA::MethodBase::GetVariableTransform() const
TStringTMVA::MethodBase::GetWeightFileDir() const
TStringTMVA::MethodBase::GetWeightFileName() const
Bool_tTMVA::MethodBase::HasTrainingTree() const
Bool_tTMVA::MethodBase::Help() const
Bool_tTMVA::MethodBase::IsNormalised() const
TDirectory*TMVA::MethodBase::LocalTDir() const
Bool_tTMVA::Configurable::LooseOptionCheckingEnabled() const
voidTObject::MakeZombie()
voidTObject::MakeZombie()
voidTMVA::Configurable::ReadOptionsFromStream(istream& istr)
voidTMVA::Configurable::ResetSetFlag()
voidTMVA::MethodBase::SetNormalised(Bool_t norm)
voidTMVA::MethodBase::SetNvar(Int_t n)
voidTMVA::MethodBase::SetSignalReferenceCut(Double_t cut)
voidTMVA::MethodBase::SetWeightFileDir(TString fileDir)
voidTMVA::MethodBase::SetWeightFileName(TString)
voidTMVA::MethodBase::Statistics(TMVA::Types::ETreeType treeType, const TString& theVarName, Double_t&, Double_t&, Double_t&, Double_t&, Double_t&, Double_t&, Bool_t norm = kFALSE)
Bool_tTMVA::MethodBase::TxtWeightsOnly() const
Bool_tTMVA::MethodBase::Verbose() const
voidTMVA::Configurable::WriteOptionsToStream(ostream& o, const TString& prefix) const
private:
Double_tAdaBoost(vector<TMVA::Event*,allocator<TMVA::Event*> >, TMVA::DecisionTree* dt)
Double_tBagging(vector<TMVA::Event*,allocator<TMVA::Event*> >, Int_t iTree)
voidInitBDT()

Data Members

public:
enum TMVA::MethodBase::EWeightFileType { kROOT
kTEXT
};
enum TMVA::MethodBase::ECutOrientation { kNegative
kPositive
};
enum TObject::EStatusBits { kCanDelete
kMustCleanup
kObjInCanvas
kIsReferenced
kHasUUID
kCannotPick
kNoContextMenu
kInvalidObject
};
enum TObject::[unnamed] { kIsOnHeap
kNotDeleted
kZombie
kBitMask
kSingleKey
kOverwrite
kWriteDelete
};
enum TObject::EStatusBits { kCanDelete
kMustCleanup
kObjInCanvas
kIsReferenced
kHasUUID
kCannotPick
kNoContextMenu
kInvalidObject
};
enum TObject::[unnamed] { kIsOnHeap
kNotDeleted
kZombie
kBitMask
kSingleKey
kOverwrite
kWriteDelete
};
public:
TMVA::MsgLoggerTMVA::Configurable::fLoggermessage logger
protected:
vector<TString>*TMVA::MethodBase::fInputVarsvector of input variables used in MVA
TMVA::MsgLoggerTMVA::MethodBase::fLoggermessage logger
Int_tTMVA::MethodBase::fNbinsnumber of bins in representative histograms
Int_tTMVA::MethodBase::fNbinsHnumber of bins in evaluation histograms
TMVA::Ranking*TMVA::MethodBase::fRankingpointer to ranking object (created by derived classifiers)
private:
Bool_tfAutomaticuse user given prune strength or automatically determined one using a validation sample
TStringfBoostTypestring specifying the boost type
Double_tfBoostWeightntuple var: boost weight
TH1F*fBoostWeightHistweights applied in boosting
TH1F*fBoostWeightVsTreeweights applied in boosting vs tree number
vector<double>fBoostWeightsthe weights applied in the individual boosts
Double_tfDeltaPruneStrengthstep size in pruning, is adjusted according to experience of previous trees
TH1F*fErrFractHisterror fraction vs tree number
Double_tfErrorFractionntuple var: misclassification error fraction
vector<TMVA::Event*,allocator<TMVA::Event*> >fEventSamplethe training events
vector<DecisionTree*>fForestthe collection of decision trees
Int_tfITreentuple var: ith tree
TTree*fMonitorNtuplemonitoring ntuple
Int_tfNCutsgrid used in cut applied in node splitting
Int_tfNTreesnumber of decision trees requested
Bool_tfNoNegWeightsInTrainingignore negative event weights in the training
Int_tfNodeMinEventsmin number of events in node
TH1I*fNodesAfterPruningVsTreenNodesAfterPruning vs tree number
TH1I*fNodesBeforePruningVsTreenNodesBeforePruning vs tree number
TMVA::DecisionTree::EPruneMethodfPruneMethodmethod used for prunig
TStringfPruneMethodSprune method option String
Double_tfPruneStrengtha parameter to set the "amount" of pruning..needs to be adjusted
Bool_tfRandomisedTreeschoose a random subset of possible cut variables at each node during training
TMVA::SeparationBase*fSepTypethe separation used in node splitting
TStringfSepTypeSthe separation (option string) used in node splitting
Int_tfUseNvarsthe number of variables used in the randomised tree splitting
Bool_tfUseWeightedTreesuse average classification from the trees, or have the individual trees trees in the forest weighted (e.g. log(boostweight) from AdaBoost
Bool_tfUseYesNoLeafuse sig or bkg classification in leave nodes or sig/bkg
vector<TMVA::Event*,allocator<TMVA::Event*> >fValidationSamplethe Validation events
vector<Double_t>fVariableImportancethe relative importance of the different variables
static const Int_tfgDebugLeveldebug level determining some printout/control plots etc.

Class Charts

Inheritance Inherited Members Includes Libraries
Class Charts

Function documentation

MethodBDT(const TString& jobName, const TString& methodTitle, TMVA::DataSet& theData, const TString& theOption = "", TDirectory* theTargetDir = 0)
 the standard constructor for the "boosted decision trees"

 MethodBDT (Boosted Decision Trees) options:
 know options:
 nTrees=Int_t:    number of trees in the forest to be created
 BoostType=       the boosting type for the trees in the forest (AdaBoost e.t.c..)
                  known: AdaBoost
                         Bagging
 SeparationType   the separation criterion applied in the node splitting
                  known: GiniIndex
                         MisClassificationError
                         CrossEntropy
                         SDivSqrtSPlusB
 nEventsMin:      the minimum number of events in a node (leaf criteria, stop splitting)
 nCuts:           the number of steps in the optimisation of the cut for a node
 UseYesNoLeaf     decide if the classification is done simply by the node type, or the S/B
                  (from the training) in the leaf node
 UseWeightedTrees use average classification from the trees, or have the individual trees
                  trees in the forest weighted (e.g. log(boostweight) from AdaBoost
 PruneMethod      The Pruning method:
                  known: NoPruning  // switch off pruning completely
                         ExpectedError
                         CostComplexity
                         CostComplexity2
 PruneStrength    a parameter to adjust the amount of pruning.
                  Should be large enouth such that overtraining is avoided");
 NoNegWeightsInTraining  Ignore negative weight events in the training.
 UseRandomisedTrees  choose at each node splitting a random set of variables
 UseNvars         use UseNvars variables in randomised trees
MethodBDT(TMVA::DataSet& theData, const TString& theWeightFile, TDirectory* theTargetDir = NULL)
 constructor for calculating BDT-MVA using previously generatad decision trees
 the result of the previous training (the decision trees) are read in via the
 weightfile. Make sure the "theVariables" correspond to the ones used in
 creating the "weight"-file
void DeclareOptions()
 define the options (their key words) that can be set in the option string
 know options:
 nTrees=Int_t:    number of trees in the forest to be created
 BoostType=       the boosting type for the trees in the forest (AdaBoost e.t.c..)
                  known: AdaBoost
                         Bagging
 SeparationType   the separation criterion applied in the node splitting
                  known: GiniIndex
                         MisClassificationError
                         CrossEntropy
                         SDivSqrtSPlusB
 nEventsMin:      the minimum number of events in a node (leaf criteria, stop splitting)
 nCuts:           the number of steps in the optimisation of the cut for a node
 UseYesNoLeaf     decide if the classification is done simply by the node type, or the S/B
                  (from the training) in the leaf node
 UseWeightedTrees use average classification from the trees, or have the individual trees
                  trees in the forest weighted (e.g. log(boostweight) from AdaBoost
 PruneMethod      The Pruning method:
                  known: NoPruning  // switch off pruning completely
                         ExpectedError
                         CostComplexity
                         CostComplexity2
 PruneStrength    a parameter to adjust the amount of pruning. Should be large enouth such that overtraining is avoided");
 NoNegWeightsInTraining  Ignore negative weight events in the training.
 UseRandomisedTrees  choose at each node splitting a random set of variables
 UseNvars         use UseNvars variables in randomised trees
void ProcessOptions()
 the option string is decoded, for available options see "DeclareOptions"
void InitBDT( void )
 common initialisation with defaults for the BDT-Method
~MethodBDT( void )
destructor
void InitEventSample( void )
 write all Events from the Tree into a vector of Events, that are
 more easily manipulated.
 This method should never be called without existing trainingTree, as it
 the vector of events from the ROOT training tree
void Train( void )
 some option, not yet set as "choosable option".. more for internal testing
Double_t PruneTree(TMVA::DecisionTree* dt, Int_t itree)
 prune a tree adjusting the prunestrength using the "test sample" until
 the best efficiency on the test sample is obtained. In principle the
 test sample should not be used for that but rather a third validation
 sample, or the trainng sample with "cross validation". The latter is
 planned but will come later.
Double_t TestTreeQuality(TMVA::DecisionTree* dt)
 test the tree quality.. in terms of Miscalssification
Double_t Boost(vector<TMVA::Event*,allocator<TMVA::Event*> > , TMVA::DecisionTree* dt, Int_t iTree)
 apply the boosting alogrithim (the algorithm is selecte via the the "option" given
 in the constructor. The return value is the boosting weight
Double_t AdaBoost(vector<TMVA::Event*,allocator<TMVA::Event*> > , TMVA::DecisionTree* dt)
 the AdaBoost implementation.
 a new training sample is generated by weighting
 events that are misclassified by the decision tree. The weight
 applied is w = (1-err)/err or more general:
            w = ((1-err)/err)^beta
 where err is the fraction of misclassified events in the tree ( <0.5 assuming
 demanding the that previous selection was better than random guessing)
 and "beta" beeing a free parameter (standard: beta = 1) that modifies the
 boosting.
Double_t Bagging(vector<TMVA::Event*,allocator<TMVA::Event*> > , Int_t iTree)
 call it Bootstrapping, re-sampling or whatever you like, in the end it is nothing
 else but applying "random Weights" to each event.
void WriteWeightsToStream(ostream& o) const
 and save the Weights
void ReadWeightsFromStream(istream& istr)
 read variable names and min/max
 NOTE: the latter values are mandatory for the normalisation
 in the reader application !!!
Double_t GetMvaValue()
 return the MVA value (range [-1;1]) that classifies the
 event.according to the majority vote from the total number of
 decision trees
 In the literature I found that people actually use the
 weighted majority vote (using the boost weights) .. However I
 did not see any improvement in doing so :(
 --> this is currently switched off
void WriteMonitoringHistosToFile( void )
 here we could write some histograms created during the processing
 to the output file.
vector< Double_t > GetVariableImportance(UInt_t ivar)
 return the relative variable importance, normalized to all
 variables together having the importance 1. The importance in
 evaluated as the total separation-gain that this variable had in
 the decision trees (weighted by the number of events)
Double_t GetVariableImportance(UInt_t ivar)
 returns the measure for the variable importance of variable "ivar"
 which is later used in GetVariableImportance() to calculat the
 relative variable importances
const TMVA::Ranking* CreateRanking()
 computes ranking of input variables
void GetHelpMessage()
 get help message text

 typical length of text line:
         "|--------------------------------------------------------------|"
void MakeClassSpecific(ostream& , const TString& ) const
 make ROOT-independent C++ class for classifier response (classifier-specific implementation)
void MakeClassSpecificHeader(ostream& , const TString& ) const
 specific class header
void MakeClassInstantiateNode(TMVA::DecisionTreeNode* n, ostream& fout, const TString& className) const
 recursively descends a tree and writes the node instance to the output streem
const std::vector<TMVA::DecisionTree*>& GetForest()
{ return fForest; }
const std::vector<TMVA::Event*>& GetTrainingEvents()
{ return fEventSample; }
const std::vector<double>& GetBoostWeights()
{ return fBoostWeights; }

Author: Andreas Hoecker, Joerg Stelzer, Helge Voss, Kai Voss
Last change: root/tmva $Id: MethodBDT.h 23334 2008-04-19 18:38:57Z brun $
Last generated: 2008-06-25 08:48
Copyright (c) 2005: *

This page has been automatically generated. If you have any comments or suggestions about the page layout send a mail to ROOT support, or contact the developers with any questions or problems regarding ROOT.