ViewVC Help
View File | Revision Log | Show Annotations | Root Listing
root/cvsroot/UserCode/Friis/TancNote/note/tanc_nn_training.tex
Revision: 1.1
Committed: Wed Mar 31 01:22:23 2010 UTC (15 years, 1 month ago) by friis
Content type: application/x-tex
Branch: MAIN
Log Message:
Slow but steady

File Contents

# User Rev Content
1 friis 1.1 The samples used to train the TaNC neural networks are typical of the signals
2     and backgrounds found in common physics analyses using taus. The signal--type
3     training sample is composed of reconstucted tau--candidates that are matched to
4     generator level hadronic tau decays coming from simulated $Z \rightarrow
5     \tau^{+}\tau^{-}$ events. The background training sample consists of reconstructed
6     tau--candidates in simulated QCD $2\rightarrow2$ hard scattering events. For
7     both signal and background samples, 20\% of the events are reserved as a
8     statistically independent sample to evaluate the performance of the neural nets
9     after the training is completed. The TaNC uses the ``MLP'' neural network
10     implementation provided by the TMVA software package, described in ~\cite{TMVA}.
11    
12     The signal and background samples are split into five subsamples corresponding
13     to each reconstructed decay mode. An additional selection is applied to each
14     subsample by requiring a ``leading pion'': either a charged hadron or gamma
15     candidate with transverse momentum greater than 5 GeV$/c$. A large number of
16     QCD training events is required as the leading pion selection and the
17     requirement that the decay mode match one of the dominant modes given in table
18     ~\ref{tab:decay_modes} are both effective discriminants. For each subsample,
19     10000 signal and background tau--candidates are reserved to be used internally
20     by the TMVA software to test for over--training. The number of signal and
21     background entries used for each decay mode subsample is given in table
22     ~\ref{tab:trainingEvents}.
23    
24     \begin{table}
25     \centering
26     \begin{tabular}{lc|c|}
27     \multirow{2}{*}{Decay Mode} & \multicolumn{2}{c}{Training Events} \\
28     & Signal & Background \\
29     \hline
30     $\pi^{-}$ & blah & blah \\
31     $\pi^{-}\pi^0$ & blah & blah \\
32     $\pi^{-}\pi^0\pi^0$ & blah & blah \\
33     $\pi^{-}\pi^{-}\pi^{+}$ & blah & blah \\
34     $\pi^{-}\pi^{-}\pi^{+}\pi^0$ & blah & blah \\
35     \hline
36     Total number of events & blah & blah \\
37     \hline
38     Training preselection efficiency & blah & blah \\
39     \end{tabular}
40     \label{tab:trainingEvents}
41     \caption{Number of events used for neural network training for each
42     selected decay mode.}
43     \end{table}
44    
45    
46     The neural nets use asi nput variables the transverse momentum and $\eta$ of the
47     tau--candidates. These variables are included as their correlations with other
48     observables can increase the separation power of the ensemble of observables.
49     For example, the opening angle in $\Delta R$ for signal tau--candidates is
50     inversely related to the transverse momentum, while for background events the
51     correlation is very small (see~\cite{DavisTau}). In the training signal and
52     background samples, there is significant discrimination power in the $P_T$
53     spectrum. However, it is desirable to eliminate any systematic dependence of
54     the neural network output on $P_T$ and $\eta$, as a physics analysis that makes use
55     of the TaNC algorithm will in general have a different $P_T$ and $\eta$
56     spectrum. \fixme(this sentance sucks) The dependence on $P_T$ and $\eta$ is
57     removed by applying a $P_T$ and $\eta$ dependent weight to the tau--candidates
58     when training the neural nets.
59    
60     The weights are defined such that weighted $P_T-\eta$ two--dimensional distributions
61     are identical for the signal and background samples after the ``leading pion''
62     and decay mode selections have been applied. The weights are defined
63     \begin{align}
64     W(P_T, \eta) &=& {\rm less}(p_{sig}(P_T, \eta), p_{bkg}(P_T, \eta))\\
65     w_{sig}(P_T, \eta) &=& W(P_T, \eta)/p_{sig}(P_T, \eta) \\
66     w_{bkg}(P_T, \eta) &=& W(P_T, \eta)/p_{bkg}(P_T, \eta)
67     \end{align}
68    
69     \fixme(workingpoint)
70