ViewVC Help
View File | Revision Log | Show Annotations | Root Listing
root/cvsroot/UserCode/benhoob/cmsnotes/OSPAS2011/datadriven.tex
Revision: 1.10
Committed: Mon Jun 20 13:25:18 2011 UTC (13 years, 10 months ago) by benhoob
Content type: application/x-tex
Branch: MAIN
CVS Tags: HEAD
Changes since 1.9: +5 -2 lines
Log Message:
Minor updates

File Contents

# User Rev Content
1 benhoob 1.5 \section{Counting Experiments}
2 benhoob 1.1 \label{sec:datadriven}
3 benhoob 1.3
4 benhoob 1.8 To look for possible BSM contributions, we define 2 signal regions that reject all but
5 benhoob 1.3 0.1\% of the dilepton $t\bar{t}$ events, by adding requirements of large \MET\ and \Ht:
6    
7     \begin{itemize}
8 benhoob 1.4 \item high \MET\ signal region: \MET\ $>$ 275~GeV, \Ht\ $>$ 300~GeV,
9     \item high \Ht\ signal region: \MET\ $>$ 200~GeV, \Ht\ $>$ 600~GeV.
10 benhoob 1.3 \end{itemize}
11    
12     For the high \MET\ (high \Ht) signal region, the MC predicts 2.6 (2.5) SM events,
13     dominated by dilepton $t\bar{t}$; the expected LM1 yield is 17 (14) and the
14 benhoob 1.4 expected LM3 yield is 6.4 (6.7). The signal regions are indicated in Fig.~\ref{fig:met_ht}.
15 benhoob 1.5 These signal regions are tighter than the one used in our published 2010 analysis since
16 benhoob 1.8 with the larger data sample they allow us to explore phase space farther from the core
17     of the SM distributions.
18    
19 benhoob 1.3
20 benhoob 1.5 We perform counting experiments in these signal regions, and use three independent methods to estimate from data the background in the signal region.
21 benhoob 1.8 The first method is a novel technique which is a variation of the ABCD method, which we used in our 2010 analysis~\cite{ref:ospaper},
22 benhoob 1.4 and exploits the fact that \HT\ and $y \equiv \MET/\sqrt{H_T}$ are nearly uncorrelated for the $t\bar{t}$ background;
23 benhoob 1.1 this method is referred to as the ABCD' technique. First, we extract the $y$ and \Ht\ distributions
24     $f(y)$ and $g(H_T)$ from data, using events from control regions which are dominated by background.
25 benhoob 1.3 Because $y$ and \Ht\ are weakly-correlated, the distribution of events in the $y$ vs. \Ht\ plane is described by:
26 benhoob 1.1
27     \begin{equation}
28 benhoob 1.5 \label{eq:abcdprime}
29 benhoob 1.1 \frac{\partial^2 N}{\partial y \partial H_T} = f(y)g(H_T),
30     \end{equation}
31    
32     allowing us to deduce the number of events falling in any region of this plane. In particular,
33     we can deduce the number of events falling in our signal regions defined by requirements on \MET\ and \Ht.
34    
35 benhoob 1.5 We measure the $f(y)$ and $g(H_T)$ distributions using events in the regions indicated in Fig.~\ref{fig:abcdprimedata},
36     and predict the background yields in the signal regions using Eq.~\ref{eq:abcdprime}.
37     %Next, we randomly sample values of $y$ and \Ht\ from these distributions; each pair of $y$ and \Ht\ values is a pseudo-event.
38     %We generate a large ensemble of pseudo-events, and find the ratio $R_{S/C}$, the ratio of the
39     %number of pseudo-events falling in the signal region to the number of pseudo-events
40     %falling in a control region defined by the same requirements used to select events
41     %to measure $f(y)$ and $g(H_T)$. We then
42     %multiply this ratio by the number events which fall in the control region in data
43     %to get the predicted yield, ie. $N_{pred} = R_{S/C} \times N({\rm control})$.
44 benhoob 1.8 To estimate the statistical uncertainty in the predicted background, the bin contents
45     of $f(y)$ and $g(H_T)$ are smeared according to their Poisson uncertainties.
46     We have studied this technique using toy MC studies based on
47 benhoob 1.3 event samples of similar size to the expected yield in data for 1 fb$^{-1}$.
48 benhoob 1.4 Based on these studies we correct the predicted background yields by factors of 1.2 $\pm$ 0.5
49 benhoob 1.1 (1.0 $\pm$ 0.5) for the high \MET\ (high \Ht) signal region.
50    
51    
52     The second background estimate, henceforth referred to as the dilepton transverse momentum ($\pt(\ell\ell)$) method,
53     is based on the idea~\cite{ref:victory} that in dilepton $t\bar{t}$ events the
54     \pt\ distributions of the charged leptons and neutrinos from $W$
55     decays are related, because of the common boosts from the top and $W$
56     decays. This relation is governed by the polarization of the $W$'s,
57     which is well understood in top
58     decays in the SM~\cite{Wpolarization,Wpolarization2} and can therefore be
59     reliably accounted for. We then use the observed
60     $\pt(\ell\ell)$ distribution to model the $\pt(\nu\nu)$ distribution,
61     which is identified with \MET. Thus, we use the number of observed
62 benhoob 1.3 events with $\HT > 300\GeV$ and $\pt(\ell\ell) > 275\GeV$
63 benhoob 1.5 ($\HT > 600\GeV$ and $\pt(\ell\ell) > 200\GeV$ )
64 benhoob 1.1 to predict the number of background events with
65 benhoob 1.3 $\HT > 300\GeV$ and $\MET > 275\GeV$ ($\HT > 600\GeV$ and $\MET > 200\GeV$).
66 benhoob 1.5 In practice, we apply two corrections to this prediction, following the same procedure as in Ref.~\cite{ref:ospaper}.
67 benhoob 1.10 The first correction accounts for the fact that we require \met\ $>$ 50\GeV in the preselection
68     but there is no corresponding requirement on \ptll; this correction
69     is $K_{50}=1.5 \pm 0.3$ ($1.3 \pm 0.2$) for the high \MET\ (high \Ht) signal region.
70     The second correction factor accounts for the $W$ polarization in \ttbar\ events, as well
71     as detector effects such as hadronic energy scale; this correction is $K_C = 1.5 \pm 0.5$ ($1.3 \pm 0.4$) for the
72 benhoob 1.5 high \MET (high \Ht) signal region.
73 benhoob 1.1
74     Our third background estimation method is based on the fact that many models of new physics
75 benhoob 1.5 produce an excess of SF with respect to OF lepton pairs, while for the \ttbar\ background the
76 benhoob 1.9 rates of SF and OF lepton pairs are the same, as discussed in Sec.~\ref{sec:fit}.
77 benhoob 1.7 Here we perform a counting experiment, by quantifying the excess of SF vs. OF pairs using the
78 benhoob 1.1 quantity
79    
80     \begin{equation}
81     \label{eq:ofhighpt}
82 benhoob 1.5 \Delta = R_{\mu e}N(ee) + \frac{1}{R_{\mu e}}N(\mu\mu) - N(e\mu).
83 benhoob 1.1 \end{equation}
84    
85 benhoob 1.6 This quantity is predicted to be 0 for processes with
86 benhoob 1.1 uncorrelated lepton flavors. In order for this technique to work, the kinematic selection
87     applied to events in all dilepton flavor channels must be the same, which is not the case
88 benhoob 1.3 for our default selection because the $Z$ mass veto is applied only to same-flavor channels.
89 benhoob 1.4 Therefore when applying the OF subtraction technique we also apply the $Z$ mass veto
90 benhoob 1.1 to the $e\mu$ channel.
91    
92     All background estimation methods based on data are in principle subject to signal contamination
93     in the control regions, which tends to decrease the significance of a signal
94     which may be present in the data by increasing the background prediction.
95     In general, it is difficult to quantify these effects because we
96 benhoob 1.5 do not know what signal may be present in the data. Having three
97 benhoob 1.1 independent methods (in addition to expectations from MC)
98     adds redundancy because signal contamination can have different effects
99 benhoob 1.6 in the different control regions for the three methods.
100 benhoob 1.1 For example, in the extreme case of a
101     BSM signal with identical distributions of $\pt(\ell \ell)$ and \MET, an excess of events might be seen
102 benhoob 1.3 in the ABCD' method but not in the $\pt(\ell \ell)$ method.
103 benhoob 1.1