ViewVC Help
View File | Revision Log | Show Annotations | Root Listing
root/cvsroot/UserCode/claudioc/OSNote2010/limit.tex
(Generate patch)

Comparing UserCode/claudioc/OSNote2010/limit.tex (file contents):
Revision 1.1 by claudioc, Fri Oct 29 02:29:40 2010 UTC vs.
Revision 1.32 by claudioc, Thu Jan 13 05:53:16 2011 UTC

# Line 1 | Line 1
1   \section{Limit on new physics}
2   \label{sec:limit}
3 < Nothing yet.
3 >
4 > %{\bf \color{red} The numbers in this Section need to be double checked.}
5 >
6 > \subsection{Limit on number of events}
7 > \label{sec:limnumevents}
8 > As discussed in Section~\ref{sec:results}, we see one event
9 > in the signal region, defined as SumJetPt$>$300 GeV and
10 > \met/$\sqrt{\rm SumJetPt}>8.5$ GeV$^{\frac{1}{2}}$.
11 >
12 > The background prediction from the SM Monte Carlo is 1.3 events.
13 > %, where the uncertainty comes from
14 > %the jet energy scale (30\%, see Section~\ref{sec:systematics}),
15 > %the luminosity (10\%), and the lepton/trigger
16 > %efficiency (10\%)\footnote{Other uncertainties associated with
17 > %the modeling of $t\bar{t}$ in MadGraph have not been evaluated.
18 > %The uncertainty on $pp \to \sigma(t\bar{t})$ is also not included.}.
19 > The data driven background predictions from the ABCD method
20 > and the $P_T(\ell\ell)$ method are $1.3 \pm 0.8({\rm stat}) \pm 0.3({\rm syst})$
21 > and $2.1 \pm 2.1({\rm stat}) \pm 0.6({\rm syst})$, respectively.
22 >
23 > These three predictions are in good agreement with each other
24 > and with the observation of one event in the signal region.
25 > We calculate a Bayesian 95\% CL upper limit\cite{ref:bayes.f}
26 > on the number of non SM events in the signal region to be 4.1.
27 > We have also calculated this limit using
28 > % a profile likelihood method
29 > % as implemented in
30 > the cl95cms software\cite{ref:cl95cms},
31 > and we also find 4.1.  (This is not surprising, since cl95cms
32 > also gives baysean upper limits with a flat prior).
33 > These limits were calculated using a background prediction of $N_{BG} = 1.4 \pm 0.8$
34 > events, the error-weighted average of the ABCD and $P_T(\ell\ell)$ background
35 > predictions.  The upper limit is not very sensitive to the choice of
36 > $N_{BG}$ and its uncertainty.
37 >
38 > To get a feeling for the sensitivity of this search to some
39 > popular SUSY models, we remind the reader of the number of expected
40 > LM0 and LM1 events from Table~\ref{tab:sigcont}: $8.6 \pm 1.6$
41 > events and $3.6 \pm 0.5$ events respectively, where the uncertainties
42 > are from energy scale (Section~\ref{sec:systematics}), luminosity,
43 > and lepton efficiency.
44 >
45 >
46 > \subsection{Outreach}
47 > \label{sec:outreach}
48 > Conveying additional useful information about the results of
49 > a generic ``signature-based'' search such as the one described
50 > in this note is a difficult issue.  
51 > Here we attempt to present our result in the most general
52 > way.
53 >
54 > Models of new physics in the dilepton final state
55 > can be confronted in an approximate way by simple
56 > generator-level studies that
57 > compare the expected number of events in 34.0~pb$^{-1}$
58 > with our upper limit of 4.1 events.  The key ingredients
59 > of such studies are the kinematical cuts described
60 > in this note, the lepton efficiencies, and the detector
61 > responses for SumJetPt and \met/$\sqrt{\rm SumJetPt}$.
62 > The muon identification efficiency is $\approx 95\%$;
63 > the electron identification efficiency varies from $\approx$ 63\% at
64 > $P_T = 10$ GeV to 91\% for $P_T > 30$ GeV.  The isolation
65 > efficiency in top events varies from $\approx 83\%$ (muons)
66 > and $\approx 89\%$ (electrons) at $P_T=10$ GeV to
67 > $\approx 95\%$ for $P_T>60$ GeV.
68 > %{\bf \color{red} The following numbers were derived from Fall 10 samples. }
69 > The average detector
70 > responses for SumJetPt and $\met/\sqrt{\rm SumJetPt}$ are
71 > $1.02 \pm 0.05$ and $0.94 \pm 0.05$ respectively, where
72 > the uncertainties are from the jet energy scale uncertainty.
73 > The experimental resolutions on these quantities are 11\% and
74 > 16\% respectively.
75 >
76 > To justify the statements in the previous paragraph
77 > about the detector responses, we plot
78 > in Figure~\ref{fig:response} the average response for
79 > SumJetPt and \met/$\sqrt{\rm SumJetPt}$ in MC, as well as the
80 > efficiency for the cuts on these quantities used in defining the
81 > signal region.
82 > % (SumJetPt $>$ 300 GeV and \met/$\sqrt{\rm SumJetPt} > 8.5$
83 > % Gev$^{\frac{1}{2}}$).  
84 > %{\bf \color{red} The following numbers were derived from Fall10 samples }
85 > We find that the average SumJetPt response
86 > in the Monte Carlo is about 1.02, with an RMS of order 11\% while
87 > the response of \met/$\sqrt{\rm SumJetPt}$ is approximately 0.94 with an
88 > RMS of 16\%.
89 >
90 > %Using this information as well as the kinematical
91 > %cuts described in Section~\ref{sec:eventSel} and the lepton efficiencies
92 > %of Figures~\ref{fig:effttbar}, one should be able to confront
93 > %any existing or future model via a relatively simple generator
94 > %level study by comparing the expected number of events in 35 pb$^{-1}$
95 > %with our upper limit of 4.1 events.
96 >
97 > \begin{figure}[tbh]
98 > \begin{center}
99 > \includegraphics[width=\linewidth]{selectionEffDec10.png}
100 > \caption{\label{fig:response} Left plots: the efficiencies
101 > as a function of the true quantities for the SumJetPt (top) and
102 > tcMET/$\sqrt{\rm SumJetPt}$ (bottom) requirements for the signal
103 > region as a function of their true values.  The value of the
104 > cuts is indicated by the vertical line.
105 > Right plots: The average response and its RMS for the SumJetPt
106 > (top) and tcMET/$\sqrt{\rm SumJetPt}$ (bottom) measurements.
107 > The response is defined as the ratio of the reconstructed quantity
108 > to the true quantity in MC.  These plots are done using the LM0
109 > Monte Carlo, but they are not expected to depend strongly on
110 > the underlying physics.
111 > %{\bf \color{red} These plots were made with Fall10 samples. }
112 > }
113 > \end{center}
114 > \end{figure}
115 >
116 >
117 >
118 > %%%  Nominal
119 > % -----------------------------------------
120 > % observed events                         1
121 > % relative error on acceptance        0.000
122 > % expected background                 1.400
123 > % absolute error on background        0.770
124 > % desired confidence level             0.95
125 > % integration upper limit             30.00
126 > % integration step size              0.0100
127 > % -----------------------------------------
128 > % Are the above correct? y
129 > %    1  16.685     0.29375E-06
130 > %
131 > % limit: less than     4.112 signal events
132 >
133 >
134 >
135 > %%%  Add 20% acceptance uncertainty based on LM0
136 > % -----------------------------------------
137 > % observed events                         1
138 > % relative error on acceptance        0.200
139 > % expected background                 1.400
140 > % absolute error on background        0.770
141 > % desired confidence level             0.95
142 > % integration upper limit             30.00
143 > % integration step size              0.0100
144 > % -----------------------------------------
145 > % Are the above correct? y
146 > %    1  29.995     0.50457E-06
147 > %
148 > % limit: less than     4.689 signal events
149 >
150 >
151 > \subsection{mSUGRA scan}
152 > \label{sec:mSUGRA}
153 > We also perform a scan of the mSUGRA parameter space, as recomended
154 > by the SUSY group convenors\cite{ref:scan}.
155 > The goal of the scan is to determine an exclusion region in the
156 > $m_0$ vs. $m_{1/2}$ plane for
157 > $\tan\beta=3$,
158 > sign of $\mu = +$, and $A_{0}=0$~GeV.  This scan is based on events
159 > generated with FastSim.
160 >
161 > The first order of business is to verify that results using
162 > Fastsim and Fullsim are compatible.  To this end we compare the
163 > expected yield for the LM1 point in FullSim (3.56 $\pm$ 0.06) and
164 > FastSim (3.29 $\pm$ 0.27), where the uncertainties are statistical only.
165 > These two numbers are in agreement, which gives us confidence in
166 > usinf FastSim for this study.
167 >
168 > The FastSim events are generated with different values of $m_0$
169 > and $m_{1/2}$ in steps of 10 GeV.  For each point in the
170 > $m_0$ vs. $m_{1/2}$ plane, we compute the expected number of
171 > events at NLO.  We then also calculate an upper limit $N_{UL}$
172 > using cl95cms at each point using the following inputs:
173 > \begin{itemize}
174 > \item Number of BG events = 1.40 $\pm$ 0.77
175 > \item Luminosity uncertainty = 11\%
176 > \item The acceptance uncertainty is calculated at each point
177 > as the quadrature sum of
178 > \begin{itemize}
179 > \item The uncertainty due to JES for that point, as calculated
180 > using the method described in Section~\ref{sec:systematics}
181 > \item A 5\% uncertainty due to lepton efficiencies
182 > \item An uncertaity on the NLO cross-section obtained by varying the
183 > factorization and renormalization scale by a factor of two\cite{ref:sanjay}.
184 > \item The PDF uncertainty on the product of cross-section and acceptance
185 > calculated using the method of Reference~\cite{ref:pdf}.
186 > \end{itemize}
187 > \item We use the ``log-normal'' model for the nuisance parameters
188 > in cl95cms
189 > \end{itemize}
190 >
191 > An mSUGRA point is excluded if the resulting $N_{UL}$ is smaller
192 > than the expected number of events.  Because of the quantization
193 > of the available MC points in the $m_0$ vs $m_{1/2}$ plane, the
194 > boundaries of the excluded region are also quantized.  We smooth
195 > the boundaries using the method recommended by the SUSY
196 > group\cite{ref:smooth}.  In addition, we show a limit
197 > curve based on the LO cross-section, as well as the
198 > ``expected'' limit curve.  The expected limit curve is
199 > calculated using the CLA function also available in cl95cms.
200 > The results are shown in Figure~\ref{fig:msugra}
201 >
202 >
203 > \begin{figure}[tbh]
204 > \begin{center}
205 > \includegraphics[width=\linewidth]{exclusion.pdf}
206 > \caption{\label{fig:msugra}\protect Exclusion curves in the mSUGRA parameter space,
207 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs.  THIS IS STILL MISSING
208 > THE PDF UNCERTAINTIES.}
209 > \end{center}
210 > \end{figure}
211 >
212 >
213 > \subsubsection{Check of the nuisance parameter models}
214 > We repeat the procedure outlined above but changing the
215 > lognormal nuisance parameter model to a gaussian or
216 > gamma-function model.  The results are shown in
217 > Figure~\ref{fig:nuisance}.  (In this case,
218 > to avoid smoothing artifacts, we
219 > show the raw results, without smoothing).
220 >
221 > \begin{figure}[tbh]
222 > \begin{center}
223 > \includegraphics[width=0.5\linewidth]{nuissance.png}
224 > \caption{\label{fig:nuisance}\protect Exclusion curves in the
225 > mSUGRA parameter space,
226 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs
227 > using different models for the nuisance parameters.
228 > Red:gaussian.  Blue:lognormal or gamma.  
229 > THIS IS STILL MISSING
230 > THE PDF UNCERETAINTIES.  MAYBE GOOD TO MAKE THIS PLOT A BIT
231 > PRETTIER, EG, CANNOT DISTINGUISH THINGS WHEN USING BLACK AND
232 > WHITE PRINTER}
233 > \end{center}
234 > \end{figure}
235 >
236 > We find that the set of excluded points is identical for the
237 > lognormal and gamma models.  There are small differences for the
238 > gaussian model. Following the recommendation of Reference~\cite{ref:cousins},
239 > we use the lognormal nuisance parameter model.
240 >
241 >
242 > \clearpage
243 >
244 >
245 > \subsubsection{Effect of signal contamination}
246 > \label{sec:contlimit}
247 > Signal contamination could affect the limit by inflating the
248 > background expectation.  In our case we see no evidence of signal
249 > contamination, within statistics.
250 > The yields in the control regions  
251 > $A$, $B$, and $C$ (Table~\ref{tab:datayield}) are just
252 > as expected in the SM, and the check
253 > of the $P_T(\ell \ell)$ method in the control region is
254 > also consistent with expectations (Table~\ref{tab:victory}).
255 > Since we have two data driven methods, with different
256 > signal contamination issues, giving consistent
257 > results that are in agreement with the SM, we
258 > argue for not making any correction to our procedure
259 > because of signal contamination.  In some sense this would
260 > be equivalent to using the SM background prediction, and using
261 > the data driven methods as confirmations of that prediction.
262 >
263 > Nevertheless, here we explore the possible effect of
264 > signal contamination.  The procedure suggested to us
265 > is the following:
266 > \begin{itemize}
267 > \item At each point in mSUGRA space we modify the
268 > ABCD background prediction from $A_D \cdot C_D/B_D$ to
269 > $(A_D-A_S) \cdot (C_D-C_S) / (B_D - B_S)$, where the
270 > subscripts $D$ and $S$ referes to the number of observed data
271 > events and expected SUSY events, respectively, in a given region.
272 > \item Similarly, at each point in mSugra space we modify the
273 > $P_T(\ell\ell)$ background prediction from
274 > $K \cdot K_C \cdot D'_D$ to $K \cdot K_C \cdot (D'_D - D'_S)$,
275 > where the subscript $D$ and $S$ are defined as above.
276 > \item At each point in mSUGRA space we recalculate $N_{UL}$
277 > using the weighted average of the modified $ABCD$ and
278 > $P_T(\ell\ell)$ method predictions.  
279 > \end{itemize}
280 >
281 > \noindent This procedure results in a reduced background prediction,
282 > and therefore a less stringent $N_{UL}$.  
283 >
284 > Note, however, that in some cases this procedure is
285 > nonsensical.  For example, take LM0 as a SUSY
286 > point.  In region $C$ we have a SM prediction of 5.1
287 > events and $C_D = 4$ in agreement with the Standard Model,
288 > see Table~\ref{tab:datayield}.  From the LM0 Monte Carlo,
289 > we find $C_S = 8.6$ events.   Thus, including information
290 > on $C_D$ and $C_S$ should {\bf strengthen} the limit, since there
291 > is clearly a deficit of events in the $C$ region in the
292 > LM0 hypothesis.  Instead, we now get a negative ABCD
293 > BG prediction (which is nonsense, so we set it to zero),
294 > and therefore a weaker limit.
295 >
296 > \begin{figure}[tbh]
297 > \begin{center}
298 > \includegraphics[width=0.5\linewidth]{sigcont.png}
299 > \caption{\label{fig:sigcont}\protect Exclusion curves in the
300 > mSUGRA parameter space,
301 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs
302 > with (blue) and without (red) the effects of signal contamination.
303 > THIS IS STILL MISSING
304 > THE PDF UNCERETAINTIES.  MAYBE GOOD TO MAKE THIS PLOT A BIT
305 > PRETTIER, EG, CANNOT DISTINGUISH THINGS WHEN USING BLACK AND
306 > WHITE PRINTER}
307 > \end{center}
308 > \end{figure}
309 >
310 > Despite these reservations, we follow the procedure suggested
311 > to us.  A comparison of the exclusion region with and without
312 > the signal contamination is shown in Figure~\ref{fig:sigcont}
313 > (with no smoothing).  The effect of signal contamination is
314 > small.
315 >
316 > \subsubsection{mSUGRA scans with different values of tan$\beta$}
317 > \label{sec:tanbetascan}
318 >
319 > For completeness, we also show the exclusion regions calculated
320 > using $\tan\beta = 10$ and $\tan\beta = 50$.
321 >
322 > NOT DONE YET.  HERE I SUGGEST THAT WE PUT 3 CURVES (NLO LIMITS,
323 > NO SIGNAL CONTAMINATION) ON THE SAME M0-M1/2 PLOT PERHAPS
324 > LEAVING OUT THE REGIONS EXCLUDED BY OTHER EXPERIMENTS.
325 >
326 >
327 >
328 >
329 >

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines