ViewVC Help
View File | Revision Log | Show Annotations | Root Listing
root/cvsroot/UserCode/claudioc/OSNote2010/limit.tex
(Generate patch)

Comparing UserCode/claudioc/OSNote2010/limit.tex (file contents):
Revision 1.2 by claudioc, Thu Nov 11 12:34:17 2010 UTC vs.
Revision 1.35 by claudioc, Fri Jan 14 00:21:24 2011 UTC

# Line 1 | Line 1
1   \section{Limit on new physics}
2   \label{sec:limit}
3  
4 < {\bf \color{red} The numbers in this Section need to be double checked.}
4 > %{\bf \color{red} The numbers in this Section need to be double checked.}
5  
6 + \subsection{Limit on number of events}
7 + \label{sec:limnumevents}
8   As discussed in Section~\ref{sec:results}, we see one event
9   in the signal region, defined as SumJetPt$>$300 GeV and
10   \met/$\sqrt{\rm SumJetPt}>8.5$ GeV$^{\frac{1}{2}}$.
11  
12 < The background prediction from the SM Monte Carlo is
13 < 1.4 $\pm$ 0.5 events, where the uncertainty comes from
14 < the jet energy scale (30\%, see Section~\ref{sec:systematics}),
15 < the luminosity (10\%), and the lepton/trigger
16 < efficiency (10\%)\footnote{Other uncertainties associated with
17 < the modeling of $t\bar{t}$ in MadGraph have not been evaluated.}.
12 > The background prediction from the SM Monte Carlo is 1.3 events.
13 > %, where the uncertainty comes from
14 > %the jet energy scale (30\%, see Section~\ref{sec:systematics}),
15 > %the luminosity (10\%), and the lepton/trigger
16 > %efficiency (10\%)\footnote{Other uncertainties associated with
17 > %the modeling of $t\bar{t}$ in MadGraph have not been evaluated.
18 > %The uncertainty on $pp \to \sigma(t\bar{t})$ is also not included.}.
19   The data driven background predictions from the ABCD method
20 < and the $P_T(\ell\ell)$ method are 1.5 $\pm$ 0.9 and
21 < $1.8^{+2.5}_{-1.8}$ events respectively.
20 > and the $P_T(\ell\ell)$ method are $1.3 \pm 0.8({\rm stat}) \pm 0.3({\rm syst})$
21 > and $2.1 \pm 2.1({\rm stat}) \pm 0.6({\rm syst})$, respectively.
22  
23   These three predictions are in good agreement with each other
24   and with the observation of one event in the signal region.
25 < We calculate a baysean 95\% CL upper limit\cite{ref:bayes.f}
25 > We calculate a Bayesian 95\% CL upper limit\cite{ref:bayes.f}
26   on the number of non SM events in the signal region to be 4.1.
27 < This was calculated using a background prediction of $N_{BG}=1.4 \pm 1.0$
28 < events.  The upper limit is not very sensitive to the choice of
27 > We have also calculated this limit using
28 > % a profile likelihood method
29 > % as implemented in
30 > the cl95cms software\cite{ref:cl95cms},
31 > and we also find 4.1.  (This is not surprising, since cl95cms
32 > also gives baysean upper limits with a flat prior).
33 > These limits were calculated using a background prediction of $N_{BG} = 1.4 \pm 0.8$
34 > events, the error-weighted average of the ABCD and $P_T(\ell\ell)$ background
35 > predictions.  The upper limit is not very sensitive to the choice of
36   $N_{BG}$ and its uncertainty.
37  
38   To get a feeling for the sensitivity of this search to some
39   popular SUSY models, we remind the reader of the number of expected
40 < LM0 and LM1 events from Table~\ref{tab:sigcontABCD}: $5.6 \pm 1.1$
41 < events and $2.2 \pm 0.3$ respectively, where the uncertainties
40 > LM0 and LM1 events from Table~\ref{tab:sigcont}: $8.6 \pm 1.6$
41 > events and $3.6 \pm 0.5$ events respectively, where the uncertainties
42   are from energy scale (Section~\ref{sec:systematics}), luminosity,
43   and lepton efficiency.
44  
45 < In Figures XX and YY we provide the response functions for the
46 < SumJetPt and \met/$\sqrt{\rm SumJetPt}$ cuts used in our analysis,
47 < {\em i.e.} the efficiencies of the experimental cuts as a function of
48 < the true quantities.  Using this information as well as the kinematical
49 < cuts described in Section~\ref{sec:eventSel} and the lepton efficiencies
50 < of Figures~\ref{fig:effttbar}, one should be able to confront
51 < any existing or future model via a relatively simple generator
52 < level study by comparing the expected number of events in 35 pb$^{-1}$
53 < with our upper limit of 4.1 events.
45 >
46 > \subsection{Outreach}
47 > \label{sec:outreach}
48 > Conveying additional useful information about the results of
49 > a generic ``signature-based'' search such as the one described
50 > in this note is a difficult issue.  
51 > Here we attempt to present our result in the most general
52 > way.
53 >
54 > Models of new physics in the dilepton final state
55 > can be confronted in an approximate way by simple
56 > generator-level studies that
57 > compare the expected number of events in 34.0~pb$^{-1}$
58 > with our upper limit of 4.1 events.  The key ingredients
59 > of such studies are the kinematical cuts described
60 > in this note, the lepton efficiencies, and the detector
61 > responses for SumJetPt and \met/$\sqrt{\rm SumJetPt}$.
62 > The muon identification efficiency is $\approx 95\%$;
63 > the electron identification efficiency varies from $\approx$ 63\% at
64 > $P_T = 10$ GeV to 91\% for $P_T > 30$ GeV.  The isolation
65 > efficiency in top events varies from $\approx 83\%$ (muons)
66 > and $\approx 89\%$ (electrons) at $P_T=10$ GeV to
67 > $\approx 95\%$ for $P_T>60$ GeV.
68 > %{\bf \color{red} The following numbers were derived from Fall 10 samples. }
69 > The average detector
70 > responses for SumJetPt and $\met/\sqrt{\rm SumJetPt}$ are
71 > $1.02 \pm 0.05$ and $0.94 \pm 0.05$ respectively, where
72 > the uncertainties are from the jet energy scale uncertainty.
73 > The experimental resolutions on these quantities are 11\% and
74 > 16\% respectively.
75 >
76 > To justify the statements in the previous paragraph
77 > about the detector responses, we plot
78 > in Figure~\ref{fig:response} the average response for
79 > SumJetPt and \met/$\sqrt{\rm SumJetPt}$ in MC, as well as the
80 > efficiency for the cuts on these quantities used in defining the
81 > signal region.
82 > % (SumJetPt $>$ 300 GeV and \met/$\sqrt{\rm SumJetPt} > 8.5$
83 > % Gev$^{\frac{1}{2}}$).  
84 > %{\bf \color{red} The following numbers were derived from Fall10 samples }
85 > We find that the average SumJetPt response
86 > in the Monte Carlo is about 1.02, with an RMS of order 11\% while
87 > the response of \met/$\sqrt{\rm SumJetPt}$ is approximately 0.94 with an
88 > RMS of 16\%.
89 >
90 > %Using this information as well as the kinematical
91 > %cuts described in Section~\ref{sec:eventSel} and the lepton efficiencies
92 > %of Figures~\ref{fig:effttbar}, one should be able to confront
93 > %any existing or future model via a relatively simple generator
94 > %level study by comparing the expected number of events in 35 pb$^{-1}$
95 > %with our upper limit of 4.1 events.
96 >
97 > \begin{figure}[tbh]
98 > \begin{center}
99 > \includegraphics[width=\linewidth]{selectionEffDec10.png}
100 > \caption{\label{fig:response} Left plots: the efficiencies
101 > as a function of the true quantities for the SumJetPt (top) and
102 > tcMET/$\sqrt{\rm SumJetPt}$ (bottom) requirements for the signal
103 > region as a function of their true values.  The value of the
104 > cuts is indicated by the vertical line.
105 > Right plots: The average response and its RMS for the SumJetPt
106 > (top) and tcMET/$\sqrt{\rm SumJetPt}$ (bottom) measurements.
107 > The response is defined as the ratio of the reconstructed quantity
108 > to the true quantity in MC.  These plots are done using the LM0
109 > Monte Carlo, but they are not expected to depend strongly on
110 > the underlying physics.
111 > %{\bf \color{red} These plots were made with Fall10 samples. }
112 > }
113 > \end{center}
114 > \end{figure}
115 >
116 >
117 >
118 > %%%  Nominal
119 > % -----------------------------------------
120 > % observed events                         1
121 > % relative error on acceptance        0.000
122 > % expected background                 1.400
123 > % absolute error on background        0.770
124 > % desired confidence level             0.95
125 > % integration upper limit             30.00
126 > % integration step size              0.0100
127 > % -----------------------------------------
128 > % Are the above correct? y
129 > %    1  16.685     0.29375E-06
130 > %
131 > % limit: less than     4.112 signal events
132 >
133 >
134 >
135 > %%%  Add 20% acceptance uncertainty based on LM0
136 > % -----------------------------------------
137 > % observed events                         1
138 > % relative error on acceptance        0.200
139 > % expected background                 1.400
140 > % absolute error on background        0.770
141 > % desired confidence level             0.95
142 > % integration upper limit             30.00
143 > % integration step size              0.0100
144 > % -----------------------------------------
145 > % Are the above correct? y
146 > %    1  29.995     0.50457E-06
147 > %
148 > % limit: less than     4.689 signal events
149 >
150 >
151 > \subsection{mSUGRA scan}
152 > \label{sec:mSUGRA}
153 > We also perform a scan of the mSUGRA parameter space, as recomended
154 > by the SUSY group convenors\cite{ref:scan}.
155 > The goal of the scan is to determine an exclusion region in the
156 > $m_0$ vs. $m_{1/2}$ plane for
157 > $\tan\beta=3$,
158 > sign of $\mu = +$, and $A_{0}=0$~GeV.  This scan is based on events
159 > generated with FastSim.
160 >
161 > The first order of business is to verify that results using
162 > Fastsim and Fullsim are compatible.  To this end we compare the
163 > expected yield for the LM1 point in FullSim (3.56 $\pm$ 0.06) and
164 > FastSim (3.29 $\pm$ 0.27), where the uncertainties are statistical only.
165 > These two numbers are in agreement, which gives us confidence in
166 > using FastSim for this study.
167 >
168 > The FastSim events are generated with different values of $m_0$
169 > and $m_{1/2}$ in steps of 10 GeV.  For each point in the
170 > $m_0$ vs. $m_{1/2}$ plane, we compute the expected number of
171 > events at NLO.  We then also calculate an upper limit $N_{UL}$
172 > using cl95cms at each point using the following inputs:
173 > \begin{itemize}
174 > \item Number of BG events = 1.40 $\pm$ 0.77
175 > \item Luminosity uncertainty = 11\%
176 > \item The acceptance uncertainty is calculated at each point
177 > as the quadrature sum of
178 > \begin{itemize}
179 > \item The uncertainty due to JES for that point, as calculated
180 > using the method described in Section~\ref{sec:systematics}
181 > \item A 5\% uncertainty due to lepton efficiencies
182 > \item An uncertaity on the NLO cross-section obtained by varying the
183 > factorization and renormalization scale by a factor of two\cite{ref:sanjay}.
184 > \item The PDF uncertainty on the product of cross-section and acceptance
185 > calculated using the method of Reference~\cite{ref:pdf}.
186 > \end{itemize}
187 > \item We use the ``log-normal'' model for the nuisance parameters
188 > in cl95cms
189 > \end{itemize}
190 >
191 > An mSUGRA point is excluded if the resulting $N_{UL}$ is smaller
192 > than the expected number of events.  Because of the quantization
193 > of the available MC points in the $m_0$ vs $m_{1/2}$ plane, the
194 > boundaries of the excluded region are also quantized.  We smooth
195 > the boundaries using the method recommended by the SUSY
196 > group\cite{ref:smooth}.  In addition, we show a limit
197 > curve based on the LO cross-section, as well as the
198 > ``expected'' limit curve.  The expected limit curve was
199 > calculated using the CLA function also available in cl95cms.
200 > In general we found that the ``expected'' limit is very close
201 > to the observed limit, which is not surprising since the
202 > expected BG (1.4 $\pm$ 0.8 events) is fully consistent
203 > with the observation (1 event). Because of the quantization,
204 > we find that the expected and observed limits are either
205 > identical or differ by one or at most two grid points.
206 > We have approximated the expected limit as the observed limit
207 > minus 10 GeV\footnote{We show the expected limit only because
208 > this is what is recommended by SUSY management. We believe that
209 > quoting the agreement between the expected BG and the
210 > observation should be enough....}.
211 > Finally, we note that the sross-section uncertainties due to
212 > variations of the factorization
213 > and renormalization scale are not included for the LO curve.
214 > The results are shown in Figure~\ref{fig:msugra}
215 >
216 >
217 > \begin{figure}[tbh]
218 > \begin{center}
219 > \includegraphics[width=\linewidth]{exclusion_noPDF.pdf}
220 > \caption{\label{fig:msugra}\protect Exclusion curves in the mSUGRA parameter space,
221 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs.  THIS IS STILL MISSING
222 > THE PDF UNCERTAINTIES.  WE ALSO WANT TO IMPROVE THE SMOOTHING PROCEDURE.}
223 > \end{center}
224 > \end{figure}
225 >
226 >
227 > \subsubsection{Check of the nuisance parameter models}
228 > We repeat the procedure outlined above but changing the
229 > lognormal nuisance parameter model to a gaussian or
230 > gamma-function model.  The results are shown in
231 > Figure~\ref{fig:nuisance}.  (In this case,
232 > to avoid smoothing artifacts, we
233 > show the raw results, without smoothing).
234 >
235 > \begin{figure}[tbh]
236 > \begin{center}
237 > \includegraphics[width=0.5\linewidth]{nuissance.png}
238 > \caption{\label{fig:nuisance}\protect Exclusion curves in the
239 > mSUGRA parameter space,
240 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs
241 > using different models for the nuisance parameters.
242 > PDF UNCERTAINTIES ARE NOT INCLUDED.}
243 > \end{center}
244 > \end{figure}
245 >
246 > We find that different assumptions on the PDFs for the nuisance
247 > parameters make very small differences to the set of excluded
248 > points.
249 > Following the recommendation of Reference~\cite{ref:cousins},
250 > we use the lognormal nuisance parameter model as the default.
251 >
252 >
253 > % \clearpage
254 >
255 >
256 > \subsubsection{Effect of signal contamination}
257 > \label{sec:contlimit}
258 >
259 > Signal contamination could affect the limit by inflating the
260 > background expectation.  In our case we see no evidence of signal
261 > contamination, within statistics.
262 > The yields in the control regions  
263 > $A$, $B$, and $C$ (Table~\ref{tab:datayield}) are just
264 > as expected in the SM, and the check
265 > of the $P_T(\ell \ell)$ method in the control region is
266 > also consistent with expectations (Table~\ref{tab:victory}).
267 > Since we have two data driven methods, with different
268 > signal contamination issues, giving consistent
269 > results that are in agreement with the SM, we
270 > argue for not making any correction to our procedure
271 > because of signal contamination.  In some sense this would
272 > be equivalent to using the SM background prediction, and using
273 > the data driven methods as confirmations of that prediction.
274 >
275 > Nevertheless, here we explore the possible effect of
276 > signal contamination.  The procedure suggested to us
277 > for the ABCD method is to modify the
278 > ABCD background prediction from $A_D \cdot C_D/B_D$ to
279 > $(A_D-A_S) \cdot (C_D-C_S) / (B_D - B_S)$, where the
280 > subscripts $D$ and $S$ refer to the number of observed data
281 > events and expected SUSY events, respectively, in a given region.
282 > We then recalculate $N_{UL}$ at each point using this modified
283 > ABCD background estimation.  For simplicity we ignore
284 > information from the $P_T(\ell \ell)$
285 > background estimation.  This is conservative, since
286 > the $P_T(\ell\ell)$ background estimation happens to
287 > be numerically larger than the one from ABCD.
288 >
289 > Note, however, that in some cases this procedure is
290 > nonsensical.  For example, take LM0 as a SUSY
291 > point.  In region $C$ we have a SM prediction of 5.1
292 > events and $C_D = 4$ in agreement with the Standard Model,
293 > see Table~\ref{tab:datayield}.  From the LM0 Monte Carlo,
294 > we find $C_S = 8.6$ events.   Thus, including information
295 > on $C_D$ and $C_S$ should {\bf strengthen} the limit, since there
296 > is clearly a deficit of events in the $C$ region in the
297 > LM0 hypothesis.  Instead, we now get a negative ABCD
298 > BG prediction (which is nonsense, so we set it to zero),
299 > and therefore a weaker limit.
300 >
301 >
302 >
303 >
304 > \begin{figure}[tbh]
305 > \begin{center}
306 > \includegraphics[width=0.5\linewidth]{sigcont.png}
307 > \caption{\label{fig:sigcont}\protect Exclusion curves in the
308 > mSUGRA parameter space,
309 > assuming $\tan\beta=3$, sign of $\mu = +$ and $A_{0}=0$~GeVs
310 > with and without the effects of signal contamination.
311 > PDF UNCERTAINTIES ARE NOT INCLUDED.}  
312 > \end{center}
313 > \end{figure}
314 >
315 > A comparison of the exclusion region with and without
316 > signal contamination is shown in Figure~\ref{fig:sigcont}
317 > (with no smoothing).  The effect of signal contamination is
318 > small, of the same order as the quantization of the scan.
319 >
320 >
321 > \subsubsection{mSUGRA scans with different values of tan$\beta$}
322 > \label{sec:tanbetascan}
323 >
324 > For completeness, we also show the exclusion regions calculated
325 > using $\tan\beta = 10$ (Figure~\ref{fig:msugratb10}).
326 >
327 > \begin{figure}[tbh]
328 > \begin{center}
329 > \includegraphics[width=\linewidth]{exclusion_tanbeta10.pdf}
330 > \caption{\label{fig:msugratb10}\protect Exclusion curves in the mSUGRA parameter space,
331 > assuming $\tan\beta=10$, sign of $\mu = +$ and $A_{0}=0$~GeVs.  THIS IS STILL MISSING
332 > THE PDF UNCERTAINTIES.  WE ALSO WANT TO IMPROVE THE SMOOTHING PROCEDURE.}
333 > \end{center}
334 > \end{figure}
335 >
336 >
337 >
338 >
339 >
340 >

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines