stascal methods for experimental parcle physicstrj/trjethstatsday3.pdfstascal methods for...
TRANSCRIPT
T.JunkSta+s+csETHZurich30Jan‐3Feb 1
Sta$s$calMethodsforExperimentalPar$clePhysics
TomJunk
PauliLecturesonPhysicsETHZürich
30January—3February2012
Day3:BayesianInference
T.JunkSta+s+csETHZurich30Jan‐3Feb 2
ReasonsforAnotherKindofProbability• Sofar,we’vebeen(mostly)usingtheno+onthatprobabilityisthelimitofafrac+onoftrialsthatpassacertaincriteriontototaltrials.
• Systema+cuncertain+esinvolvemanyharderissues.Experimentalistsspendmuchoftheir+meevalua+ngandreducingtheeffectsofsystema+cuncertainty.
• Wealsowantmorefromourinterpreta+ons‐‐wewanttobeabletomakedecisionsaboutwhattodonext.
• WhichHEPprojecttofundnext?• Whichtheoriestoworkon?• WhichanalysistopicswithinanexperimentarelikelytobefruiXul?
Thesearealldifferentkindsofbetsthatweareforcedtomakeasscien+sts.Theyarefraughtwithuncertainty,subjec+vity,andprejudice.
Non‐scien+stsconfrontuncertaintyandtheneedtomakedecisionstoo!
T.JunkSta+s+csETHZurich30Jan‐3Feb 3
Bayes’TheoremLaw of Joint Probability:
Events A and B interpreted to mean “data” and “hypothesis”
{x} = set of observations {ν} = set of model parameters
A frequentist would say: Models have no “probability”. One model’s true, others are false. We just can’t tell which ones (maybe the space of considered models does not contain a true one). Better language:
describes our belief in the different models parameterized by {ν}
€
p({ν} | data) =L(data |{ν})π (ν)
L(data |{ ′ ν })π ({ ′ ν })d{ ′ ν }∫
€
p({ν} | data)
T.JunkSta+s+csETHZurich30Jan‐3Feb 4
Bayes’Theorem is called the “posterior probability” of the model parameters
is called the “prior density” of the model parameters
The Bayesian approach tells us how our existing knowledge before we do the experiment is “updated” by having run the experiment.
This is a natural way to aggregate knowledge -- each experiment updates what we know from prior experiments (or subjective prejudice or some things which are obviously true, like physical region bounds).
Be sure not to aggregate the same information multiple times! (groupthink)
We make decisions and bets based on all of our knowledge and prejudices
“Every animal, even a frequentist statistician, is an informal Bayesian.” See R. Cousins, “Why Isn’t Every Physicist a Bayesian”, Am. J. P., Volume 63, Issue 5, pp. 398-410
€
p({ν} | data)
€
π ({ν})
T.JunkSta+s+csETHZurich30Jan‐3Feb 5
How I remember Bayes’s Theorem
Posterior “PDF” (“Credibility”)
“Likelihood Function” (“Bayesian Update”)
“Prior belief distribution”
Normalize this so that
for the observed data
T.JunkSta+s+csETHZurich30Jan‐3Feb 6
BayesianApplica$ontoHEPData:SeBngLimitsonanewprocesswithsystema$cuncertain$es
€
L(r,θ) = PPoiss(data | r,θ)bins∏
channels∏
Whererisanoverallsignalscalefactor,andθrepresentsallnuisanceparameters.
€
PPoiss(data | r,θ) =(rsi(θ) + bi(θ))
ni e−(rsi (θ )+bi (θ ))
ni!
whereniisobservedineachbini,siisthepredictedsignalforafiducialmodel(SM),andbiisthepredictedbackground.
Dependenceofsiandbionθincludesrate,shape,andbin‐by‐binindependentuncertain+esinarealis+cexample.
T.JunkSta+s+csETHZurich30Jan‐3Feb 7
BayesianLimitsIncludinguncertain+esonnuisanceparametersθ
€
′ L (data | r) = L(data | r,θ)π (θ)dθ∫whereπ(θ)encodesourpriorbeliefinthevaluesoftheuncertainparameters.UsuallyGaussiancenteredonthebestes+mateandwithawidthgivenbythesystema+c.Theintegralishigh‐dimensional.MarkovChainMCintegra+onisquiteuseful!
Usefulforavarietyofresults:
€
0.95 =
′ L (data | r)π (r)dr0
rlim
∫
′ L (data | r)π (r)dr0
∞
∫
Typicallyπ(r)isconstantOtherop+onspossible.Sensi$vitytopriorsaconcern.
Limits:
PosteriorDen
sity=L′(r)×π(r)
=r
ObservedLimit
5%ofintegral
T.JunkSta+s+csETHZurich30Jan‐3Feb 8
BayesianCrossSec$onExtrac$on
€
′ L (data | r) = L(data | r,θ)π (θ)dθ∫Samehandlingofnuisanceparametersasforlimits
€
0.68 =
′ L (data | r)π (r)drrlow
rhigh
∫
′ L (data | r)π (r)dr0
∞
∫€
r = rmax−(rmax −rlow )+(rhigh−rmax )
Usually:shortestintervalcontaining68%oftheposterior(otherchoicespossible).Usetheword“credibility”inplaceof“confidence”
Ifthe68%CLintervaldoesnotcontainzero,thentheposterioratthetopandbolomareequalinmagnitude.Theintervalcanalsobreakupintosmallerpieces!(example:WWTGC@LEP2
Themeasuredcrosssec+onanditsuncertainty
9T.JunkSta+s+csETHZurich30Jan‐3Feb
ExtendingOurUsefulTipAboutLimitsIttakesalmostexactly3expectedsignaleventstoexcludeamodel.
Ifyouhavezeroeventsobserved,zeroexpectedbackground,andnosystema+cuncertain+es,thenthelimitwillbe3signalevents.
Calls=expectedsignal,b=expectedbackground.r=s+bisthetotalpredic+on.
€
L(n = 0,r) =r0e−r
0!= e−r = e−(s+b )
€
0.95 =
′ L (data | r)π (r)dr0
rlim
∫
′ L (data | r)π (r)dr0
∞
∫=−e−(s+b )
0
rlim
−e−(s+b )0
∞ = e−rlim
Thebackgroundratecancels!For0observedevents,thesignallimitdoesnotdependonthepredictedbackground(oritsuncertainty).ThisisalsotrueforCLslimits,butnotPCLlimits(whichgetstrongerwithmorebackground)
Ifp=0.05,thenr=‐ln(0.05)=2.99573
T.JunkSta+s+csETHZurich30Jan‐3Feb 10
A Handy Limit Calculator
D0(hlp://www‐d0.fnal.gov/Run2Physics/limit_calc/limit_calc.html)hasaweb‐based,menu‐drivenBayesianlimitcalculatorforasinglecoun+ngexperiment,withuncorrelateduncertain+esontheacceptance,background,andluminosity.Assumesauniformprioronthesignalstrength.Computes95%CLlimits(“CredibilityLevel”)
T.JunkSta+s+csETHZurich30Jan‐3Feb 11
Sensitivity of upper limit to Even a “flat” Prior
L. Demortier, Feb. 4, 2005
T.JunkSta+s+csETHZurich30Jan‐3Feb 12
Systema+cUncertain+esEncodedaspriorsonthenuisanceparametersπ({θ}).
Canbequiteconten+ous‐‐injec+onoftheoryuncertain+esandresultsfromotherexperiments‐‐howmuchdowetrustthem?
Donotinjectthesameinforma+ontwice.
Someuncertain+eshavesta+s+calinterpreta+ons‐‐canbeincludedinLasaddi+onaldata.Othersarepurelyaboutbelief.Theoryerrorsovendonothavesta+s+calinterpreta+ons.
T.JunkSta+s+csETHZurich30Jan‐3Feb 13
Aside: Uncertainty on our Cut Values? (answer: no)
• Systematic uncertainty -- covers unknown differences between model predictions and the “truth”
• We know what values we set our cuts to.
• We aren’t sure the distributions we’re cutting on are properly modeled.
• Try to constrain modeling with control samples (extrapolation assumptions)
• Estimating systematic errors by “varying cuts” isn’t optimal -- try to understand bounds of mismodeling instead.
T.JunkSta+s+csETHZurich30Jan‐3Feb 14
Integra$ngoverSystema$cUncertain$esHelpsConstraintheirValueswithData
€
′ L (data | r) = L(data | r,θ)π (θ)dθ∫Nuisanceparameters:θParameterofInterest:r
Example:supposewehaveabackgroundratepredic+onthat’s50%(frac+onally)uncertain‐‐goesintoπ(θ).Butonlyanarrowrangeofbackgroundratescontributessignificantlytotheintegral.Thekernelfallstozerorapidlyoutsideofthatrange.
Canmakeaposteriorprobabilitydistribu+onforthebackgroundtoo‐‐narrowbeliefdistribu+on.
T.JunkSta+s+csETHZurich30Jan‐3Feb 15
Coping with Systematic Uncertainty • “Profile:”
• Maximize L over possible values of nuisance parameters include prior belief densities as part of the χ2 function (usually Gaussian constraints)
• “Marginalize:” • Integrate L over possible values of nuisance parameters (weighted by their prior belief functions -- Gaussian, gamma, others...) • Consistent Bayesian interpretation of uncertainty on nuisance parameters
• Aside: MC “statistical” uncertainties are systematic uncertainties
T.JunkSta+s+csETHZurich30Jan‐3Feb 16
Example of a Pitfall in Fitting Models
• Fitting a polynomial with too high a degree • Can extrapolations be trusted?
CEM16_TRK8
Trigger x-section extrapolation vs. luminosity
Lum E30
Trig
ger R
ate
T.JunkSta+s+csETHZurich30Jan‐3Feb 17
OtherPiNallsofFiBngUsuallymethodsrelyingonprofilingandmarginalizingprovidenumericallysimilarresults,butthereareexcep+ons.
Some+mesalikelihoodfunc+onhasmul+plemaxima.
Predic+on=10+2+3.Observedata=12.What’sthebestfitnuisanceparameter?ItsUncertainty?Integra+ngoverthewholeshapeprovidesthemostinforma+on.
Some+mesalikelihoodfunc+onhasadiscon+nuousfirstderiva+ve(careshouldbetakentoavoidthis,butsome+mesithappens.e.g.usingBarlowandBeeston’sTFrac+onFilerinalikelihoodfunc+on).
MINUITgetsstuckincorners.Uncertaintyinfitvalueisalsoill‐defined.
L
θ
L
θ
T.JunkSta+s+csETHZurich30Jan‐3Feb 18
AsymmetricUncertain$esandPriors
Measurements,andeventheore+calcalcula+ons,frequentlyareassignedasymmetricuncertain+es:
Value=10+2‐1,ormoreextremely,10+2+2(ouch).Whentheuncertain+eshavethesamesignonbothsides,itisworthwhiletocheckandseewhythisisthecase.
Example–weseekabumpinamassdistribu+onbycoun+ngeventsinasmallwindowaroundwherethebumpissought.
Thedetectorcalibra+onhasanenergyuncertainty(magne+cfieldorchamberalignmentfortracks,ormuchlargereffect,calorimeterenergyscalesforjets).
Shivthecalibra+onscaleup–predictedpeakshivsoutofthewindowdownwardshivinexpectedsignalpredic+on.
Shivthecalibra+ondown–predictedpeakshivsoutoftheothersideofthewindowdownwardshivinexpectedsignalpredic+on
T.JunkSta+s+csETHZurich30Jan‐3Feb 19
TreatmentofAsymmetricUncertain$es
Thesecasesareprelyclear–theunderlyingparameter,theenergyscale,hasa(Gaussian?Yourchoice)distribu+on,whileithasanonlinear,possiblynon‐monotonicimpactonthemodelpredic+on.
Thesameparametermayhavealinear,symmetricalimpactonanothermodelpredic+on,andwewillhavetotreatthemascorrelatedinsta+s+calanalysistools.
Treatmentisambiguouswhenlilleisknownwhytheuncertain+esareasymmetric,oritisnotclearhowtoextrapolate/interpolatethem.
SeeR.Barlow,“AsymmetricSystema+cErrors”,arXiv:physics/0306138“AsymmetricSta+s+calErrors”,arXiv:physics/0406120
T.JunkSta+s+csETHZurich30Jan‐3Feb 20
Quadra$cImpactsofAsymmetricUncertain$es
R.Barlow
T.JunkSta+s+csETHZurich30Jan‐3Feb 21
R.Barlow
Resul$ngPriorDistribu$onsforalterna$vehandlingofAsymmetricImpacts
T.JunkSta+s+csETHZurich30Jan‐3Feb 22
OtherIdeasonHandlingAsymmetricUncertain$es
• Quadra+cinterpola+onforsmallvaluesoftheuncertainparameter‐‐avoidsthekinkatzero
• Gradualswitchover(useanexponen+alorotherasympto+cfunc+on)tolinearforlargevaluesofthenuisanceparameter‐‐avoidslargequadra+cdivergencefrommoresensiblelinearextrapola+on
Arbitrary!Butthisone’snice.Whatareourcriteriaforwhat’s“nice”?
Preservethemeanofthepriordistribu+ontobethecentralvalue.Otherwisepeoplewillcomplainofbias.
Preservethemedianofthepriordistribu+ontobethecentralvalue.Otherwiseanup‐varia+onintheparameterwillproduceadown‐varia+onintheimpactedpredic+on.
Preservethemodeofthepriordistribu+onThebestfitvalueshouldbethecentralpredic+on.
Wemaybeaskingtoomuch!Whatdoes1+10‐1mean,anyway?
T.JunkSta+s+csETHZurich30Jan‐3Feb 23
Sta$s$calUncertain$esonSystema$cUncertain$es?Answer:No.Butsomesystema+cuncertain+esaredifficulttoevaluateproperly.
SeeRogerBarlow’s“Systema+cerrors:FactsandFic+ons”,arXiv:hep‐ex/0207026
Theidea:Ifasystema+cuncertaintyises+matedbycomparingtwodatasamplesortwoMCsamples,ordatavs.MC,thenifoneorbothofthemhavealimitedsize,thenthemagnitudeofthesystema+ccanbepoorlyconstrained.
Ideally,workharder(runmoreMC)togetabelerpredic+onoftheexpectedsignalandbackground,underallassump+onsofsystema+cvaria+on.
MonteCarloSta$s$calUncertaintyisaSystema$cUncertaintybutdon’tdouble‐countitforeachseparateMCvaria+onofeachnuisanceparameter.Easytodobycomparingcentralvs.variedMCsamples.
Sta+s+callyweaktestsshouldbehandedascrosschecks.Iftheyareconsistent,considerthetesttohavepassed,butdonotaddsystema+cuncertainty.Iftheyfail,however,andadiscrepancybetweentwoMC’sordataandMCcannotbeunderstoodandfixed,thenasystema+cuncertaintyiscalledfor.
T.JunkSta+s+csETHZurich30Jan‐3Feb 24
EvenBayesianshavetobealiRleFrequen$st• Ahard‐coreBayesianwouldsaythattheresultsofanexperimentshoulddependonlyonthedatathatareobserved,andnotonotherpossibledatathatwerenotobserved.
Alsoknownasthe“likelihoodprinciple”
• Butwes+llwantthesensi+vityes+mated!Anexperimentcangetastrongupperlimitnotbecauseitwaswelldesigned,butbecauseitwaslucky.
Howtoop+mizeananalysisbeforedataareobserved?
So‐‐runMonteCarlosimulatedexperimentsandcomputeaFrequen+stdistribu+onofpossiblelimits.Takethemedian‐‐metricindependentandlesspulledbytails.
ButevenBayesian/Frequen+stshavetobeBayesian:usethePrior‐Predic+vemethod‐‐varythesystema+csoneachcpseudoexperimentincalcula+ngexpectedlimits.Toomitthisstepignoresanimportantpartoftheireffects.
T.JunkSta+s+csETHZurich30Jan‐3Feb 25
BayesianExample:CDFHiggsSearchatmH=160GeV(anolderone)
=r
Posterior=L′(r)×π(r)
=r
ObservedLimit
5%ofintegral
T.JunkSta+s+csETHZurich30Jan‐3Feb
WhatTheseLookLikefora5.0σObserva+on
CDFSingleTop,3.2�‐1
T.JunkSta+s+csETHZurich30Jan‐3Feb 27
EvenBayesianshavetobealiRleFrequen$stWewouldliketoknowhowthecrosssec+oncalcula+onsbehaveinanensembleofpossibleexperimentaloutcomes.
Procedure:
• Injectasignal.• Varysystema+csoneachpseudoexperiment(whichintegratesoverthemintheensemble)• CalculateBayesiancrosssec+onforeachoutcomeandplotdistribu+on.• Blacklineisthemedian,notthemean• Checkthewidthofthedistribu+onagainstthequoteduncertain+es.Specifically,thedistribu+onof(meas‐inject)/uncertaintyShouldbeaUnit‐widthGaussian(whennotupagainstzero).
ThisisinfactaNeymanconstruc+on!CandoFeldman‐Cousinswiththis(correctforfitbiases,ifany).
T.JunkSta+s+csETHZurich30Jan‐3Feb 28
SomeFeaturesoftheLinearityPlotThedistribu+onoffitoutcomesataninjectedsignalof0isadeltafunc+onatzerowith50%ofthetotalamount.Theother50%ofthedistribu+onhasawidthfromthemeasurementresolu+on.
Whencompu+ngpulls,usetheuperrorifthemeasuredvalueisbelowtheinjectedrate,andthedownerrorifitisabove.
Forafullysystema+cs‐dominatedmeasurement,thebandedgesshouldbestraightlinespoin+ngattheorigin.(e.g,iftheonlyuncertaintywereacceptance).Alsolargelythecaseforhighs/bsta+s+cs‐limitedmeasurements.
Forthismeasurement,therewasasmallsignalandalarge,uncertainbackground.Thetotaluncertaintyonthesignalislessdependentonthevalueofthesignal.
Usingthefitvalueoftheuncertaintycanbebiasing–alsoquoteexpectedfituncertainty
T.JunkSta+s+csETHZurich30Jan‐3Feb 29
AnExampleWhereUsualBayesianSo\wareDoesn’tWork• TypicalBayesiancodeassumesfixedbackground,signalshapes(withsystema+cs)‐‐scalesignalwithascalefactorandsetthelimitonthescalefactor• Butwhatifthekinema+csofthesignaldependonthecrosssec+on?Example‐‐MSSMHiggsbosondecaywidthscaleswithtan2β,asdoestheproduc+oncrosssec+on.• Solu+on‐‐doa2Dscanandatwo‐hypothesistestateachmA,tanβpoint
T.JunkSta+s+csETHZurich30Jan‐3Feb 30
PriorsinNon‐Cross‐Sec$onParameters
Example:takeaflatpriorinmH;canwediscovertheHiggsbosonbyprocessofelimina+on?(assumesexactlyoneHiggsbosonexists,andotherSMassump+ons)
Example:Flatpriorinlog(tanβ)‐‐evenwithnosensi+vity,cansetnon‐triviallimits..
T.JunkSta+s+csETHZurich30Jan‐3Feb 31
BayesianDiscovery?
BayesFactor
€
B = ′ L (data | rmax ) / ′ L (data | r = 0)
Similardefini+ontotheprofilelikelihoodra+o,butinsteadofmaximizingL,itisaveragedovernuisanceparametersinthenumeratoranddenominator.
Similarcriteriaforevidence,discoveryasprofilelikelihood.
Physicistswouldliketocheckthefalsediscoveryrate,andthenwe’rebacktop‐values.
But‐‐oddbehaviorofBcomparedwithp‐valueforevenasimplecase
J.Heinrich,CDF9678hlp://newton.hep.upenn.edu/~heinrich/bfexample.pdf
T.JunkSta+s+csETHZurich30Jan‐3Feb 32
TevatronHiggsCombina$onCross‐CheckedTwoWays
Verysimilarresults‐‐• Comparableexclusionregions• Samepalernofexcess/deficitrela+vetoexpecta+on
n.b.UsingCLs+blimitsinsteadofCLsorBayesianlimitswouldextendthebolomoftheyellowbandtozerointheaboveplot,andtheobservedlimitwouldfluctuateaccordingly.We’dhavetoexplainthe5%ofmHvalueswerandomlyexcludedwithoutsufficientsensi+vity.
r lim=
T.JunkSta+s+csETHZurich30Jan‐3Feb 33
MeasurementandDiscoveryareVeryDifferentBuzzwords:• Measurement=“PointEs+ma+on”• Discovery=“HypothesisTes+ng”
Youcanhaveadiscoveryandapoormeasurement!Example:Expectedb=2x10‐7events,expectedsignal=1event,observe1event,nosystema+cs.
p‐value~2x10‐7isadiscovery!(hardtoexplainthateventwithjustthebackgroundmodel).Buthave±100%uncertaintyonthemeasuredcrosssec+on!
Inaone‐binsearch,allteststa+s+csareequivalent.Butaddinasecondbin,andthemeasuredcrosssec+onbecomesapoorerteststa+s+cthanthera+oofprofilelikelihoods.
Inallprac+cality,discriminantdistribu+onshaveawidespectrumofs/b,eveninthesamehistogram.Butsomegoodbinswithb<1event
T.JunkSta+s+csETHZurich30Jan‐3Feb 34
Advantages and Disadvantages of Bayesian Inference
• Advantages: • Allows input of a priori knowledge:
• positive cross-sections • positive masses
• Gives you “reasonable” confidence intervals which don’t conflict with a priori knowledge • Easy to produce cross-section limits • Depends only on observed data and not other possible data • No other way to treat uncertainty in model-derived parameters
• Disadvantages: • Allows input of a priori knowledge (AKA “prejudice”) (be sure not to put it in twice...) • Results are metric-dependent (limit on cross section or coupling constant? -- square it to get cross section). • Coverage not guaranteed • Arbitrary edges of credibility interval (see freq. explanation)
T.JunkSta+s+csETHZurich30Jan‐3Feb 35
AnotherApplica$onofBayesianReasoning:TheKalmanFilter
UsedinHEPtofittracksinapar+cledetector
FromtheWikipediaar+cle
T.JunkSta+s+csETHZurich30Jan‐3Feb 36
Outliers • Sometimes they’re obvious, often they are not. • Best to make sure that the uncertainties on all points honestly include all known effects. Understand them!
L. Ristori, Instantaneous Luminosity at CDF vs. time (a Tevatron store in 2005)
hours
Lum E30
T.JunkSta+s+csETHZurich30Jan‐3Feb 37
Summary
Sta+s+cs,likephysics,isalotoffun!
It’scentraltoourjobasscien+sts,andabouthowhumanknowledgeisobtainedfromobserva+on.
Lotsofwaystoaddressthesameproblems.
Manyques+onsdonothaveasingleanswer.Roomforuncertainty.Probabilityanduncertaintyaredifferentbutrelated.
Thinkabouthowyourfinalresultwillbeextractedfromthedatabeforeyoudesignyourexperiment/analysis‐‐keepthinkingaboutitasyouimproveandop+mizeit.
T.JunkSta+s+csETHZurich30Jan‐3Feb 38
ExtraMaterial
T.JunkSta+s+csETHZurich30Jan‐3Feb 39
Bayesian Upper Limit Calculation
data = n b = background rate s = signal rate (= cross section when luminosity=1)
Multiply by a flat prior π(s) = 1 and find the limit by integrating:
Not too tricky; easy to explain. • But where did π(s) come from? • What to do about systematic uncertainty on signal and background?
T.JunkSta+s+csETHZurich30Jan‐3Feb 40
Frequentist Analysis of Significance of Data
• Most experiments yield outcomes with measure ~0
• A better question: Assuming the null hypothesis is true, what are the chances of observing something as much like the test hypothesis as we did (or more)? used to reject the null hypothesis if small
• Another question: If test hypothesis is true, what are the chances that we’d see something as much like the null hypothesis as we did (or more)? used to reject the test hypothesis if small
It is possible to reject both hypotheses! (but not with C+F or Bayesian techniques).
T.JunkSta+s+csETHZurich30Jan‐3Feb 41
Frequentist Interpretation of Data • Relies on an abstraction -- an infinite ensemble of repetitions of the experiment. Can speak of probabilities as fractions of experiments.
• Constructed to give proper coverage:
95% CL intervals contain the true value 95% of the time, and do not contain the true value 5% of the time, if the experiment is repeated.
• Two kinds of errors: • Accepting test hypothesis if it is false • Excluding test hypothesis if it is true
• Two kinds of success • Accepting test hypothsis if it is true • Excluding test hypothesis if it is false
Difference between “power” and “coverage”
T.JunkSta+s+csETHZurich30Jan‐3Feb 42
Undesirable Behavior of Limit-Setting Procedures • Empty confidence intervals: we know with 100% certainty that an empty confidence interval doesn’t contain the true value, even though the technique produces correct 95% coverage in an ensemble of possible experiments. Odd situation when we know we’re in the “unlucky” 5%.
• Ability of an experiment to exclude a model to which there is no sensitivity. Classic example: fewer selected data events than predicted by SM background. Can sometimes rule out SM b.g. hypothesis at 95% CL and also any signal+background hypothesis, regardless of how small the signal is.
Annoying, but not actually flaws of a technique • Experiments with less sensitivity (lower s, or higher b, or bigger errors) can set more stringent limits if they are lucky than more sensitive experiments • Increasing systematic errors on b can result in more stringent limits (happens if an excess is observed in data).
T.JunkSta+s+csETHZurich30Jan‐3Feb 43
Solution to Annoying Problems -- Expected Limits
• Sensitivity ought to be quoted as the median expected limit (or median discovery probability) or median expected error bar in a large ensemble of possible experiments, not the observed one. Called “a priori limits” in CDF Run 1 parlance.
• Systematic errors will always weaken the expected limits (observed limits may do anything!)
• Best way to compare which analysis is best among several choices -- optimize cuts based on expected limits is optimal
Approximations to expected limit:
Approxima+ontoexpecteddiscoverysignificance
T.JunkSta+s+csETHZurich30Jan‐3Feb 44
Systematic Uncertainties in Fequentist Approaches • Can construct multi-dimensional Confidence intervals, with each nuisance parameter (=source of uncertainty) constrained by some measurement.
• Not all nuisance paramters can be constrained this way -- some are theoretical guesses with belief distributions instead of pure statistical experimental errors.
• Systematic uncertainty is uncertainty in the predictions of our model: e.g., p(data|Standard Model) is not completely well determined due to nuisance parameters
• One approach -- “ensemble of ensembles” -- include in the ensemble variations of the nuisance parameters.
(even Frequentists have to be a little Bayesian sometimes)
T.JunkSta+s+csETHZurich30Jan‐3Feb 45
IndividualCandidatesCanMakeaBigDifference
ifs/bishighenoughneareachone.
Finemassgrid‐‐smoothinterpola+onofpredic+ons‐‐someanalysisswitchoversatdifferentmHforop+miza+onpurposes
AtLEP‐‐canfollowindividualcandidates’interpreta+onsasfunc+onsoftestmass
T.JunkSta+s+csETHZurich30Jan‐3Feb 46
APiNall‐‐NotEnoughMC(ordatainsidebandregions)ToMakeAdequatePredic$ons
Cousins,TuckerandLinnemanntelluspriorpredic+vep‐valuesundercoverwith0±0eventsarepredictedinacontrolsample.
CTLProposeaflatpriorintruerate,usejointLFincontrolandsignalsamples.Problemis,themeanexpectedeventrateinthecontrolsampleisnobs+1incontrolsample.Finebinning→biasinbackgroundpredic+on.
Overcoversfordiscovery,undercoversforlimits?
AnExtremeExample(namesremoved)