New Physics at the LHC. A Les Houches Report: Physics at TeV Colliders 2009 - New Physics Working Group
G. Brooijmans, C. Grojean, G.D. Kribs, C. Shepherd-Themistocleous, K. Agashe, L. Basso, G. Belanger, A. Belyaev, K. Black, T. Bose, R. Brunelière, G. Cacciapaglia, E. Carrera, S.P. Das, A. Deandrea, S. De Curtis, A.-I. Etienvre, J.R. Espinosa, S. Fichet, L. Gauthier, S. Gopalakrishna, H. Gray, B. Gripaios, M. Guchait, S.J. Harper, C. Henderson, J. Jackson, M. Karagöz, S. Kraml, K. Lane, T. Lari, S.J. Lee, J.R. Lessard, Y. Maravin, A. Martin, B. McElrath, G. Moreau, S. Moretti, D.E. Morrissey, M. Mühlleitner, D. Poland, G.M. Pruna, A. Pukhov, A.R. Raklev, T. Robens, R. Rosenfeld, H. Rzehak, G.P. Salam, S. Sekmen, G. Servant, R.K. Singh, B.C. Smith, M Spira, M.J. Strassler, I. Tomalin, M. Tytgat, M. Vos, J.G. Wacker, P. v. Weitershausen, K.M. Zurek
NNEW PHYSICS AT THE LHC: A LES HOUCHES REPORT
Physics at TeV Colliders 2009 – New Physics Working Group
G. Brooijmans , C. Grojean , , G.D. Kribs and C. Shepherd-Themistocleous (convenors) K. Agashe , L. Basso , , G. Belanger , A. Belyaev , , K. Black , T. Bose , R. Brunelière ,G. Cacciapaglia , , E. Carrera , , S.P. Das , A. Deandrea , , S. De Curtis ,A.-I. Etienvre , J.R. Espinosa , S. Fichet , L. Gauthier , S. Gopalakrishna , H. Gray ,B. Gripaios , M. Guchait , S.J. Harper , C. Henderson , J. Jackson , , M. Karagöz ,S. Kraml , K. Lane , T. Lari , S.J. Lee , J.R. Lessard , Y. Maravin , A. Martin ,B. McElrath , G. Moreau , S. Moretti , , , D.E. Morrissey , M. Mühlleitner , D. Poland ,G.M. Pruna , , A. Pukhov , A.R. Raklev , T. Robens , R. Rosenfeld , H. Rzehak ,G.P. Salam , S. Sekmen , G. Servant , , R.K. Singh , B.C. Smith , M Spira ,M.J. Strassler , I. Tomalin , M. Tytgat , M. Vos , J.G. Wacker , P. v. Weitershausen ,and K.M. Zurek Abstract
We present a collection of signatures for physics beyond the standard model that need to be ex-plored at the LHC. First, are presented various tools developed to measure new particle massesin scenarios where all decays include an unobservable particle. Second, various aspects of su-persymmetric models are discussed. Third, some signatures of models of strong electroweaksymmetry are discussed. In the fourth part, a special attention is devoted to high mass reso-nances, as the ones appearing in models with warped extra dimensions. Finally, prospects formodels with a hidden sector/valley are presented. Our report, which includes brief experimen-tal and theoretical reviews as well as original results, summarizes the activities of the “NewPhysics” working group for the “Physics at TeV Colliders" workshop (Les Houches, France,8–26 June, 2009).
Acknowledgements
We would like to heartily thank the funding bodies, the organisers (G. Bélanger, F. Boudjema,L. di Ciaccio, P.A. Delsart, S. Gascon, C. Grojean, J.P. Guillet, S. Kraml, R. Lafaye, G. Moreau,E. Pilon, G. Salam, P. Slavich and D. Zerwas), the staff and the other participants of the LesHouches workshop for providing a stimulating and lively environment in which to work. a r X i v : . [ h e p - ph ] M a y Physics Department, Columbia University, New York , USA CERN, Physics Departement, Theory Unit, Geneva, Switzerland IPhT, CEA–Saclay, Gif-sur-Yvette, France Department of Physics, University of Oregon, Eugene, USA Particle Physics Department, STFC, Rutherford Appleton Laboratory, Didcot, UK Maryland Center for Fundamental Physics, Department of Physics, U. of Maryland, USA School of Physics & Astronomy, University of Southampton, Highfield, Southampton, UK LAPTH, Université de Savoie, CNRS, Annecy-le-Vieux, France Laboratory for Particle Physics and Cosmology, Harvard University, Cambridge, USA Department of Physics, Boston University, Boston, USA Fakultät für Mathematik und Physik, Albert-Ludwigs-Universität, Freiburg, Germany Université Lyon 1, Villeurbanne, France Institut de Physique Nucléaire de Lyon, CNRS/IN2P3, UMR5822, Villeurbanne, France Department of Physics, Boston University, Boston, USA AHEP Group, Institut de Física Corpuscular – C.S.I.C./Universitat de València, Spain INFN, Sesto Fiorentino, Firenze, Italy IRFU/Service de physique des particules, CEA–Saclay, Gif-sur-Yvette, France ICREA and IFAE, Universitat Autònoma de Barcelona, Barcelona, Spain LPSC, UJF Grenoble 1, CNRS/IN2P3, Grenoble, France The Institute of Mathematical Sciences, C.I.T. Campus, Taramani, Chennai, India Physics Department, California Institute of Technology, Pasadena, USA Department of High Energy Physics, Tata Institute of Fundamental Research, Mumbai, India CERN, Geneva, Switzerland H.H. Wills Physics Laboratory, University of Bristol, UK University of Oxford, Subdepartment of Particle Physics, Oxford, UK INFN, Sezione di Milano, Milano, Italy Department of Particle Physics, Weizmann Institute of Science, Rehovot, Israel Department of Physics and Astronomy, University of Victoria, Victoria, Canada Kansas State University, Manhattan, USA Fermi National Accelerator Laboratory, Batavia, USA Heidelberg University, Heidelberg, Germany Laboratoire de Physique Théorique, Université Paris XI, Orsay, France Dipartimento di Fisica Teorica, Università di Torino, Torino, Italy TRIUMF, Vancouver, Canada Institute for Theoretical Physics, Karlsruhe Institute of Technology, Karlsruhe, Germany Jefferson Physical Laboratory, Harvard University, Cambridge, USA Skobeltsyn Inst. of Nuclear Physics, Moscow State University, Moscow, Russia Oskar Klein Centre, Department of Physics, Stockholm University, Stockholm, Sweden Department of Physics and Astronomy, University of Glasgow, Glasgow, UK Instituto de Fisica Teorica – UNESP, Sao Paulo, Brazil LPTHE, UPMC Univ. Paris 6, CNRS UMR 7589, Paris, France Department of Physics, Florida State University, Tallahassee, FL 32306, USA Institut für Theoretische Physik und Astrophysik, Würzburg, Germany Paul Scherrer Institute, Villigen PSI, Switzerland Department of Physics and Astronomy, Rutgers University, Piscataway, USA Department of Physics and Astronomy, University of Gent, Gent, Belgium IFIC — centre mixte Univ. València/CSIC, Valencia, Spain Theory Group, SLAC, Menlo Park, USA University of Michigan, Ann Arbor, USA ontents
Introduction 6
G. Brooijmans, C. Grojean, G.D. Kribs and C. Shepherd-Themistocleous
Mass determination methods 7
L. Basso et al.
B. Gripaios
Supersymmetry 35
A.R. Raklev, G.P. Salam and J.G. Wacker
M. Mühlleitner, H. Rzehak and M. Spira
S. Fichet and S. Kraml
G. Belanger, S. Kraml, A. Pukhov and R.K. Singh
Strong EWSB 65
J. Espinosa, C. Grojean and M. Mühlleitner
K. Black et al.
High mass resonances 98
K. Agashe et al. Z (cid:48) discovery potential at the LHC in the minimal B − L model 117 L. Basso et al.
11 Single custodian production in warped extra dimensional models 129
S. Gopalakrishna, G. Moreau and R.K. Singh
12 Four top final states 137
G. Servant, M. Vos, L. Gauthier and A.-I. Etienvre G. Brooijmans, G. Moreau and R.K. Singh
14 Effects of nearby resonances at colliders 148
G. Cacciapaglia, A. Deandrea and S. De Curtis
Hidden sectors 153
15 An exotic photon cloud trigger for CMS 154
C. Henderson
16 Long-lived exotica production at the LHC/Tevatron 160
M.J. Strassler and I. Tomalin
17 A benchmark SUSY Abelian hidden sector 171
D.E. Morrissey, D. Poland and K.M. Zurek ntroduction G. Brooijmans, C. Grojean, G.D. Kribs and C. Shepherd-Themistocleous
The LHC has started colliding proton beams at a center of mass energy of 7 TeV, usheringin a new era of physics at the energy frontier. The exploration of physics in the multi-TeVenergy domain will take another major step forward in 2013 when the LHC will run at a centerof mass energy close to 14 TeV.The minimal discovery scenario for the LHC is the Higgs boson, but it is likely that therewill be a lot more. In the case of observation of the Higgs boson, the mechanism responsible forstabilizing its mass at the electroweak scale should manifest itself. If the Standard Model Higgsboson is shown not to exist, other new particles or interactions fulfilling its role in regulatingthe massive vector boson scattering cross-section should be observed. The LHC’s discoverypotential spans a broad spectrum, including the direct production of the dark matter componentsin the universe and the manifestation of new degrees of freedom in space-time. In this reporta wide variety of new physics signals are studied, exploring mostly areas that have emergedrecently.The first two contributions examine the various tools developed to measure new particlemasses in scenarios where all decays include an unobservable particle, usually a dark mattercandidate. The performance of these tools is evaluated and compared for different new physicsscenarios, illustrating their complementary strengths.A second group of studies use Supersymmetry as a working model. These evaluate mul-tiple aspects: how to observe R-parity violating decays of gluinos, the impact of SUSY-QCDcorrections to MSSM Higgs production, how to distinguish Supersymmetry from Gauge-HiggsUnification, and the allowed region for CP-violating phases in the MSSM.This is followed by two contributions in the area of strong electroweak symmetry break-ing, one on the impact of a composite nature of the Higgs boson on the LHC Higgs discoverypotential, and the second a study of the LHC discovery reach for techni-vector mesons in theirdecays to electroweak vector bosons.Another possibility is that new high mass resonances will produced at the LHC. A firstarticle on that topic reviews processes inspired by models of warped extra dimensions and showsthe results of some applications of the special techniques needed in their discovery. In othercontributions to this section, the LHC discovery potential at multiple center of mass energiesfor the specific case of a Z (cid:48) B − L boson is revisited, the production of heavy Kaluza-Klein quarksand four-top final states are studied, the LHC sensitivity to very wide high mass t ¯ t resonancesis examined, and the effects of nearby resonances are considered.The final set of studies included in this report tackle the novel signatures introduced inrecent hidden sector models. One of these proposes a new trigger scheme for signatures withmultitudes of low energy photons, another defines a set of trigger, reconstruction and analysisbenchmarks that should be appropriate for early LHC searches, and a third explores what wouldhappen if the MSSM were coupled to a new gauged hidden sector with characteristic energyscale in the GeV region.This report does not attempt to present an exhaustive picture of new physics at LHC.However, in presenting a wide variety of signatures motivated by very different models andexploring the performance of sets of techniques to observe and/or measure these new physics5cenarios, it will hopefully serve as a useful resource for the exploitation of the LHC physicspotential. 6 ass determination methods ontribution 1Comparison of mass determination methods at the LHC L. Basso, R. Brunelière, T. Lari, J.-R. Lessard, B. McElrath, T. Robens,S. Sekmen, M. Tytgat and P. v. Weitershausen
Abstract
For any BSM theory, the underlying particle mass spectrum will beamong the first available information on the new physics involved. Amultitude of techniques is currently available to determine the massesof new particles in these models from measured data. Here, we reporton an ongoing study in which different mass determination methodsare applied to a common SUSY event sample, generated including ageneric collider detector simulation. The event sample was producedwith and without the explicit generation of an additional hard jet by thehard matrix element to investigate possible effects of extra hard jet ra-diation. We report on first results of this study for several of the morecommonly used mass determination methods. Introduction
The start of data taking at CERN’s Large Hadron Collider at the end of 2009 [1] promises thebeginning of an exciting era for both Standard Model physics and beyond the Standard Modelsearches. Currently, many BSM models are on the market, which promise to solve some SMinherent puzzles such as the hierarchy problem, or the absence of dark matter candidates. Thesemodels typically introduce additional massive particles, where coupling and mass exclusionlimits are obtained from past and current BSM collider searches [2]. For the center of massenergies at the LHC, many allowed scenarios exist where the new particles are produced at arelatively high rate, and typically decay through long decay chains containing both SM andBSM decay products. The measurement of these BSM masses at the LHC will be among thefirst available information about the structure of the underlying theory.In the past years, a large number of methods, widely varying in applicability and accuracy,have been proposed for measuring the masses of the new particles at colliders (for a compre-hensive review, see Ref. [3]). Many of the well established methods have already been tested tohigh accuracy in realistic experimental setups, where parton showers, hadronization, and detec-tor simulation are all included. However, similar studies for many of the more recently proposedvariables, as well as a consistent comparison of the existing methods are still lacking. Here, weinitiate a comparative investigation of various mass determination methods. For this, we usecommon Monte Carlo samples for the mSUGRA scenario SPS1a [4] , where parton shower,decays and hadronization were included and a generic LHC detector response was modelledwith a fast detector simulator. We also produced event samples where one hard jet is explicitelygenerated by the hard matrix element, and matched with the parton shower using the MLM The superpartner masses and production cross-sections for this scenario are given in App. A.
14 TeV . We focused on aluminosity of
10 fb − . Our results should apply for the first stage analyses at the nominal LHCenergy. Event generation and detector simulation
In this project, we have generated events for supersymmetric processes and the two main back-ground processes. In order to include a better description of initial state radiation with high P T and three-body decays which could affect mass determination, event generation of the super-symmetric signal has been performed in three steps:1. Matrix element generation has been done with Madgraph 4.4.24 [6, 7]. Samples weredivided according to the different final states ˜ g ˜ g , ˜ g ˜ q , ˜ q ˜ q and χχ and the number of QCDradiations (0 or 1). ˜ g means a gluino, ˜ q is a squark and χ is either a chargino or a neu-tralino. Samples generated with no or one additional QCD radiation will be named in thefollowing → and → processes, respectively.2. As a second step, the particles produced during matrix element event generation are pro-vided to BRIDGE. BRIDGE v1.8 [8] is used to decay supersymmetric particles accordingto its own decay rates using all possible 2 and 3-body decays.3. Finally, decayed events are passed to Pythia [9] version 6.420 for parton showering andhadronization. The merging of samples with different parton multiplicity is also per-formed during this last step using the MLM matching scheme as explained in [5, 10].The main matching parameter Q match , used to determine whether a jet after showering ismatched to one of the initial partons, is set to
40 GeV . In order to avoid double countingbetween e.g. ˜ g ˜ g and ˜ g ˜ q ¯ q [5], events from the latter process including an intermediategluino resonance are excluded.The W + jets and t ¯ t + jets backgrounds have been generated with Alpgen [11] plus Pythia [9]generators using the standard MLM matching procedure. Simulation for a multipurpose LHC detector response was implemented using the fast simula-tion package Delphes 1.8 [12]. Simulation includes a tracking system embedded into a magneticfield, calorimeters, a muon system, and very forward detectors arranged along the beamline, andtakes into account the effect of magnetic field, the granularity of the calorimeters and subdetec-tor resolutions. We have used the default detector configuration. Definitions of objects used inthe analysis are given in App. B.All generated signal and background samples were stored in the Monte Carlo DatabaseMCDB [13]. 9
Mass variables
The effective mass ( M eff ) is used to estimate the SUSY mass scale ( M SUSY ). For hadronic pro-cesses, M SUSY usually refers to the masses of the strongly interacting SUSY particles. Authorsin [14] use M SUSY as the lowest of these masses, while the author in [15] defines it to be theiraverage. Similarly, there is no universal way to define M eff . The most widely used is describedin equation (1). M eff = p T , + p T , + p T , + p T , + E missT (1)It is mainly used in the 4 jets + E missT channel. Nevertheless, the 2 jets + 2 leptons + E missT channel will also be studied in this note. In this latter case, p T , and p T , of equation (1)correspond to the leptons transverse momenta instead of the jets transverse momenta.Independently of the definition of M eff and M SUSY , the strategy is always the same. Thecorrelation between these two is determined by simulating many points in SUSY parameterspace. This correlation has been shown to be linear, although the correlation coefficient variessignificantly depending on the exact definition of these two variables and the SUSY modelused [15]. ˆ s / ) The ˆ s / variable, equation (2), is another variable used to establish the SUSY energy scale.It is designed to have its distribution peak at the threshold center of mass energy ( ˆ s / ) ofthe studied process [16]. In the context of SUSY processes produced in hadron colliders, thethreshold value of ˆ s / corresponds to about twice the mass of the lightest gluino or squark. ˆ s / ( M inv ) ≡ (cid:112) E − P z + (cid:113) ( E missT ) + M (2)The total visible energy is E = (cid:80) i E i and the corresponding longitudinal momentum is P z = (cid:80) i E i cos θ i , where index i labels the calorimeter towers. M inv is the sum of the masses ofall the particles that cannot be detected (invisible). When muons are present in the event, theirenergy is added to E and their longitudinal momentum is added to P z . When particles in the final state escape detection, their momentum can only be constrained fromthe missing momentum in the transverse plane. Therefore, a simple variable that can extract theabsolute masses of intermediate particles is the transverse mass. Such a variable does not relyon the event topology. The only requirement for the variable to work properly is that all themissing energy comes from the same particle (which mass is to be reconstructed). If such arequirement is not matched, it gets harder to extract information. A typical suitable event is ofthe form: A + X → B (vis) + C (inv) + X , (3)where the particle A decays into some visible ( B ) and some invisible ( C ) particles. We use X to identify anything else taking part in the event and not being important here.10everal definitions of the transverse mass exist in the literature: we quote here Barger’sTransverse Mass [17]: M T = (cid:18)(cid:113) M (vis) + (cid:126)p T2 (vis) + | /p T | (cid:19) − ( (cid:126)p T (vis) + /(cid:126)p T ) , (4)where (vis) means the sum over the visible particles one wants to consider. A general featureof this variable is the prominent peak and sharp edge at the absolute mass of the parent particle A . However, this variable assumes that all the missing energy comes from one particle, whichis not generally true for BSM models with pair production and a stable particle, such as in theMSSM considered here. M T Stransverse Mass and M T Kink
The M T stransverse mass collider observable first introduced in [19] is useful to measuremasses of pair produced particles, with each of them decaying to one or more directly visi-ble particles and one invisible particle leading to missing transverse momentum. It was shownthat the endpoint of the M T distribution yields an estimate of the mass of the decaying particle,provided that the mass of the invisible daughter is known. The method is especially suited forR-parity conserving SUSY models, where superparticles are pair produced and the LSP at theend of the decay chains is stable and undetectable. As an example, we give the expression for M T as originally derived in [19] for slepton pair production pp → X + ˜ l ˜ l → X + l ˜ χ l ˜ χ M T ≡ min (cid:54) p + (cid:54) p = (cid:54) p T (cid:2) max (cid:8) m T ( p l T , (cid:54) p ) , m T ( p l T , (cid:54) p ) (cid:9)(cid:3) , (5)with m T ( p l i T , (cid:54) p i ) = m l i + m χ + 2( E T l i E T i − p T l i (cid:54) p i ) and E T = (cid:112) p T + m and where theminimization runs over all possible 2-momenta, (cid:54) p , (corresponding to the unknown 2-momentaof the two neutralinos), such that their sum equals the total missing transverse momentum, (cid:54) p T ,observed in the event. The condition that the mass of the invisible daughter is known beforehandis of course a problem since none of the SUSY particle masses have been determined yet.However, this problem can be avoided with the M T kink method introduced in [20], in whichthe M T endpoint distribution considered as function of a trial mass for the invisible particlemay reveal a kink yielding the exact two unknown particle masses separately. In the example ofslepton pair production given above, where the decay of the mother particle contains 1 visibleparticle, the strength of the kink depends on p T ( X ) (or the total p T of the slepton pair system)and the kink is expected to disappear for p T → [21, 22]. In contrast to other methods, mass determination from edges does not rely on a specific eventtopology. The method is typically used for long decay chains of the form A → B + C → B + D + E → ... (6)where the intermediate decay chain particles are taken onshell; in general, it can be used toextract masses from decay chains of arbitrary length . From the four-momenta of the outgoing Notice that this is not the case for chiral bosons, for which the Jacobian peak is absent [18]. For → and → decays, only relative mass differences can be determined. m ab...n = ( p a + p b + ... + p n ) (7)are constructed. The minimal and maximal allowed values of these variables, which are visibleas “edges” in the respective distributions, are completely determined by phase space and givenin terms of the decay-chain masses only, therefore being independent of the total energy ofthe process. The explicit analytic form of the distribution endpoints depends on relative masshierarchies between the intermediate onshell particles; in case of no a priori knowledge, allpossible sets of inversion relations need to tested. Studies of edges have been presented ine.g. [14, 23–25], and these (and similar) variables have found wide applications.In our present study, we focus on the decay chain ˜ q → ˜ χ q → ˜ llq → ˜ χ llq, (8)where we consider the following variables m ll = ( p l + p l ) , m qll = ( p l + p l + p q ) ,m ql (low) = min { ( p l + p q ) , ( p l + p q ) } , m ql (high) = max { ( p l + p q ) , ( p l + p q ) } The endpoints in the distributions of these variables are denoted by m max ll , m max qll , m max ql (low) , m max ql (high) . Unlike the preceding methods, the estimator in the case of Polynomial methods is the mass it-self, and not an auxiliary variable. They work by hypothesizing a kinematic topology consistentwith the particles in the final state, and for each assignment of visible objects to external legs,deriving a polynomial equation for the event. This polynomial is a function of unknown kine-matic quantities and masses. One may then consider different ways to solve these polynomialsby making further assumptions. Applications of these ideas were pursued in Ref. [26] for a sin-gle decay chain with some masses known. Considering both decay chains simultaneously canpotentially give us more information and allow a better determination of the masses [20, 27, 28].Ref. [27] considered symmetric decay chains with two intermediate resonances on each side,Ref. [29, 30] considered symmetric chains with three intermediate resonances on each side, andRef. [31] used this same symmetric 3-resonance topology but omitted the quadratic missingmass shell condition and instead used a likelihood to achieve similar results. The relationshipbetween these variables and the M T "kink" observable was explored in Ref. [32]. Figure 1:
The event topology considered. ( M Z =) ( p + p + p + p ) = ( p + p + p + p ) , ( M Y =) ( p + p + p ) = ( p + p + p ) , ( M X =) ( p + p ) = ( p + p ) , ( M N =) p = p . (9)where p i is the 4-momentum for particle i ( i = 1 . . . . Since the only invisible particles are and and since we can measure the missing transverse energy, there are two more constraints: p x + p x = p xmiss , p y + p y = p ymiss . (10)Given the 6 constraints in Eqs. (9) and (10) and 8 unknowns from the 4-momenta of the missingparticles, there remain two unknowns per event. The system is under-constrained and cannot besolved. This situation changes if we use a second event with the same decay chains, under theassumption that the invariant masses are the same in the two events. Denoting the 4-momentain the second event as q i ( i = 1 . . . , we have 8 more unknowns, q and q , but 10 moreequations, q = q = p , ( q + q ) = ( q + q ) = ( p + p ) , ( q + q + q ) = ( q + q + q ) = ( p + p + p ) , ( q + q + q + q ) = ( q + q + q + q ) = ( p + p + p + p ) ,q x + q x = q xmiss , q y + q y = q ymiss . (11)Altogether, we have 16 unknowns and 16 equations. The system can be solved numerically andwe obtain discrete solutions for p , p , q , q and thus the masses m N , m X , m Y , and m Z . Notethat the equations always have 8 complex solutions, but we will keep only the real and positiveones which we henceforth call “solutions”. The code used to solve the polynomials is publiclyavailable in Ref. [33]. Results
All analyses use object definitions and cuts as given in App. B if not stated otherwise.
The effective mass distribution is shown in Fig. 2.The M eff from SUSY events can be clearly distinguished from the backgrounds consid-ered . This makes M eff a good variable for early SUSY discovery. Moreover, given a goodunderstanding of the backgrounds, the M eff distribution from SUSY events could be deduced.From Fig. 3, the peak of the distribution could be established with a precision of 10 to 100 GeV.Nevertheless, to estimate M SUSY from the distribution, the corresponding SUSY model needsto be known. This is needed to find the correlation between M eff and M SUSY via MC simula-tion. However, an effective mass analysis cannot discriminate between different SUSY models.Consequently, external input from other studies is needed to estimate M SUSY when using M eff . Due to computing constraints, multi-jets from QCD have not been simulated. They could be a significantsource of background for the 4 jets + E missT channel although we are confident that requiring E missT > GeV inthe analysis would keep this type of background under control. ffective Mass (GeV)0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b − eve n t s / f b channel missT SUSYW+JetsTTBAR
Effective Mass (GeV)0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b − eve n t s / f b channel missT SUSYW+JetsTTBAR
Figure 2:
The distribution of the effective mass for the 4 jets + E missT channel (left) and the 2 jets + 2leptons + E missT channel (right). The SUSY events are in purple, while the background from top-antitopand W+jets events are in red and green respectively. Effective Mass (GeV)0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b SUSY missT missT
Figure 3:
The distribution of the effective mass for the 4 jets + E missT channel (blue) and the 2 jets + 2leptons + E missT channel (red) when only SUSY events are present. ˆ s / ) Although ˆ s / is model independent, it needs M inv as input. In the SUSY context, it means thatthe neutralino mass needs to be known. The dependence of ˆ s / on M inv is shown in Fig. 4.Another issue with the ˆ s / variable is that it is very sensitive to initial state radiation (ISR).The solution proposed by the authors in [16] is to use only calorimeter towers with | η | smallerthan η max in the calculation of E and P z , equation (2). They choose η max = 1 . based on thefact that this is where the CMS barrel ends. The effect of using different η max is shown in Fig.4.The ˆ s / distribution of SUSY and background events (without QCD multi-jets) in the4 jets + E missT and 2 jets + 2 leptons + E missT channels can be seen in Fig. 5. The ˆ s / SUSYdistribution peaks at a different position than the SM model background making ˆ s / a goodvariable for early SUSY discovery. Nevertheless, it is doubtful that it will be useful in establish-ing the SUSY scale. First, Fig.4 shows that ˆ s / (0) peaks at about 2000 GeV while ˆ s / (1000)peaks around 3000 GeV. It means that a bias of about twice the lightest SUSY particle masswould be introduced. Second, the choice of η max cut can induce another significant deviation.For example, the ˆ s / (0) distribution of the top-antitop background should peak at twice the top14 (GeV) s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b channel missT Inv. Mass = 0 GeVInv. Mass = 200 GeVInv. Mass = 400 GeVInv. Mass = 600 GeVInv. Mass = 800 GeVInv. Mass = 1000 GeV (GeV) s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b channel missT = 1.4 max ! = 2.5 max ! = 3.2 max ! no cut Figure 4:
The left hand side plot shows ˆ s / for six different values of M inv , 0 (blue), 200 (red) , 400(light green), 600 (purple), 800 (cyan) and 1000 (dark green) GeV. The right hand side figure shows the ˆ s / (0) distribution for different η max cuts: η max = 1 . (blue), η max = 2 . (red), η max = 3 . (green)and no η max cut (purple). Both plots are using the 4 jets + E missT channel. mass ( ∼ GeV). However, from Fig. 5, it peaks at 500 GeV in the 2 jets + 2 leptons + E missT channel and at 600 GeV for the 4 jets + E missT channel. It is also worrisome that the variable ˆ s / is channel dependent. These observations lead to conclude that ˆ s / requires a very trustworthyMC to properly understand the effect of the η max cut. (GeV) s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b = 1.4) max ! channel ( missT SUSYW+JetsTTBAR (GeV) s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b s0 500 1000 1500 2000 2500 3000 3500 4000 − eve n t s / f b = 1.4) max ! channel ( missT SUSYW+JetsTTBAR
Figure 5:
The ˆ s / (0) distribution for the 4 jets + E missT (left) and 2 jets + 2 leptons + E missT (right)channels. η max = 1 . is used as described in the text. The general assumption for this variable to work properly is to have only one source of missingenergy. The presence of more than one source in an event (both real particles and detectorleaks) generally spoils the results. In fact, the fraction of events for which the missing energyis effectively coming from just one source, matching the definition of the variable, is small.Therefore, instead of a well defined peak with a sharp edge on a flat distribution, we can expecta smooth distribution peaking at the correct mass value. This is indeed what we see. Figure6 shows the transverse mass distributions for two opposite sign leptons: e + e − (left) and µ + µ − (right) for the SUSY → scenario. The distributions for the SUSY → scenario are verysimilar. 15 (GeV) T M0 50 100 150 200 250 300 350 400 450 500050100150200250 -1 Electrons (GeV) T M0 50 100 150 200 250 300 350 400 450 500020406080100120140160180200220240 -1 Muons
Figure 6:
Transverse mass distribution for e + e − pairs (left) and µ + µ − pairs (right) for the SUSY → scenario. As suggested previously, looking at figure 6 we see a continuous distribution peaking atthe correct (cid:101) χ mass value ( m χ = 181 . GeV). Since the peak is not very prominent, moredetailed analysis of the background is required for a quantitative statement. Also, the shapeof the distribution is not really characteristic: similar studies in the literature [34] showed thattypical SM backgrounds can lead to the same shape.Notice that, given the low statistics, fluctuations may be misunderstood as peaks in thedistribution. Both electron and muon distributions show possible secondary peaks at
GeVand
GeV respectively, none of which corresponding to actual particles in the spectrumgiving rise to pairs of (opposite sign) charged leptons.The conclusion from this study is that the application of the transverse mass to processeswith more than one missing energy source may yield some information, but ultimately, the(simple) transverse mass is not a suitable variable for SUSY or UED events, since more thanone particle is escaping the detection and the definition of eq. (3) is not matched.When applied to the proper events instead, this variable is very powerful in addressingquantitatively the intermediate particle’s mass, as shown in [34] for the B − L model . M T Stransverse Mass and M T Kink
At the SPS1a point used here, the most abundantly produced slepton is the ˜ τ . Fig. 7 shows the M T distribution obtained for the SUSY → sample for both same sign and opposite sign ˜ τ pair production pp → X + ˜ τ ˜ τ → X + τ ˜ χ τ ˜ χ using parton level information and using theexact mass of the invisible LSP ˜ χ , 96.7 GeV (see Table 1). For the computation of M T theOxbridge M T / Stransverse Mass Library [35] was used. As by construction m ˜ τ ≥ M T , theendpoints of these distributions are expected to be a good estimate of the ˜ τ mass, 134.5 GeV,which is clearly the case here. For this particular channel, where each of the ˜ τ decays to 1 vis-ible and 1 invisible daughter, the endpoint of the M T distribution, M max T ( M χ , p T ) , consideredas function of the trial LSP mass, M χ , is expected to exhibit a kink at M χ = m ˜ χ , only whenthe τ ˜ χ τ ˜ χ system is recoiling with significant p T against X [21, 22], as will be demonstratedbelow. As an example, Fig. 8 (left) shows the M max T distribution as function of the trial LSPmass for the same sign ˜ τ pair production events in the ˜ g ˜ q production channel. The distributionwas fitted in the low ( < M χ < GeV) and high ( < M χ < GeV) M χ range with the Only at the parton-level. The detector level analysis is still on going. M max T ( M χ , p T ) taken from [22], which describe the region below ( M χ < m ˜ χ ) and above ( M χ > m ˜ χ ) the kink . The latter functions implicitly also depend on m ˜ χ and m ˜ τ .Instead of doing a 2-dimensional fit we performed a 1-dimensional fit of the M χ dependenceonly, leaving the average p T of the sample as a free parameter in the fit together with the twounknown particle masses, m ˜ χ and m ˜ τ . Here, the average p T ( X ) of the event sample was de-termined to be 234 GeV. As can be seen, both fits describe the M max T perfectly in their respectiveregions and the two curves cross each other at the expected kink position at ( m ˜ χ , m ˜ τ ) . Thefitted mass values were m ˜ χ = 96 ± ± GeV and m ˜ τ = 133 ± ± GeV, whereas (cid:104) p T (cid:105) = 205 ±
58 (370 ± GeV for the low (high) M χ fit respectively, which is in very goodagreement with the actual mass and (cid:104) p T (cid:105) values. The observed kink is expected to disappear for p T ( X ) going to zero. This effect can indeed be seen in Fig. 8 (right) where the M T endpointdistribution is plotted for events with p T ( X ) < GeV and compared to the case without any p T ( X ) cut. The position of the kink is determined by the two particle masses only and thereforeindependent of p T ( X ) , but the kink itself becomes less pronounced for small p T ( X ) values. For p T ( X ) → the entire M T endpoint distribution can be described by one single function of thetwo unknown particle masses [22]. From the above observations, we conclude that the M T (GeV) T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b ! pp q~g~ ! pp q~q~ ! pp ""! pp (GeV) T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b ! pp q~g~ ! pp q~q~ ! pp ""! pp Figure 7: M T distribution for same sign (left) and opposite sign (right) SUSY ˜ τ pair production onparton level. kink method is at least in principle applicable to the ˜ τ pair production channel considered hereand may yield an accurate determination of both the LSP and ˜ τ mass.Turning to an analysis at detector level, in Fig. 9 the M T distributions are shown for τ pair production for the SUSY → sample, where events with exactly 2 same sign or oppositesign τ jets were selected and where the exact LSP mass was used for the M T computation.Due to experimental resolution, possible τ misidentification and due to τ jets not originatingfrom ˜ τ decays, the sharp edges of the distributions are blurred compared to the correspondingparton level distributions in Fig. 7. The precise position of the endpoints will also be affected bye.g. the tau jet calibration. Note also that a study of the different backgrounds for this particularchannel was not yet included here. The extraction of the LSP and ˜ τ mass values from actualmeasured data therefore requires further investigation and will definitely be more challenging. See Eqn. (4.10)-(4.13) of that reference. (GeV) ! M0 50 100 150 200 250 ( G e V ) m ax T M fit !"
Figure 8: M max T distribution as function of the trial LSP mass, M χ , for same sign SUSY ˜ τ pair produc-tion in the ˜ g ˜ q production channel, with the fit revealing the kink at ( m ˜ χ , m ˜ τ ) as described in the text(left) and with different p T ( X ) cuts (right). (GeV) T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b ! pp q~g~ ! pp q~q~ ! pp ""! pp (GeV) T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b T2 M0 50 100 150 200 250 300 - eve n t s / G e V / f b ! pp q~g~ ! pp q~q~ ! pp ""! pp Figure 9: M T distribution for same sign (left) and opposite sign (right) τ pair production for the SUSYsample. We have analysed the signal chain as given in (8), for both → and → SUSY samples,considering l = µ only. For the parameter point considered here, the theoretical values for theendpoints are given by m max ll = 81 GeV , m max qll = 455 /
448 GeV m max ql (low) = 320 /
315 GeV , m max ql (high) = 398 /
392 GeV for initial ( ˜ d, ˜ s ) and ( ˜ u, ˜ c ) squarks respectively. Our experimental signature is exactly one µ + µ − pair, both at generator and detector level. In addition to this "dimuon" signal, we also investi-gate the behaviour of the "pure" signal, which was selected by additionally requiring the exis-tence of the ˜ χ → µ L ˜ µ R decay on generator level. Note that this sample also includes eventswhere the neutralino was not produced according to (8); the majority of these additional eventscomes from direct (cid:101) χ (cid:101) χ production and subsequent decays. In our analysis, we applied stan-18ard Delphes cuts and detector level object definitions (cf appendix), as well as lepton isolationfor all leptons considered. In addition, we cut out the Z peak as well as all invariant massesbelow
10 GeV in m ll for all variables. The overall pure (dimuon) signal cross sections on de-tector level, which take the above mentioned cuts as well as object definitions into account, are .
22 pb (0 .
35 pb) for the → and .
24 pb (0 .
40 pb) for the → sample .Considering the pure signal only, the characteristic triangle-shaped distribution of the m ll vari-able [36] can easily be reproduced on generator level and persists on the detector level, cf.Fig. 10. The dimuon signal contains additional background which peaks at lower energies.About two thirds of the background can be attributed to stau pairpoduction, with subsequentleptonic tau decays. Since the tau does not distinguish between first and second family leptons,this background can be nearly completely reduced by subtracting the m ll distribution for eventswith the e − µ + and e + µ − signatures, respectively (see Fig. 11). After the subtraction, the ex-pected triangular shape is recovered and the edge is clearly visible.Unlike m ll , the m qll , m ql (low) , m ql (high) variables involve identifying the correct quark jet.As an example, we here discuss m qll , where similar results were obtained for the other vari-ables. First, we consider the behaviour of the pure signal without additional background, wherewe now additionally require a squark parent for the (cid:101) χ , such that events stem from the decaychain (8) only. As for m ll , the distribution shape doesn’t change much when moving fromgenerator to detector level, given the correct identification of the jet , cf. Fig. 12. In general,however, combinations with either one of the two hardest jets in the event have to be consid-ered, and each variable will then inevitably include misidentified jets. In Ref. [36], a subtractionmethod similar to the opposite sign opposite flavour subtraction as described above was used.The background resulting from incorrectly identified jets is eliminated by subtracting a massdistribution with a random uncorrelated hard jet, for instance the hardest or second hardest jetfrom a previous event candidate. However, for the low luminosity considered here, this sub-traction method does not immediately result in the expected shape distributions, and furtherinvestigation is needed .Finally, we want to comment on the inclusion of additional backgrounds. Specifically, thepreliminary results for the edge mass method presented here did not take SM background intoaccount. In the high luminosity study [36], however, this background was well under controlafter applying similar suppression techniques as discussed for the SUSY induced backgroundabove. Summarizing, we can say that, given that the SM background is under control, the m ll edge, including all SUSY induced background, is clearly visible even at an early stage ofdata taking, and can be used to constrain the number of unknown masses by one. However,full knowledge of the relative mass spectrum includes edge measurements involving jets. Thesehave proven to be more challenging, and further studies are needed in order to obtain the correctjet assignment for these variables on detector level. We used the Delphes lepton isolation criteria with no track with p T > in the dR = 0 . cone. The relative contributions to the pure signal on detector level for the ˜ g ˜ g / ˜ g ˜ q / ˜ q ˜ q , and (cid:101) χ (cid:101) χ samples are , , , and for the → and , , , and , for the → sample respectively. We here used a χ minimalization in order to identify the "proper" jet at detector level, in order to test detectoreffects on the pure signal distribution. In Ref. [37], a more dedicated study results in percent-range errors for distributions including jets, for a slighltydifferent point in SUSY parameter space. ll (GeV)0 20 40 60 80 100 - eve n t s / G e V / f b mll (GeV)0 20 40 60 80 100 - eve n t s / G e V / f b Figure 10:
The m ll distribution for events exhibiting a µ + µ − signature as well as involving the ˜ χ → µ L ˜ µ R decay, both on generator level (left) and detector level (right), for → sample. The left (right)plot corresponds to 6049 (2173) events. The expected edge at
81 GeV is clearly visible in both samples. mll (GeV)0 20 40 60 80 100 - eve n t s / G e V / f b Figure 11:
The m ll distribution for all µ + µ − events after subtracting the uncorrelated lepton back-ground. The plot corresponds to 2435 events after subtraction. The expected edge at
81 GeV is clearlyvisible. mqll (GeV)0 100 200 300 400 500 600 700 800 900 1000 - eve n t s / G e V / f b mqll (GeV)0 100 200 300 400 500 600 700 800 900 1000 - eve n t s / G e V / f b Figure 12:
The m qll distribution for µ + µ − events involving the ˜ χ → µ L ˜ µ R decay, using the properquark on generator level (left) and corresponding jet on detector level (right). This is from the → sample, corresponding to 5241 and 1803 events respectively. .6 Polynomial Intersection Here we considered the topology of Refs. [29–31], shown in Fig. 1. This occurs in SPS1a inlarge numbers with taus on the external legs, because the stau is the NLSP. This is generic acrossSUSY models in the "coannihilation region", in which the correct relic density is achieved bythe enhanced annihilation cross section due to the near degeneracy of the stau and neutralino. Toachieve precise results one can restrict to only events with smuons or selectrons instead of staus,but the statistics are much lower. Instead here we tried using the taus themselves. We define atau as either an isolated muon, electron, or hadronic tau candidate as defined by Delphes. Thereis inherently missing energy in the tau decays, so we expect resonances to be smeared comparedto refs. [29–31]. Additionally we require:– 2 or more jets with p T > GeV (only p T > GeV jets are considered)– all possible combinations of jets and tau’s are considered.To solve the system of equations presented previously, one must choose two events. Refs.[29, 30] computed all N ( N − possible pairs for N events to avoid questions about subsetsize, which is very CPU intensive, but in principle one can Monte Carlo over pair choice (withreplacement) and as the number of pairs approaches infinity this is mathematically equivalentto taking all possible pairs. In practice the error on mass determination is fundamentally set bythe number of events, therefore one should not need very many more solutions than the numberof events before the errors from pair choice are sub-dominant. Therefore instead of plotting thesolutions from all N ( N − pairs of events, we Monte Carlo’ed over pair choice, plotting allsolutions from each pair, until the number of entries in the histograms were 10 times the numberof events. Future work should quantify the errors on mass determination as a function of thenumber of pairs chosen.There are several possible particles which can appear at each point in the chain, generallywith similar masses, so that no double-peak structures are seen. We have used the entire SPS1a → dataset, so the heaviest particle M Z is always a squark (possibly with an upstreamgluon) with masses from − GeV. The second heaviest M Y is the χ at 181 GeV. Thethird heaviest M X is dominantly ˜ τ at 135 GeV, and the lightest M N is the ˜ χ at 97 GeV. Mass (GeV)0 100 200 300 400 500 600 700 800 900 1000 N u m b e r o f S o l u t i on s Masses (2 jet + 4 tau)
Figure 13: M N (black solid), M X (red dashed), M Y (green dot-dash), and M Z (blue dashed) polynomialsolutions. Our results are shown in Fig.13. We find unsurprisingly that the slepton and neutralinomass peaks are broader than in Refs. [29, 30], due to the extra missing energy from neutrinos.21
Conclusions
In this writeup, we reported on the first results of an ongoing comparative study of different massdetermination methods. We used a common Monte Carlo data sample, which was generated forthe MSSM mSugra point SPS1a, for a proton-proton collider with a c.m. energy of
14 TeV and an integrated luminosity of
10 fb − . Our sample includes parton shower, hadronization,and detector simulation. We investigated several mass determination variables. Most of thesewere specifically designed for a scenario with long decay chains and missing energy from oneor more invisible final state particles. At this stage of the study, comparative statements cannotyet be made. Therefore, we only comment on the status of the analyses and point to directionswhich need to be taken in further investigation.– Effective mass
The effective mass variable is designed to determine the lowest or averageBSM mass scale in the considered process. It only uses transverse information of theinvolved particles and does not rely on additional mass assumptions. In this study, wefound that, assuming the background to be under control, the distribution peaks at theexpected values. However, for a thorough investigation of any BSM model, a parameterscan needs to be done which establishes the relation between M eff and M BSM ; therefore,the final interpretation of the result is highly model dependent.–
Square of shat-min
Similar to the effective mass, the square of shat-min tries to deter-mine an overall scale of the BSM process by exploration of the threshold region of BSMparticle pair-production. In contrast to M eff , this variable directly relies on an additionalinput of the LSP mass, which needs to be determined elsewhere. Furthermore, this vari-able is highly sensitive to initial state radiation. Cutting out ISR events with a rapiditycut leads to a high cut dependence of the result. We therefore conclude that, although inprinciple applicable, the effects of different rapidity cuts need to be further under controlbefore this variable can be used to determine a mass scale for new physics.– Transverse mass
In contrast to the other variables considered in this report, the trans-verse mass is not applicable for scenarios where the missing energy stems from morethan one particle; in a way, our results can be seen as a test of an a priori false assump-tion. As expected, we do not obtain a distinct peak in the M T distributions, but rather abroad spectrum which however peaks at the expected value. This is caused by a smallnumber of events which have effectively one source of missing energy. The distributioncan furthermore be polluted by additional background; therefore, the use of the transversemass is quite limited in the scenario considered here.– M T Stransverse Mass and M T kink The stransverse mass, M T , has the advantageover the transverse mass that the missing energy in the process can come from morethan one particle. At the parton level, the endpoint of the M T distribution of the eventsample considered here could effectively be used to estimate the mass of the decayingparticle. However, the mass of the invisible particle is required as input and needs to bedetermined elsewhere. The latter problem could be overcome by the M T Kink method inwhich the M T endpoint, M max T , is considered as function of a trial mass corresponding tothe invisible particle. A kink effectively appears at a position which depends only on themass of the LSP and the considered decaying particle. The strength of the kink was seento depend on the total P T of the decaying particle pair. Both the mass of the decayingparticle and the LSP could be extracted quite well from a fit to the M max T trial massdependence. However, further investigation at detector level, including the considerationof the SM and BSM background as well as reconstruction inefficiencies, is needed before22 definite statement can be made about the use of this method in our present study.– Edges
Using the information of edges of invariant mass distributions is one of the moreclassical methods for BSM mass determination. It is in principle applicable to any eventtopology which involves on-shell decays of (B)SM particles. In our study, we foundthat edge measurements which only rely on the leptonic information of the event caneasily be determined, especially after a simple background reduction. However, edgemeasurements involving jets are much more challenging, and for the low luminosity con-sidered here, we did not manage to efficiently subtract the background stemming fromwrong combinatorics. This point needs further investigation before any statement aboutjet-related quantities can be made. From the measurement of the dilepton mass only, thenumber of unknown masses can be reduced by one. Note however that, depending onrelative mass hierarchies within the decay chains, different inversion relations hold forextraction of the correct mass assignments. This can in principle lead to further misinter-pretations, even when the complete edge information is available.–
Polynomial Intersection
The polynomial intersection method uses exact solutions for thekinematic configurations of long decay chains with intermediate on-shell particles, andin general can only be applied to specific topologies, as it relies on the overall numberof unknowns and constraints in the considered system. In this study, we investigated atopology with eight external legs, assuming symmetric decay chains. This allows for anexact solution of the polynomial equation system if any two events of the same topologyare combined. Pair assignment in N events as well as the related error determinationproved to pose the biggest challenge in our study. Instead of combining all possible pairchoices, we used a random Monte Carlo pair assignment. The resulting distributionsfor the masses peak at the expected values, where peaks are broad mainly due to extraenergy losses in the tau decays. A big advantage of this method is that all intermediatemasses can be determined and fitted simulaneously. Future investigation concerns errorestimation as well as the inclusion of all backgrounds.We consider this report as a starting point for a more thorough investigation. More detailedstudies adressing the issues mentioned above, as well as the inclusion of all background, areneeded before we can compare different variables in a quantitative way. However, our resultsalready point to advantages and drawbacks of the variables considered here, and further in-vestigation and eventual synergies of different determination methods will hopefully lead topromising results in the near future. Appendices
App. A
SPS1a spectrum
The SPS1a spectrum use here was generated using SOFTSUSY [38] version 2.0.5, with m t =175 GeV . We give the mass values for particles relevant in this study in Table 1, and total crosssections for the → and → samples in Table 2. Branching ratios have been calculatedusing BRIDGE [8] and are available upon request . With slight numerical variations, we reproduce the decay tables in appendix D of [39]. d L . d R . u L . u R . b . b . t . t . l L . l R . τ . τ . ν l . ν τ . g . (cid:101) χ − . (cid:101) χ − . (cid:101) χ . (cid:101) χ . | (cid:101) χ | . (cid:101) χ . Table 1:
Relevant masses for SPS1a in GeV. u = ( u, c ) , d = ( d, s ) , l = ( e, µ ) . X X → → q ˜ q ( j ) 6 .
56 7 . q ˜ g ( j ) 19 .
52 21 . g ˜ g ( j ) 4 .
53 5 . (cid:101) χ (cid:101) χ ( j ) 1 .
97 4 . Table 2:
Production cross sections in pb for p p → X X , for a cm energy of
14 TeV . CTEQ6L1PDFs were used. → sample includes an explicitly generated hard jet, where hard is defined by p T, jet >
40 GeV . App. B
Delphes precuts and object definitions
In all detector level analyses, a minimal set of cuts was used, corresponding to the Delphes [12]pre set cuts. We also list the object definitions on detector level used in all analyses. Additionalcuts might have been applied for different variables; cf the respective subsections for furtherdetails.
Delphes pre cuts – electron/ positron definition: | η | < . in the tracker, p T >
10 GeV – muon definition: | η | < . in the tracker, p T >
10 GeV – taujet definition : p T >
10 GeV – jet definition: p T >
20 GeV ; CDF jet cluster algorithm [40] was used, with R = 0 . – lepton isolation criteria (if applied): no track with p T > in a cone with dR = 0 . around the considered lepton Analysis object definitions – Missing transverse energy: requires E miss T >
100 GeV .– jet criteria: p T, jet >
50 GeV , | η | jet < – electron/ muon: isolated; no track with p T > in a cone with dR = 0 . around theconsidered lepton– any signal involing n leptons: exactly n isolated leptons at detector level Acknowledgements
We thank Benjamin Fuks, Claude Duhr, and Priscila de Aquino for their help during the setupof Feynrules during an earlier version of this study. We also thank Xavier Rouby for clarifying For a more detailed description of the reconstruction algorithm see [12]. ontribution 2LHC mass measurement, algebraic singularities, and thetransverse mass
B. Gripaios
Abstract
I consider the recently-proposed ‘algebraic singularity method’ for mea-suring masses of invisible particles produced at the LHC in arbitrarydecay topologies. I apply the method to the simplest case of a singleparent particle decaying to an invisible daughter particle and a visibledaughter particle, and show that it gives a local approximation to theusual transverse mass variable. In doing so, I identify some issues thatmay need to be taken into consideration in generalizing the algebraicsingularity method to more complicated decay topologies. One is that,in order to measure masses unambiguously with this method, one mayneed to identify not only the presence, but also the nature, of singulari-ties in experimental distributions. Introduction
Invisible particles will be a fact of life at the LHC, nolens volens . Whether they take the formof neutrinos, or dark matter candidates, or even visible particles that escape into dead regionsof the detector, invisible particles will be omnipresent. The problem with invisible particlesis that they carry away kinematic information in events in which they are present, making thereconstruction of events, and hence particle mass measurements, a non-trivial exercise. In thepresence of a concrete dynamical model, missing information is not necessarily a problem, inthat one can simply marginalize the likelihood that comes from the matrix element. But ifwe profess to be ignorant of dynamics (which is certainly the case if we are looking for newphysics), then we must address the question of what can be learnt from the residual kinematicinformation alone.We have known for a long time that the situation is not hopeless. Indeed, in the canonicalexample of a W -boson undergoing a decay to a charged lepton and an invisible neutrino, thetransverse mass variable was exploited long ago in UAs 1 and 2 to measure the mass of the W [41, 42], and even today it provides the best individual measurement [43]. A more modernexample is the top quark, pair produced at the Tevatron and undergoing a decay in the di-leptonic channel: t → b l ν . Here there is an extra complication, in that each decay inthe pair produces an invisible particle, and even more information is lost. Nevertheless, theso-called m T variable [44, 45], has recently successfully been used to measure the top massin this channel [46, 47]. At the LHC, we can expect (or at least hope) to encounter even more More precisely, the measured missing transverse momentum in an event constrains only the sum of the trans-verse momenta of the two neutrinos. W s, resulting in a topology in which a single particle decays to two invisible neutrinos: h → W → l ν [48]. Even worse (or better, depending on one’s perspective), the LHC mayproduce an invisible dark-matter candidate, whose unknown (and, in contrast to the neutrino,non-negligible) mass further increases the number of unknowns. Dark-matter candidates mayalso be multiply produced, if there is a discrete symmetry that guarantees their stability. A finalcomplication is that theories of physics beyond the Standard Model, such as supersymmetry,typically predict a plethora of new states clustered around the TeV scale. Given the presenceof light SM states, these are likely to undergo cascade decays, resulting in sizable combinatoricambiguities in observed final (and initial) states.In recent years, a large number of methods have been proposed for measuring the massesof particles produced in these topologies; reviews may be found in [3,49] and elsewhere in theseproceeedings (along with a complete set of references). Although most of these methods are,to a large extent, ad hoc , in that they focus on a particular decay topology, a number of resultsof a more general nature have been obtained along the way. Among these is the observationthat in longer decay chains, the system of kinematic equations from one or more events maybe sufficient to solve directly for the masses [50]. Even for shorter decays chains, we nowknow that all masses can be measured, given sufficiently many events. Indeed, even in thedecay topology with the fewest constraints, namely one (or more) parent particle(s) undergoinga two-body decay to a visible and an invisible particle, the kinematics allow both the masses ofthe parent and the invisible daughter to be measured [51]. As a corollary, one has the resultthat all masses can be measured in any set of decays where each decay contains only oneinvisible daughter particle, no matter how many visible particles are involved. A third resultis that for variables, such as the transverse mass, that enjoy boundedness properties, issues ofcombinatorics (which arise from assignments of particles to decays [52] or of radiation to theinitial or final state) can be solved by extremization with respect to assignments.Furthermore, we have also begun to arrive at a deeper understanding of kinematics it-self. The breakthrough in this direction came from Cheng and Han [32], who observed thatthe m T variable mentioned above has an interpretation as ‘the’ natural kinematic function forthe topology of pair-produced particles undergoing identical two-body decays, in the follow-ing sense. Imagine writing down the kinematic constraints, corresponding to conservation of(energy-)momentum, and the mass-shell conditions for some assumed decay topology. Now,for a given event, in which some of the energy-momenta are measured and some are not, onemay ask whether the measured momenta impose any constraint on the unknown masses thatappear in the kinematic constraints. Apparently, the answer is negative, because the constraintsare just a set of underconstrained polynomials in the unmeasured momenta and masses. Infact the answer is affirmative, essentially because the masses and energies are restricted to takevalues in R + , whereas the solutions of polynomial equations generally take values in C . Theupshot is that, each event divides the space of unknown mass parameters into an allowed regionand a disallowed region. The boundary of the two regions is defined by the function m T .From this we learn not only that the ad hoc variable m T is a natural kinematic object, butwe also learn that m T encodes all of the information about particle masses that is contained inan event of this topology. This means that, absent dynamic information or other assumptions,there seems to be little point in searching for an alternative variable to measure masses in thistopology.An obvious follow-up question is: what is the function that defines the kinematic bound-27ry for other decay topologies? For simple cases, this is easily answered: For a single parent,two-body decay, it is the transverse mass [53], whereas for asymmetric pair decays (where ei-ther parents or daughters (or both) are different) one is led to a generalized version of m T [53].Unfortunately, addressing this question on a case-by-case basis becomes increasingly difficultas the the decay topology, and the set of kinematic constraints, become increasingly complex.Very recently, I.-W. Kim has proposed [54] a related method, which, although only ap-proximate (in a sense to be defined below) allows an elegant, and more importantly general,algorithm for mass measurement to be defined, for any decay topology.The starting point for the method is to note that the full phase space (defined by thevarious momenta, subject to the kinematic constraints) is smooth (modulo singularities arisingfrom soft and collinear divergences of massless particles, which do not concern us here). Butwhen some of the particles are invisible, we must project out the kinematic variables that gounmeasured; the resulting observable phase space is a singular manifold. The singularities inobservable phase space give rise to singularities in the distributions of functions on observablephase space, which in turn give rise to sharp features that can be easily identified in experimentaldata, notwithstanding the presence of smoothly-varying backgrounds or detector acceptances.These singularities generalize the well-known edges that appear in invariant- or transverse-massdistributions.With this observation in hand, one can define an algorithm for measuring masses, schemat-ically described as follows. Firstly, assume some decay topology, and write down the corre-sponding kinematic constraints. Secondly, identify the locations of the singular points in ob-servable phase space. Thirdly, construct a co-ordinate, called the singularity co-ordinate, in thevicinity of a given singular point, which: (i) vanishes at the singularity; (ii) corresponds to adirection normal to the singular phase space; and (iii) is normalized such that every event hasthe same significance. Fourthly, for each event, find the nearest singularity, and the value of theassociated singularity co-ordinate for that event. Fifthly, plot the distribution over events of thesingularity co-ordinate for all possible guesses for the unknown mass values. When the massguesses are correct, the distribution will feature a singularity of the origin.The reader may already have noted three potential thorns in the side of the algorithm.Firstly, the notion of ‘nearest singularity’ needs an explicit definition if there is more than one.Secondly, one might fear that the local approximation made will be of limited use for a sampleof experimental events that are spread roughly uniformly in phase space [55]. Thirdly, themethod only guarantees the presence of a singularity at the origin of the singularity co-ordinatewhen the hypothesized masses are the correct ones. It does not guarantee the converse, namelythe absence of a singularity at the origin when the hypothesized masses are incorrect.In what follows, I hope to shed some light on these issues by applying the method, ver-batim , to the simplest decay topology, namely a single parent particle undergoing a two-bodydecay to visible particle and an invisible particle. In doing so, we will see that the singularity co-ordinate is just a local approximation to the usual transverse mass variable. We will also see inthese examples that the algorithm, as it stands, does not identify the correct masses uniquely; todo so, one needs to identify not just the presence of a singularity at the origin of the singularityco-ordinate, but also its nature.I start, in the next Section, by considering as a special case the subset of events in whichthe visible daughter particle is produced at rest. I treat the general case of moving visible For a fuller description, see [54]. A special case
Consider a single parent particle Y , of mass m Y , undergoing a two-body decay to an invisibleparticle X , of mass m X , and a visible particle V , of mass m V . Consider, for now, the restrictedsubset of events in which the momentum of the visible system V vanishes. The four-momentaof V , X and Y may, therefore, be written as V µ = ( m V , , , (1) X µ = ( p , p , p ) , (2) Y µ = ( p + m V , p , p ) , (3)and the three mass-shell constraints may be written as V µ V µ = m V , X µ X µ = m X , and Y µ Y µ = m Y . To make the analysis as straightforward as possible, I solve the constraint X µ X µ = m X for p , such that the remaining constraint is g ≡ p + p − M = 0 , (4)where I set (cid:32) m Y − m X − m V m V (cid:33) − m X ≡ M . (5)It is now very simple to apply the method of [54]. The full phase space is defined bythree momenta, namely p and p , subject to the constraint (4). Geometrically, phase space is atwo-sphere embedded in R . According to [54], we should now split the momentum variablesinto those momenta which are measured in an experiment (or ‘known unknowns’ [56]), viz. p ,and those ‘unknown unknowns’ which are not measured, viz. p . The observable phase spaceis then obtained by projecting out the unmeasured p , and is given by the disk p ≤ M , (6)in R . The full phase space is clearly a non-singular manifold (it is, after all, just a two-sphere),but the observable phase space, obtained by projection, exhibits singularities whenever the tan-gent space to the full phase space is parallel to the direction of projection. In our simple ex-ample, this is equivalent to the simple algebraic condition ∂g∂p = 2 p . But when p = 0 ,Eq. (4) implies that p = M , so the singular points of the observable phase space correspondto the boundary of the disk in (6).Now let us build the singularity co-ordinate in the vicinity of a given singular point. Sincethe disk is rotationally symmetric, we may, without loss of generality, choose the singular pointto be at ( p , p ) = (( M, , . Following the rubric of [54], the first step is to choose a systemof orthonormal co-ordinates, ( n, t , t ) in the neighbourhood of the singularity, correspondingto directions normal and tangent to the full phase space (the two-sphere in the case at hand).Thus I write ( p , p ) = (( M + n, t ) , t ) (7)29n these co-ordinates, the constraint (4) may be written as ( M + n ) + t + t − M = 0 = ⇒ n = − M ( t + t ) + O ( n ) (8)The un-normalized singularity co-ordinate is just n ; to normalize it, we restrict the secondfundamental form II = − M ( t + t ) to the invisible direction t . The volume of phase spacein the invisible direction thus scales as (2 M II ) and the normalized singularity co-ordinateis Σ = 2
M n = 2
M δp . For a general singular point on the boundary of the disk, we find Σ = 2 p · δ p = δ ( p − M ) . So the singularity co-ordinate is just the observable ( p − M ) ,linearized about a point where p = M . Note that the singularity co-ordinate depends on theevent (through p ), on the chosen singularity (which defines δ p , and on the hypothesis for themasses (through M ).Having computed explicitly the singularity co-ordinate, we are now in a position to an-swer several questions. Firstly, what is the relation to the usual transverse mass variable? Thetransverse mass variable is defined by m T ≡ m X + m V + 2( ef − p · q ) (9)where p and q are the transverse momenta of X and V , respectively, and e ≡ (cid:112) p + m X and f ≡ (cid:112) q + m V are their transverse energies. For the special case of q = , this reduces to m T ≡ m V + m X + 2 m V (cid:113) p + m X . (10)Now we know that when we hypothesize the correct value m X for the a priori unknown massof X , the distribution of m T has its maximum at m T = m Y . That is to say, the m T distributionhas a singularity (an edge, in fact) at m T = m Y . Equivalently, we can say that the distributionof m T − m Y will be singular at the origin (when the correct hypothesis of the masses m X and m Y is chosen), or indeed that the distribution of the observable ( p − M ) has a singularity atthe origin. But as we saw above, ( p − M ) , when linearized about a singularity, is preciselythe singularity co-ordinate constructed according to the recipe of [54]. Note that it is not correctto say that the transverse mass and the singularity co-ordinate are equivalent, because the latteris linearized about a singular point, whereas the former is not. But it is correct to say that thetransverse mass and the singularity co-ordinate are equivalent in the neighbourhood of a givensingular point, modulo an overall scale factor. Nevertheless, away from the singular point, thetransverse mass and singularity co-ordinate distributions will disagree.Secondly, since there is a whole S of singular points, given by the boundary of the disk,which one should we choose to construct the singularity co-ordinate for a given event? Naïvely,the rotational invariance tells us that any one is as good as any other. But if we choose just asingle point, most events (assuming they are spread uniformly over the disk) will be a long wayaway from the singular point. To counteract this, it is suggested in [54] that for any event, weshould choose the ‘nearest’ singular point to compute the singularity co-ordinate for any event.This then raises the question of what metric defines the concept of nearness, and of whethera singular point thus defined is unique. One answer might be to define the nearest point withrespect to the metric on observable phase space induced by the Euclidean embedding; in thatcase the nearest singularity is obtained by drawing a radius through the event and finding itsintersection with the disk’s boundary. Another solution might be to minimize the singularity30o-ordinate itself, constructed with respect to all singularities. At least in the example here, thisalternative definition yields the same singular point at the nearest one, for a given event.Thirdly, what masses can we actually measure with the singularity co-ordinate in thiscase? We know on the basis of general kinematic arguments that to measure both the invisiblemasses m X and m Y , one needs events in which the parent particle Y has variable transverseboosts with respect to the laboratory frame. But here, we restricted events to the subset with q = 0 , corresponding to a fixed transverse boost of the parent. In this case we know thatwe should only be able to measure the combination of masses given by M using kinematicmethods alone. To see that this is what happens here, we need to recall how the algorithmof [54] is defined. The algorithm instructs us to construct the singularity co-ordinate for allpossible hypothetical values of the unknown masses. It then tells us that for true values of themasses, we will observe a singularity at the origin.Now, since the singularity co-ordinate is just δ ( p − M ) , it is clear that if we insteadchoose wrong values for the masses, the singularity will, in general, have a singularity that istranslated away from the origin. But if one makes a wrong guess for m X and m Y individuallythat yields the right value of M in combination, then the singularity will be at the origin.This is, of course, fully consistent with the kinematic observation that in such a subset ofevents, one can do no better than measure the combination M , and the singularity method doesno worse than any other method in this respect. But it does raise the worry that, in other cases,more than one set of mass values will give rise to a singularity at the origin, even when generalkinematic arguments tell us that the masses can be measured unambiguously. Indeed, this isexactly what will happen when we consider, in the next Section, the general case of visibleparticles with arbitrary momentum.Lastly, can we get rid of the linearization? In this simple case, we can simply take the‘known unknown’ to be p (taking values in R + ). Phase space is then a parabola in R × R + andthe projected phase space is the interval p ∈ [0 , M ] . Then the singularity co-ordinate is just p − M . So in this case, because the constraint is a single quadratic function, linearization isan unnecessary simplification. Now let me proceed to the general case, where V is produced with arbitrary four-momentum.In this case, we have events in which the parent may have an arbitrary boost with respect to thelaboratory frame, and know on general kinematic grounds that it should be possible to measureboth of the unknown masses m X and m Y . We would like to see, explicitly, whether (and if so,how) this may be achieved using the algebraic singularity method.To prevent the proliferation of unknowns, let me consider the case of -dimensionalspacetime. The energy-momentum vectors then become V µ = ( q , q, q ) , (11) X µ = ( p , p, p ) , (12) Y µ = ( p + q , p + q, p + q ) . (13)To render the analysis straightforward, I first use the two constraints V µ V µ = m V and X µ X µ = m X to solve for p and q . To wit, q = (cid:113) q + q + m V (14)31 = (cid:113) p + p + m X . (15)This leaves a set of three ‘known unknowns’, namely { p, q, q } , and one ‘unknown unknown’, p , subject to the single constraint g ≡ p q − pq − p q + m X + m V − m Y = 0 , (16)where p , q are, of course, given by Eq. (14). Geometrically, the full phase space is a three-dimensional hypersurface in R , defined by the quartic constraint (16). The observable phasespace is obtained by projecting with respect to the co-ordinate p , and is singular when ∂g∂p = 2( q p p − q ) = ⇒ q p − p q = 0 . (17)Substituting into (16), it is easy enough to show that, at the singularities, the observable mo-menta satisfy m Y = m X + m V + 2( ef − pq ) . (18)Perhaps unsurprisingly, this is just the condition that the transverse mass variable (9) be at itsmaximum.At a singularity, the normal vector to phase space has direction (cid:0) q − p (cid:1) T , (19)so that the tangent space may be defined by the three vectors (cid:0) (cid:1) T , (cid:0) (cid:1) T , (cid:0) p q (cid:1) T (20)Using these vectors to define the directions of the orthonormal co-ordinates in the neighbour-hood of the singularity, we have nt t t = sin θ cos θ θ − sin θ δpδqδq δp , (21)where I defined tan θ ≡ − q p = − q p = − fe . (22)We are now in a position to compute the singularity co-ordinate. Going through the normaliza-tion procedure given in [54], we end up with Σ = ∂g∂n n, (23) = 2 pq − qp p q ( q sin θ + p cos θ )( δp sin θ + δq cos θ ) , (24) = 2 pq − qp p q ( q δp − p δq ) , (25)32 pf − qeef ( f δp − eδq ) . (26)To show the relation with the transverse mass, linearize (9) about a point with m T = m Y .One obtains m Y − m T = 2 pf − qeef ( f δp − eδq ) . (27)So the singularity co-ordinate is equivalent to the transverse mass variable, expanded linearlyabout its maximum.Let me again make some remarks. Firstly, I pointed out at the beginning of this sectionthat, in this general case, kinematics tells us that it should be possible to measure both invisiblemasses, m X and m Y , in this case. How is this achieved using the singularity co-ordinate?For the transverse mass variable, this may be achieved, at least in principle, in the followingway [51]. Guess a value for the mass m X and compute the resulting distribution over eventsof the transverse mass (9). Extract the endpoint of the distribution. Now plot the endpoint as afunction of the guessed mass m X . Kinematics tells us that the function has a ‘kink’ [57] (thatis, is C , but not C ), at the point where m X takes its true value.For the singularity co-ordinate, we are instructed to compute the value of the co-ordinatefor all events and for all possible hypothesized values of the masses. We should then look forvalues of the masses that give rise to a singularity at the origin in the distribution over eventsof the singularity co-ordinate. Now, just as for the special case considered in the last section,although it is true that the set of mass values giving rise to a singularity at the origin containsthe point corresponding to the true mass values, it is not true that the set contains only this point.This is easily seen by considering a plot of the transverse mass distribution for various values of m X , an example of which is reproduced in Fig. 1. The point is simply that the m T distributionhas an endpoint for all values of m X , and since the distribution is C , but not C at the endpoint,it seems reasonable to describe the endpoint as singular. Therefore, the singularity co-ordinate,which is simply a linearized version of the transverse mass variable, will feature a singularity inthis sense for all values of m X . And for each value of m X , there exists a value of m Y that willmap this singularity to the origin in the singularity co-ordinate.We conclude that the set of masses, { ( m X , m Y ) } that give rise to a singularity at the originof the singularity co-ordinate does not consist of the single point corresponding to the true massvalues. But rather consists of a curve in the space ( m X , m Y ) . How then are we to find the truemasses? Two methods suggest themselves. One is to note that this curve is nothing but thekink curve discussed above, and the location of the kink gives the true masses. A second is toobserve in Fig. 1 that the nature of the singularity qualitatively changes as one approaches thetrue value of m X . That is to say, the distribution becomes discontinuous.A second remark is that, in order to compute the singularity co-ordinate according tothe prescription given in [54], one needs a definition of what is the ‘nearest singularity to agiven event’. In the special case considered in the previous Section, where the observable phasespace is a disk with the singularities living on its boundary, it seems easy enough to make anunambiguous definition. But in the more general case considered in this Section, the observablephase space is a three-volume in ( p, q, q ) , whose singular boundary is the two-surface definedby the quartic p q − pq − p q ) = M . (In the case of massless daughter particles, thisdescribes a hyperbola.) It is now unclear, at least from a geometer’s viewpoint, what the nearestsingularity should be. 33 m
400 420 440 460 480 500 520 540 560 580 600 E ve n t s
10 HERWIG
Figure 1:
Simulation of the transverse mass variable for decay of a single gluino, with different curvescorresponding to different hypothesized values of the invsible daughter mass m X . The distribution has asingular endpoint for all values of m X , which translates into a singularity in the singularity co-ordinatefor all m X . Reproduced from [58]. Conclusions
The algebraic singularity method of [54] constitutes an elegant general method for measuringmasses, exploiting singularities that arise in projecting the smooth phase space of events in-volving invisible particles onto the observable phase space. To better understand the method,I applied it to the simple example of a single parent particle undergoing a point-like decay toan invisible daughter particle and a visible daughter particle. For this decay topology, it isknown that the kinematic properties are completely captured by the transverse mass variable,and I showed that the variable that arises from the algebraic singularity method is nothing but alinearized version of the usual transverse mass.The algorithm works by looking for a singularity at the origin in the distribution of a cer-tain co-ordinate, the so-called singularity co-ordinate. Such a singularity is shown to arise whenthe hypothesis for the masses is the correct one. I argued that the method suffers from the prob-lem that many mass hypotheses (including incorrect ones) give rise to such a singularity, evenin cases where general kinematic arguments tell us that it ought to be possible to determine themasses unambiguously. To do so using the singularity method seems to require supplementingthe algorithm with a means to identify the nature of a singularity, as well as its mere presence.I also discussed the issue of the definition of the nearest singularity, required to implementthe algorithm. This issue will need to be borne in mind when applying the method to morecomplicated decay topologies, such as cascade decays.
Acknowledgements
I thank A. J. Barr, I.-W. Kim and C. G. Lester for discussions.34 upersymmetry ontribution 3Light gluinos in hiding: reconstructing R-parity violatingdecays at the Tevatron
A.R. Raklev, G.P. Salam and J.G. Wacker
Abstract
If gluinos exist with masses less than 200 GeV, they are copiously pro-duced at the Tevatron, but still may not have been discovered if theydecay through baryon number violating operators. We show that usingcuts on jet substructure can enable a discovery with existing data andeven determine the gluino’s mass. Introduction
The TeV scale is the high energy frontier and New Physics (NP) is currently searched for ina multitude of channels at hadron colliders. Searches for NP with final state jets have yieldedbounds on new resonances decaying into pairs of jets through di-jet searches [59], or into jetsand missing energy [60, 61]. These resonance searches reach up to masses at the TeV scale forresonant production and 500 GeV for pair-produced particles. However, other NP possibilitiescan appear in exclusively hadronic final states, with no missing energy, that will not be discov-erable with the di-jet search or jets and missing energy searches. These NP possibilities fail tostand out over background because hadronic final states are challenging to calibrate and analyse.Even some relatively basic searches have not been performed at the Tevatron. For instance, [62]showed that a resonance that decays into two other new particles that subsequently decay to jetshas not been explicitly searched for, even though the backgrounds are manageable. In light ofthis dearth of searches, it is possible that NP candidates could escape detection at much lowermasses than 1 TeV, not because of small couplings to Standard Model (SM) particles and lowcross sections, but because of detector signatures that are sufficiently different from the searchesperformed so far at the Tevatron.This contribution considers one such possibility: a relatively light gluino, copiously pair-produced at the Tevatron, in a model with trilinear R-parity violating (RPV) operators thatviolate baryon number. As a result, the gluino decays into three quarks, ˜ g → qqq , and the eventcontains six final state partons. It is conceivable that these six partons will produce a six-jetsignature if the gluinos are produced near threshold, however, in the busy hadronic environmentof the Tevatron, with potential multiple jet overlaps and unreconstructed jets, it is not evidentthat this will show up over SM backgrounds in searches requiring high jet-multiplicity, e.g. the all-hadronic t ¯ t channel [63]. The reconstruction of a gluino mass peak is similarly veryhard due to the large combinatorics. Taken to its other extreme, the signature of gluino pair-production far above threshold, the six final state partons can merge into two back-to-back jets,each consisting of three collimated sub-jets from the individual final-state quarks. Here di-jetsearches along the same lines as [59] may be effective, but the non-resonant origin of the gluinopair makes discovery through simple di-jet invariant mass distributions very difficult.36he current limit on gluino masses in these scenarios comes from event shapes at LEP,which give a model independent gluino mass bound of 51.0 GeV [64]. In this contribution, wewill show how jet-substructure information for hard jets yields much better expected sensitivityto gluinos, and how a discovery may be lurking in Tevatron data. Our analysis will followa pattern similar to those suggested for reconstructing RPV neutralino decays to three quarksin [65], and benefits by lessons learned from the recent large interest in reconstructing thehadronic decays of other massive particle species, such as gauge bosons, top quarks and theHiggs [66–75]. The light gluino
Searches for supersymmetry necessarily make assumptions on the sparticle spectrum, and usu-ally it is assumed that color neutral particles (neutralinos, charginos, and sleptons) are lighterthan the colored particles (gluinos and squarks), however, this choice is motivated by top-downmodel building considerations. See [76, 77] for examples of models with gluinos as the light-est supersymmetric particle (LSP) and [78, 79] for studies without top-down prejudices. If acolored LSP decays relatively quickly it avoids the standard cosmological constraints on stablecharged particles.In this contribution, the gluino is assumed to be the lightest supersymmetric particle pro-duced in a collider and it decays via the R -parity violating superpotential operator L RPV = (cid:90) d θ λ (cid:48)(cid:48) ijk U ci D cj D ck + h.c. , (1)where λ (cid:48)(cid:48) ijk are flavor dependent coupling constants that are antisymmetric in the jk indices.It is likely that there are strong flavor hierarchies in these coupling constants which lessen theconstraints on the operator. For instance, a minimal flavor violation scenario gives λ (cid:48)(cid:48) ijk ∝ y / u i y / d j y / d k . If this is the case, the gluino decays dominantly to the heavy flavor combinationof ˜ g → cbs , assuming that it is beneath the top quark threshold. However, to be conservative, wewill not make any assumptions of heavy flavours in the final state, and we will use λ (cid:48)(cid:48) = 0 . ,leading to the gluino decay ˜ g → uds and its charge conjugate. The size of the coupling avoidsboth resonant single-sparticle production and displaced vertices/metastable charged sparticles,which have signatures that should be more easily detectable.A gluino LSP that decays via the baryon number violating RPV operator completelychanges the search strategy for supersymmetry at a hadron collider. Pair-production of lightgluinos has a huge cross section, as shown in Fig. 1. If the gluino is light and undiscovered,the rest of the susy spectrum is possibly not much heavier and we give a rough outline of thegeneric phenomenology here.Any squarks produced decay quickly into the gluino, giving rise to short cascades thatrarely contain secondary leptons or missing energy, even if charginos are kinematically accessi-ble in the decay chain. The best chance for a “spectacular” leptonic event at the Tevatron is thenif the direct production of charginos and neutralinos is kinematically accessible, and they decayto a lighter slepton. The slepton will in turn decay via a prompt four-body decay, ˜ (cid:96) → (cid:96) ˜ g ¯ qq .Thus the entire event would be χχ (cid:48) → g (cid:96) j , where some of the (cid:96) may be neutrinos, givingrise to a modest amount of missing energy. A tri-lepton search might be effective with this chan-nel, however, most current searches place jet multiplicity cuts to reduce the t ¯ t background [80].The searches for tri-leptons with jet vetoes set limits of m χ ± > ∼ GeV. Given that a gluino37 igure 1:
Leading order gluino pair-production cross section versus mass at the Tevatron.
LSP gives a more challenging environment, this is an optimistic estimate of the Tevatron’s reachfor charginos and neutralinos in these scenarios.Associated production of a squark and gluino can be an important contribution to theboosted gluino spectrum because the gluino’s p T can arise through the decay of the squark;however, the final state decay products of ˜ g ˜ q → q ˜ g ˜ g are not more visible than gluino pair-production. The squark may also appear as a di-jet resonance, however, the production crosssections are much smaller. For instance, a spectrum with ( m ˜ g , m ˜ q ) = (51 GeV , GeV ) hasa production cross section of σ ˜ g ˜ q (cid:39) pb, beneath the 10 pb bound for a 400 GeV resonance inthe di-jet search [59]. Associated production could become an effective search channel for thesquark with the jet substructure methods described below. For concreteness, we consider a “worse case” scenario with only gluinos light, keeping all othersparticle masses at 1 TeV. Our conservative approach means that the discovery potential shouldbe independent of the RPV coupling flavour and value, and to first order only depend on thegluino mass. We simulate gluino pair-production with RPV decays at the √ s = 1 . TeVTevatron using the H
ERWIG
IMMY
ERWIG event generator includes spin correlations in the gluino decays. The leading-logarithmic partonshower approximation used in H
ERWIG has been shown to model jet substructure well in a widevariety of processes [87–92].The resulting events are interfaced to the F
AST J ET IVET [95] framework, with some minor modifications to cope with the simulation ofsparticles. Our background sample, consisting of QCD → events, t ¯ t , W +jet, Z +jet and W W/W Z/ZZ production is simulated with the same setup. To explore the ultimate reach ofthe Tevatron, we use statistics comparable to 10 fb − of integrated luminosity. A cut of | η | < . is imposed upon all jets as a realistic geometrical acceptance; however, no other detector effectsare included. The signal is isolated by searching for jets that contain all three quarks from the gluino decay.These jets are necessarily very hard; indeed, the distance ∆ R = (cid:112) ∆ φ + ∆ η between thedecay products of a massive particle with mass m and transverse momentum p T should be ∆ R > ∼ m/p T [68]. The sensitivity to gluinos is strongly dependent upon the jet-algorithm size38arameter, and the momentum cut and these parameters should be optimised for any specificsearch. This contribution uses values that have good sensitivity over a large gluino mass range,but no optimisation is performed. In the following we will use the k T jet-algorithm [96, 97] inthe inclusive mode, as currently used at the Tevatron [98], with the following choices for jetsize: R = 0 . , . , . , . .Gluino candidates are identified by requirements on jet-substructure of the hardest jets inthe event. For each merging i of sub-jets k and l in the jet clustering we define y i = min ( p T k , p T l ) m j ∆ R kl , (2)where p T k and p T l are the transverse momenta of the sub-jets and m j is the mass of the final jetafter all mergings. The expectation is that y and y , from the last two mergings of the jet, aredistributed very differently from ordinary QCD jets because of the three-parton structure of thegluino jet. This turns out to be the case, see Fig. 1 of [65] for the similar case of a three-quarkneutralino decay.Since we have pair-production of gluinos, there is a choice of whether to perform an in-clusive analysis searching for at least one gluino candidate jet in each event, or an exclusiveanalysis, reconstructing both gluinos. This is a balance between signal efficiency and SM back-ground rejection that should be optimized to maximize discovery potential. In order to triggerand collimate the decay products of the gluino, at least one high p T jet is required, however,this also implies another back-to-back high p T gluino. To arrive at as clean a signal sample aspossible we choose an exclusive analysis.The following cuts are used: i) we require two hard jets with p T > GeV, ii) both mustbe candidate gluino jets satisfying the substructure constraint y > . , and iii) their massesmust be within 20% of each other. The resulting jet mass spectrum is shown in Fig. 2 for agluino mass of m ˜ g = 150 GeV and different values of the jet-size parameter R . A gluino masspeak is clearly observable with a small background, consisting mostly of QCD events, whenthe jet-size R is large enough to contain the complete gluino jets. This is generically true forgluino masses up to around 200 GeV with our p T cut, where we start to loose containment ofthe gluino for the largest jet-size considered.There is some systematic bias in signal events, visible in Fig. 2, towards jet masses largerthan the nominal. This arises from the jet-algorithm sweeping up extra energy from initial stateradiation and the underlying event, however, it should be possible to calibrate this with knownparticle masses, e.g. the top quark, and to limit the bias with filtering [68] and related techniques[73, 75]. The final achievable precision on mass seems likely to be limited by statistics and thejet mass resolution of the experiment.In Fig. 3 we show the resulting signal significance, S/ √ B , as a function of gluino massfor all four jet-sizes. The significance is estimated by the number of events in a 40 GeV intervalaround the nominal mass, which corresponds to a semi-realistic experimental jet mass resolu-tion. This ignores the “looking-elsewhere” problem, but should serve as a first estimate of thepossible reach.We can see that even above the point where we start to loose containment of the gluinojet the signal stands out over the background. This is in part caused by the k T algorithm’sefficiency in sweeping up soft radiation somewhat outside of its “radius” R , and in part bythe jet substructure cut not requiring two significant structures, allowing partially reconstructed39 [GeV] j m0 50 100 150 200 250 300 − E ve n t s / G e V / f b R=0.5 [GeV] j m0 50 100 150 200 250 300 − E ve n t s / G e V / f b R=0.7 [GeV] j m0 50 100 150 200 250 300 − E ve n t s / G e V / f b R=1.0 [GeV] j m0 50 100 150 200 250 300 − E ve n t s / G e V / f b R=1.5
Figure 2:
Jet mass distribution after cuts for 10 fb − integrated luminosity (black with error bars). Alsoshown is contribution from signal events for m ˜ g = 150 GeV (red), QCD (black) and other high p T SMbackground events (
W, Z and t ¯ t ) (green). gluinos. The latter effect is visible in the mass tails for the signal distribution in Fig. 2, and alsoin the presence of vector bosons which should only have one significant structure. For low massgluinos this is clearly not desirable as it could obfuscate a signal, however, the vector bosonscan serve as a jet mass calibration tool along with the top.Figure 3 also shows the signal to background ratio as a function of the gluino mass.For light gluino masses the lowest jet-sizes allows discovery and large S/B for gluino massesfrom 70 GeV to 150 GeV, above which containment of the gluino is lost. For higher massesprogressively larger jet-sizes must be used. For gluino masses significantly above 200 GeV, therate for producing boosted gluinos is too low; the requirement of 10 signal events with gluinotransverse momentum of p T > . m ˜ g sets an upper limit to the reach of m ˜ g = 280 GeV with10 fb − of integrated luminosity. However, these heavier gluinos are sufficiently massive thattheir decay products are multiple hard jets at the Tevatron. Conclusions
This contribution has demonstrated that the Tevatron can discover new light colored particlesthat decay into complicated hadronic final states, by using events where these particles areproduced at high p T and their boosted decay products are collimated. The recently developedtechniques using jet substructure can effectively separate signal from background, allowing the40 [GeV] g~ m50 100 150 200 250 300 S i gn i f i ca n ce , R = 0.5 T k , R = 0.7 T k , R = 1.0 T k , R = 1.5 T k [GeV] g~ m50 100 150 200 250 300 S / B Figure 3:
Significance (left) and signal over background (right) as a function of gluino mass for variousvalues of the jet-size R . new particle to appear as a jet-mass resonance. This work is a proof-of-principle, but there issignificantly more work to be done. Neither of the Tevatron’s two detectors have calorimetrythat is as finely segmented as the LHC detectors, where most studies have been done so far, andthis may reduce the jet mass resolution and restrict the Tevatron reach. Some of this resolutionloss may be recovered by using tracking information, particularly at CDF [99].On the theoretical side, the optimal cuts and jet-size for a given mass have not beendetermined, nor have other jet algorithms such as Cambridge/Aachen been studied. Switchingto a more inclusive analysis may also improve sensitivity. The primary challenge is to keepsignal efficiency high due to the low number of gluinos in the high p T tail, e.g. with m ˜ g =150 GeV we see 18% of gluinos with p T > GeV reconstructed using R = 1 . . Therefore itmay be beneficial to search for one narrow gluino candidate jet with tight constraints recoilingagainst another wide jet-size gluino candidate jet. The application of loose substructure cuts tothe jets of the ordinary di-jet search is also interesting.If the squarks become light enough that associated squark-gluino production or squarkpair production becomes sizable, then the dominant source of boosted gluinos may come fromthese secondary processes. This is qualitatively similar to the models studied in [65] where thehadronically decaying LSP was being produced in cascade decays of squarks and gluinos. Acknowledgements
ARR thanks the Swedish Research Council (VR) for financial support through the Oskar KleinCentre. GPS acknowledges support from the French ANR under grant ANR-09-BLAN-0060.JGW is supported by the US DOE under contract number DE-AC02-76SF00515 and receivespartial support from the Stanford Institute for Theoretical Physics and the US DOE OutstandingJunior Investigator Award. 41 ontribution 4SUSY-QCD corrections to MSSM Higgs boson production viagluon fusion
M. Mühlleitner, H. Rzehak and M. Spira
Abstract
In the MSSM scalar h, H production is mediated by heavy quark andsquark loops. The higher order QCD corrections have been obtainedsome time ago and turned out to be large. The full SUSY-QCD correc-tions have been obtained recently including the full mass dependence ofthe loop particles. We describe our calculation and present first numer-ical results. We also address the question of the proper treatment of thelarge gluino mass limit, i.e. the consistent decoupling of heavy gluinoeffects, and present the effective Lagrangian for decoupled gluinos. Introduction
One of the major goals at the LHC is the detection of Higgs boson(s) [100–104]. In the MinimalSupersymmetric Extension of the Standard Model (MSSM) two complex Higgs doublets are in-troduced to give masses to up- and down-type fermions [105–112]. After electroweak symmetrybreaking there are five physical Higgs states, two CP-even neutral Higgs bosons h, H , one neu-tral CP-odd Higgs state A and two charged Higgs bosons H ± . At tree level, the Higgs sectorcan be parameterized by two independent parameters, the pseudoscalar Higgs boson mass M A and the ratio of the two vacuum expectation values (VEV) of the two complex Higgs doublets, tan β = v /v . The Higgs couplings to quarks and gauge bosons are modified with sin and cos of the mixing angles α and β with respect to the Standard Model (SM) couplings, where α de-notes the h, H mixing angle. The bottom (top) Yukawa couplings are enhanced (suppressed) forlarge values of tan β , so that top Yukawa couplings play a dominant role at small and moderatevalues of tan β .At the LHC and Tevatron neutral Higgs bosons are copiously produced via gluon fusion gg → h, H, A , which is mediated in the case of h, H by (s)top and (s)bottom loops [113–115].The pure QCD corrections to the (s)quark loops have been obtained including the full Higgs and(s)quark mass dependencies and increase the cross sections by ∼ % [116–122]. This resultcan be approximated by very heavy top (s)quarks with ∼ − % accuracy for tan β < ∼ [123]. In this limit the next-to-leading order (NLO) QCD [124–128] and later the next-to-next-to-leading order (NNLO) QCD corrections [129–133] have been obtained, the latter leadingto a moderate increase of 20-30%. Finite top mass effects at NNLO have been discussed in[134–138]. Finally, the estimate of the next-to-next-to-next-to-leading order effects [139–142]indicates improved perturbative convergence. The full supersymmetric (SUSY) QCD correc-tions have been obtained in the limit of heavy SUSY particle masses [143–147] and more re-cently including the full mass dependence [148]. The electroweak loop effects have been cal-culated in [149–152]. In this article we will describe in Section 2 the calculation of the full42 g Q h, H gg ˜ Q h, H gg ˜ Q h, H
Figure 1:
Diagrams contributing to gg → h, H at leading order. SUSY-QCD corrections in gluon fusion to h, H , and we will present for the first time numericalresults. In Section 3 we will discuss the consistent derivation of the effective Lagrangian for thescalar Higgs couplings to gluons after the gluino decoupling. Gluon fusion
At leading order (LO) the gluon fusion processes gg → h/H are mediated by heavy quarkand squark triangle loops, cf. Fig.1, the latter contributing significantly for squark masses < ∼ GeV. The LO cross section in the narrow-width approximation can be obtained from the h/H gluonic decay widths, [113–115, 153] σ LO ( pp → h/H ) = σ h/H τ h/H d L gg dτ h/H (1) σ h/H = π M h/H Γ LO ( h/H → gg ) σ h/H = G F α s ( µ R )288 √ π (cid:12)(cid:12)(cid:12)(cid:12)(cid:12)(cid:12)(cid:88) Q g h/HQ A h/HQ ( τ Q ) + (cid:88) (cid:101) Q g h/H (cid:101) Q A h/H (cid:101) Q ( τ (cid:101) Q ) (cid:12)(cid:12)(cid:12)(cid:12)(cid:12)(cid:12) , (2)where τ h/H = M h/H /s with s being the squared hadronic c.m. energy and τ Q/ ˜ Q = 4 m Q/ ˜ Q /M h/H .The LO form factors are given by A h/HQ ( τ ) = 32 τ [1 + (1 − τ ) f ( τ )] A h/H ˜ Q ( τ ) = − τ [1 − τ f ( τ )] (3) f ( τ ) = arcsin √ τ τ ≥ − (cid:20) log 1 + √ − τ − √ − τ − iπ (cid:21) τ < . And the gluon luminosity at the factorization scale µ F is defined as d L gg dτ = (cid:90) τ dxx g ( x, µ F ) g ( τ /x, µ F ) , where g ( x, µ F ) denotes the gluon parton density of the proton. The NLO SUSY-QCD correc-tions consist of the virtual two-loop corrections, cf. Fig.2, and the real corrections due to theradiation processes gg → gh/H, gq → qh/H and q ¯ q → gh/H , cf. Fig.3. The final result for43 /H ˜ t, ˜ bgg g h/H ˜ t, ˜ bgg ˜ g h/H ˜ t, ˜ bgg g Figure 2:
Some generic diagrams for the virtual NLO SUSY-QCD corrections to the gluonic Higgscouplings. the total hadronic cross sections can be split accordingly into five parts, σ ( pp → h/H + X ) = σ h/H (cid:104) C h/H α s π (cid:105) τ h/H d L gg dτ h/H + ∆ σ h/Hgg + ∆ σ h/Hgq + ∆ σ h/Hq ¯ q . (4)The strong coupling constant is renormalized in the MS scheme, with the top quark, gluinoand squark contributions decoupled from the scale dependence. The quark and squark massesare renormalized on-shell. The parton densities are defined in the MS scheme with five activeflavors, i.e. the top quark, the gluino and the squarks are not included in the factorizationscale dependence. After renormalization we are left with collinear divergences in the sum ofthe virtual and real corrections which are absorbed in the renormalization of the parton densityfunctions, so that the result Eq. (4) is finite and depends on the renormalization and factorizationscales µ R and µ F , respectively. The natural scale choices turn out to be µ R = µ F ∼ M h/H .The numerical results are presented for the modified small α eff scenario [154], defined by the gg g ˜ t, ˜ b h, H gq q ˜ t, ˜ b h, H ¯ qq g ˜ t, ˜ bh, H Figure 3:
Typical diagrams for the real NLO QCD corrections to the squark contributions to the gluonfusion processes. following choices of MSSM parameters [ m t = 172 . GeV], M ˜ Q = 800 GeV tan β = 30 M ˜ g = 1000 GeV µ = 2 TeV M = 500 GeV A b = A t = − . TeV . (5)In this scenario the squark masses amount to m ˜ t = 679 GeV m ˜ t = 935 GeV m ˜ b = 601 GeV m ˜ b = 961 GeV . (6)Fig. 4 displays the genuine SUSY-QCD corrections normalized to the LO bottom quark formfactor, i.e. A h/Hb ( τ b ) → A h/Hb ( τ b )(1 + C bSUSY α s π ) . The corrections can be sizeable, but can bedescribed reasonably with the usual ∆ b approximation [155, 156], if A b is renormalized in theMS scheme. 44 b approximationreal partimaginary partC b SUSY (gg " H)small eff tg $ = 30 M H [ GeV ] -40-20020406080100120140 100 150 200 250 300 350 400 450 500 Preliminary
Figure 4:
The genuine SUSY-QCD corrections normalized to the LO bottom quark form factor. Realcorrections: red (light gray), virtual corrections: blue (dark gray), compared to the ∆ b approximation(dashed lines). A b has been renormalized in the MS scheme. Decoupling of the gluinos
In this section we will address the limit of heavy quark, squark and gluino masses, where inaddition the gluinos are much heavier than the quarks and squarks. For the derivation of theeffective Lagrangian for the scalar Higgs couplings to gluons we analyze the relation betweenthe quark Yukawa coupling λ Q and the Higgs coupling to squarks λ ˜ Q in the limit of large gluinomasses. We define these couplings at leading order in the case of vanishing mixing, λ Q = g H Q m Q v , λ ˜ Q = 2 g H Q m Q v = κλ Q , with κ = 2 vg H Q , (7)where g H Q denotes the normalization factor of the MSSM Higgs couplings to quark pairs withrespect to the SM. In the following we will sketch how the modified relation between thesecouplings for scales below the gluino mass M ˜ g is derived. For details, see Ref. [157]. We startwith the unbroken relation between the running MS couplings of Eq. (7) and the correspondingrenormalization group equations (RGE) for scales above M ˜ g . If the scales decrease below M ˜ g the gluino decouples from the RGEs leading to modified RGEs which are different for the twocouplings λ ˜ Q and κλ Q so that the two couplings deviate for scales below M ˜ g . The propermatching at the gluino mass scale yields a finite threshold contribution for the evolution fromthe gluino mass scale to smaller scales, while the logarithmic structure of the matching relationis given by the solution of the RGEs below M ˜ g . In order to decouple consistently the gluinofrom the RGE for gluino mass scales large compared to the chosen renormalization scale, amomentum substraction of the gluino contribution for vanishing momentum transfer has to beperformed [158]. We refer the reader to [157] for details and give here directly the result forthe modified relation between the quark Yukawa coupling and the effective Higgs coupling tosquarks taking into account the proper gluino decoupling: g H Q m Q v = ¯ λ ˜ Q,MO ( m ˜ Q ) (cid:40) C F α s π (cid:32) log M g m Q + 32 log m Q m Q + 12 (cid:33)(cid:41) , (8)where m Q is the pole mass and M O denotes the momentum substracted coupling, which istaken at the squark mass scale, which is the proper scale choice of the effective Higgs coupling45o squarks and which is relevant for an additional large gap between the quark and squarkmasses.Taking into account the radiative corrections to the relation between the effective cou-plings after decoupling the gluinos leads to the following effective Lagrangian in the limit ofheavy squarks and quarks, L eff = α s π G aµν G aµν H v (cid:88) Q g H Q (cid:20) α s π (cid:21) + (cid:88) ˜ Q g H ˜ Q (cid:104) C SQCD α s π (cid:105) + O ( α s ) , (9)where g H ˜ Q = v ¯ λ ˜ Q,MO ( m ˜ Q ) /m Q . The cofficient C SQCD is given by C SQCD = 376 . (10)It is well-defined in the limit of large gluino masses and thus fulfills the constraint of theAppelquist–Carazzone decoupling theorem [159]. Conclusions
We have presented first results for the NLO SUSY-QCD corrections to gluon fusion into CP-even MSSM Higgs bosons, including the full mass dependence of the loop particles. The gen-uine SUSY-QCD corrections can be sizeable. We furthermore demonstrated that the gluino con-tributions can be decoupled in the large M ˜ g limit in accordance with the Appelquist–Carazzonetheorem. Acknowledgements
We thank the organizers of the 2009 Les Houches workshop for the organization of this veryinteresting and fruitful workshop, in which to participate is always a big pleasure.46 ontribution 5Discriminating SUSY models at the LHC: a case study ofgauge-Higgs unification versus mSUGRA
S. Fichet and S. Kraml
Abstract
We investigate whether a sparticle spectrum arising from supersymmet-ric gauge-Higgs unification (SGHU) can be discriminated against theminimal supergravity (mSUGRA) model by LHC measurements. Tothis aim we assume that a realistic part of the mass spectrum has beenmeasured with a reasonable accuracy and perform Markov Chain MonteCarlo fits of the two models, SGHU and mSUGRA, to the expecteddata. Introduction
The model of supersymmetric gauge-Higgs unification (SGHU) we published recently in [160]features light selectrons and smuons, which are systematically lighter than the second-lightestneutralino ˜ χ . Same-flavour opposite-sign (SFOS) dileptons stemming from ˜ χ → (cid:96) ± ˜ (cid:96) ∓ → (cid:96) ± (cid:96) ∓ ˜ χ (with (cid:96) = e, µ ) in cascade decays of squarks and gluinos are hence expected to have alarge rate in this model.The SFOS dilepton signature arising from on-shell decays of ˜ χ to sleptons is also promi-nant in the minimal supergravity (mSUGRA) model with small M [161]. Indeed, most bench-mark studies are performed within the mSUGRA model, see e.g. [24, 162, 163]. Top-down fitsof the model to expected LHC measurements look quite promising; as shown in [164] they are,however, largely dominated by the gaugino and slepton masses (or mass differences).In this contribution, we investigate whether SGHU can be discriminated against mSUGRAbased on LHC measurements. To this aim, we perform a case study for the SGHU point Dof [160]. We assume that a realistic part of the mass spectrum has been measured with a rea-sonable accuracy, and perform Markov Chain Monte Carlo (MCMC) fits of the two models tothe expected measurements. These measurements, although not sufficient to do a Lagrangianreconstruction, may permit to exclude models of supersymmetry breaking, or to conclude thatone model is more likely than another from a Bayesian point of view. If it is not the case,the posterior distributions may help identify additional observables with better discriminatingpower.In general, depending on the measurements available, there are two ways of comparingthe agreement between models and data. If the models are overconstrained by measurements,one can simply compare the maxima of their likelihoods. On the other hand, if the models areunderconstrained, continous sets of points reach the maximum of likelihood, and it becomes For details on the MCMC method see, e.g., [165–168] and references therein. MCMC fits of mSUGRAparameters to expected LHC data (at mSUGRA benchmark points) were recently done in [163, 169]. Setup of the analysis
We use a modified version of SUSPECT2.41 [170] as spectrum generator, and MICROMEGAS[171, 172] for computing additional observables. While there exist specialized fitting tools likeSFITTER [173] and FITTINO [174], these are not directly applicable to the SGHU model forvarious reasons. We have therefore programed our own MCMC analysis with a Metropolisalgorithm, largely following the procedure of [168]. Below we explain some details which arespecific to our analysis.
As reference scenario we use the SGHU point D of [160]. The (s)particle masses that areaccessible to LHC measurements are shown in Table 1. Since at present no experimental simu-lation is available for this scenario, we simply assume that the sparticle masses can be extractedfrom invariant-mass distributions (following, e.g., [29, 30]) with 3% accuracy. This is in agree-ment with the discussions in the “Spins and Masses” subgroup at this Workshop. We keep theinput m t used in [160] and assume that it will be measured to 1 GeV accuracy at the LHC.The error on the top quark mass feeds into a parametric uncertainty on m h ; we therefore take m t = 172 . ± GeV and m h = 117 . ± GeV in our fits, assuming that other theoreticaluncertainties on m h will be under control by the time these measurements become available.Finally, we consider two cases: hypothesis H0 without measurement of the heavy Higgs sectorand hypothesis H1 with measurement of the heavy Higgs sector. Throughout the analysis, wedemand that the ˜ χ is the lightest SUSY particle, LSP.Some more comments are in order. First, m ˜ q in Table 1 is the average mass of the 1st and2nd generation squarks, m ˜ q = ( m ˜ u L + m ˜ u R + m ˜ d L + m ˜ d R ) . Second, a priori we cannot knowthe chirality of the slepton in the ˜ χ → (cid:96) ± ˜ (cid:96) ∓ → (cid:96) ± (cid:96) ∓ ˜ χ decay chain: the extracted slepton mass m ˜ (cid:96) ≡ m ˜ e = m ˜ µ is the mass of either the left- or the right-chiral selectron/smuon, dependingon the mass ordering with respect to ˜ χ . If m ˜ (cid:96) L < m ˜ χ the wino-like ˜ χ decays mainly into (cid:96) ˜ (cid:96) L even if m ˜ (cid:96) R < m ˜ (cid:96) L , and it is m ˜ (cid:96) L that is measured. This is in fact the case at our referencepoint, which has m ˜ (cid:96) R = 217 GeV and m ˜ (cid:96) L = 327 GeV. If, however, m ˜ (cid:96) R < m ˜ χ < m ˜ (cid:96) L , then ˜ χ → (cid:96) ˜ (cid:96) R , and what is measured is m ˜ (cid:96) R . This is typically the case in mSUGRA. In the MCMCscans we therefore take m ˜ (cid:96) = 326 . ± . GeV as being m ˜ (cid:96) L or m ˜ (cid:96) R depending on the massordering at a particular parameter point. Third, we note that at point D both staus are heavierthan the ˜ χ and hence do not appear in the decay chains. This is neglected in this simple study;in a more sophisticated analysis, however, one should take the absence of a τ + τ − edge intoaccount.We do not include constraints from B-physics observables nor the dark matter relic densityin the fit, but use them only a posteriori . The nominal values at point D are BR ( b → sγ ) = ˜ χ m ˜ (cid:96) m ˜ χ m ˜ q m ˜ g m h m H . ± . . ± . . ±
12 1022 ± . ± . . ± . ± . Table 1:
Masses (in GeV) accessible to LHC measurements at the SGHU point D, and assumed ex-perimental errors. We consider two cases: case H0 without measurement of m H , and case H1 withmeasurement of m H . . × − , BR ( B s → µ + µ − ) = 5 . × − , and Ω h = 0 . . The familiar mSUGRA model depends on four continuous parameters — tan β , the universalgaugino mass M / , the universal scalar mass parameter M and the universal trilinear coupling A (the latter three being input at M GUT )— and the sign of µ .The SGHU model also depends on tan β , M / and sign( µ ). The boundary conditions forthe Higgs and scalar sectors are, however, considerably different from the mSUGRA case. Firstof all, the Higgs soft terms are fixed by the SGHU relation m H , = (cid:15) H Bµ − | µ | (1)at M GUT , with (cid:15) H = ± ; this is computed iteratively in our modified SUSPECT version [160].Moreover, the soft terms of the first and second generation sfermions vanish at M GUT , whilethose of the third generation are non-zero and non-universal. In the full model developedin [160], the third generation soft terms depend on the GUT-scale Yukawa couplings and twobulk mixing angles, and are computed in our modified SUSPECT version using an additionallevel of iteration. This procedure being very time consuming, we do not consider the completemodel here, but simply let the third-generation scalar soft-terms vary independently. The costof this is a larger number of free parameters, which will have repercussions on the Bayes factor,as explained in Section 2.4. On the other hand, this approach is less dependent on the modelbuilding of the matter sector.The parameters to be fitted to the data are hence: mSUGRA : tan β, M / , M , A (2) SGHU : tan β, M / , M Q , M U , M D , A t , A b , M L , M E , A τ (3)We take µ > throughout, and (cid:15) H = − in the SGHU case. Generally, both signs of µ and allsign combinations of µ and (cid:15) H should be investigated, but this is not possible here because ofCPU limitations. The choice of (cid:15) H = − is, however, justified because, as we will see, in themSUGRA case we find large negative A , dominated by the effect of A t . In the SGHU case, weknow from [160] that only one sign combination of (cid:15) H and A t gives acceptable phenomenology.An important difference between mSUGRA and SGHU lies in the gaugino and sleptonmass ratios. The gaugino masses are determined by M / in both models. The slepton masses,however, are driven by M in the mSUGRA case, while in the SGHU case they are drivenby M / and the U(1) Y D-term contribution from the S parameter, S = ( m H − m H ) +Tr( m Q − m U + m D + m R − m L ) . Roughly, m ˜ χ ≈ . M / , m ˜ χ ≈ . M / , m e R ≈ M + (0 . M / ) − . S GUT , and m e L ≈ M + (0 . M / ) + 0 . S GUT , where S GUT isthe value of S at M GUT . Note that S GUT ≡ in mSUGRA, while M ≡ in SGHU. From this49e can already estimate M / ≈ GeV in both models, M ≈ GeV in the mSUGRAcase, and S GUT ≈ − (280 GeV) in the SGHU case. Moreover, from these considerations weexpect the mass ordering m ˜ (cid:96) R < m ˜ χ < m ˜ (cid:96) L in mSUGRA, but m ˜ (cid:96) R < m ˜ (cid:96) L < m ˜ χ in SGHU.Another important difference lies in the higgsino and heavy Higgs masses. Since µ ∼ GeV at point D, the higgsino states are not accessible at LHC. The heavy Higgs masses,however, are around
GeV, which might be within reach. In order to test the discriminatingpower of the heavy Higgs sector, we perform fits without and with including a measurement ofone of the heavy Higgs masses. We here use the mass of H , but taking instead m A or m H ± iscompletely equivalent.Regarding parameter ranges, since m ˜ g (cid:39) GeV, we vary M / in [0 , GeV only.In SGHU, the scalar mass parameters are allowed to vary within [0 , GeV. The A termsare allowed to vary within specific ranges, which contain the parameter space of the full SGHUmodel: A t = [ − , GeV, A b = [ − , GeV, A τ = [ − , GeV. In themSUGRA case, the scalar masses and A ≡ A t = A b = A τ are allowed to vary withoutbounds. We do not constrain tan β . A posteriori, it does not exceed due to theoreticalconstraints from tachions and color or charge breaking. Last but not least, we use flat priors forall model parameters. For a discussion of prior (in)dependence in the presence of LHC data,see [169]. In the likelihood function, all measurements are taken into account as gaussians proportionalto exp (cid:0) − ( x th − x exp ) /σ exp (cid:1) . Here x exp and σ exp are the nominal value and assumed exper-imental error as given in Table 1, and x th is the prediction at a given parameter point. Theglobal level of convergence of the Markov chains is evaluated using the procedure describedin [165]. For the parameters which give the maximum likelihood in each case, we evaluate the68% and 95% Bayesian Credibility (BC) intervals, using the full likelihood. If the maximumlikelihood is constrained by gaussian measurements, these correspond to the usual σ and σ confidence intervals. We also evaluate the 68% and 95% BC regions from the 2D marginalizeddistributions. The Bayes factor is defined as the ratio of the posterior probability of two models given a set ofdata: K = P ( M | data ) /P ( M | data ) . (4)Assuming that both models have the same global probability, P ( M ) = P ( M ) , to describereality, this ratio is reduced to the ratio of global likelihoods: K = P ( data |M ) /P ( data |M ) .There is, however, a subtlety: assuming that a set M of data is measured implies thatthe discovery D is already done: data = M ∩ D . This implies that K = P ( M |M ∩ D ) P ( D |M ) /P ( M |M ∩ D ) P ( D |M ) . Here P ( D |M ) is the probability to make a dis-covery assuming the model M , i.e. the potential of discovery of M . For a supersymmetricmodel at the LHC, we can consider this is roughly equal to P ( M / < − TeV ) . In theparticular case we study, as we compare two supersymmetric models, this ratio cancels. Thelikelihood P ( M |M (cid:84) D ) becomes equal to (cid:82) L ( M, θ i ) P ( θ i ∩ D ) dθ i where the θ i are the pa-rameters of the model. By taking flat internal priors on the parameters, this reduces to theintegral of the likelihood over the volume of the parameter space V D allowing the discovery:50 ( M |M ∩ D ) = (cid:82) V D L ( M, θ j ) dθ j . Outside of this volume, the likelihood must be consideredas null. In our case, the Bayes factor is therefore simply reduced to the ratio of the two averagelikelihoods, computed on the discovery volume: K = (cid:104)L (cid:105) / (cid:104)L (cid:105) = N (cid:88) i =1 L ( x ( i )1 ) / N (cid:88) i =1 L ( x ( i )2 ) . (5)where the sums are over the points of Markov Chains. For two models to be discriminated, theBayes factor should be at least around 3 (30) to constitute a weak (strong) evidence. A Bayesfactor larger than 100 is considered as a decisive evidence.It is important to note that the Bayes factor favorizes models with small number of pa-rameters. This implies that the SGHU model with independant scalar soft terms we considerhere should be less favored than the complete one with only two mixing angles. A detaileddiscussion of the Bayes factor can be found in, e.g., [175]. Results
In this section, we present the results of MCMC scans which collected around points foreach case, i.e. for each of the two hypotheses in the two models. Figure 1 shows 1D and 2Dmarginalized likelihoods for the mSUGRA and SGHU model parameters under the H0 hypoth-esis (no measurement of heavy Higgses). The marginalized likelihoods for the H1 hypothesis(assumed measurement of m H ) are shown in Fig. 2. In both figures, the 2D marginalized likeli-hoods are plotted as isolines corresponding to 68% and 95% BC regions. The colored 2D mapscorrespond to the empirical averages of the sampled likelihoods. They have only indicativevalue, to show what the zones of high likelihood are, independent of the volume effect which istaken into account in the true marginalization. We recall that the 68% (95%) BC intervals aredefined by the hypersurface enclosing 68% (95%) of the integral likelihood around the maxi-mum. When this limit is identical to the boundary of the scan, this means that the distributionis too flat to give a prefered value with 68% (95%) credibility.We see from Figs. 1 and 2 that in the mSUGRA case M / , A and M and in the SGHUcase M / and A t are well constrained, but the other parameters are not. We also note a con-siderable tightening of the correlations between tan β , M / and A t ) when information on theheavy Higgs sector is added. In particular, a measurement of m H very much constrains tan β in the mSUGRA case, with the fitted value being in fact quite close to the “true” one, see Fig. 2.In the SGHU case, on the other hand, tan β is much less constrained.The values of maximal and averaged likelihoods and convergence parameter r are givenin Table 2. In the H0 hypothesis, both models fit the data very well without preference for theone or the other, the maximum likelihoods as well as the Bayes factor being close to one. This isin fact only little different in the H1 hypothesis: the mSUGRA fit still gives a high L max (cid:39) . ,and the Bayes factor is of order 2, i.e. not sufficiently large to constitute an evidence. In order toseparate the effect of the “pure” SGHU condition eq. (1) from that of the non-universal sfermionsoft terms, we also performed a fit for a SGHU model variant with universal M and A for allthree generations (in other words, mSUGRA supplemented by eq. (1)). In this case, we find L max = 0 . and (cid:104)L(cid:105) = 0 . in the H1 hypothesis, that means a Bayes factor of K ≈ with respect to strict mSUGRA, and K ≈ w.r.t. SGHU with 10 free parameters. So the smallpreference of SGHU over mSUGRA in the H1 case comes indeed from the degeneracy of theHiggs soft terms, m = m = m at M GUT ( m , = m H , + | µ | , m = | Bµ | ).51 igure 1: Marginalized likelihood distributions in 1D and 2D for the mSUGRA (orange) and SGHU(red) models in the H0 hypothesis. In the mSUGRA case, A ≡ A t and M ≡ M Q . The plots onthe diagonal show the 1D likelihoods of both models, normalized to have the same maximum. The off-diagonal plots show iso-contours of 68% and 95% BC, computed within the 2D marginalized likelihood.The upper triangle of 2D plots is the SGHU case, while the lower triangle is the mSUGRA case. Thecolor maps indicate the empirically averaged likelihoods. The axes of the 2D plots are shown on theouter boundary of the figure. The green lines/stars indicate the nominal values of point D. The 68% and 95% Bayesian credibility intervals (BCIs) for the model parameters aregiven explicitly in Table 3 for mSUGRA and in Table 4 for SGHU. For comparison, the inputvalues at point D are: tan β = 30 , M / = 500 GeV, M Q = 614 GeV, M U = 635 GeV, M D = 414 GeV, A t = − GeV, A b = − GeV, M L = 408 GeV, M E = 433 GeV, A τ = − GeV.We next ask whether indirect observables can help discriminate the two models. To thisaim, Fig. 3 shows the 1D marginalized distributions for Br ( b → sγ ) , Br ( B s → µ + µ − ) , andthe neutralino relic density Ω h as obtained from the mSUGRA and SGHU fits. The 68% and95% BCIs are given explicitly in Table 5. We see that the B-physics observables have a gooddiscriminating power in case the heavy Higgs sector is known (H1 hypothesis), but not so in theH0 hypothesis. Regarding the relic density, we note that the mSUGRA model predicts a muchtoo large Ω h ∼ . – . at 68% BC if the heavy Higgs sector is unconstrained. In the H1 case,52 igure 2: Same as Fig. 1 but for the H1 hypothesis. when m H (and hence m A and tan β ) are fixed, then the Ω h prediction within mSUGRA alsogives smaller values in agreement with WMAP observations. This is different for the SGHUmodel, for which Ω h peaks towards values smaller than ∼ . . However, the distribution israther flat and when considering the 68% or 95% BCIs, no definite conclusion can be obtained,see Table 5. L max (cid:104)L(cid:105) r mSUGRA H0 0.984 0.200 1.0037mSUGRA H1 0.742 0.080 1.0058 L max (cid:104)L(cid:105) r SGHU H0 0.995 0.221 1.0064SGHU H1 0.995 0.166 1.0065
Table 2:
Values of the maximum and averaged likelihoods, and of the convergence parameter r . Obviously, improving the model discrimination requires the measurement of additionalparts of the mass spectrum. To this end, we show in Fig. 4 the 1D marginalized likelihooddistributions for some predicted masses, in particular the masses of ˜ e R , ˜ τ , ˜ t , and ˜ χ ± . Asexpected, a very good discrimination would be obtained by measuring the ˜ e R mass (note that theposterior distributions for m ˜ e R do not overlap). Measurement of the ˜ τ and/or ˜ t masses wouldhelp reveal the non-universality of the scalar soft terms. A powerful test in particular of theSGHU condition eq. (1)) would be the determination of the µ parameter through a measurement53SUGRA H0 mSUGRA H168% BCI 95% BCI 68% BCI 95% BCI tan β [9, 27] [6, 36] [29, 35] [25,37] M / [495, 515] [485, 525] [496, 516] [487, 526] M [252, 280] [239, 292] [252, 280] [239, 292] A [ − − − − − Table 3:
68% and 95% Bayesian credibility intervals (BCIs) for the mSUGRA parameters in the H0 andH1 hypotheses.
SGHU H0 SGHU H168% BCI 95% BCI 68% BCI 95% BCI tan β [4, 43] [4, 57] [13,42] [8,56] M / [493, 512] [484, 520] [494, 512] [485,521] M Q [1, 1341] [1, 1837] [0, 1093] [0, 1689] M U [3, 1413] [3, 1766] [2, 1257] [2, 1626] A t [ − − − − − − − − Table 4:
68% and 95% BCIs for SGHU parameters in the H0 and H1 hypotheses. The limits for M D ,L ,E are very similar to those for M Q ,U . There are no reasonable limits for A b,τ . Br ( b → sγ ) × log ( Br ( B s → µ − µ + )) Ω h mSUGRA H0 [2 . , . , [2 . , .
02] [ − . , − . , [ − . , − .
0] [0 . , . , [0 . , . mSUGRA H1 [2 . , . , [2 . , .
75] [ − . , − . , [ − . , − .
0] [0 . , . , [0 . , . SGHU H0 [2 . , . , [2 . , .
55] [ − . , − . , [ − . , − .
5] [0 . , . , [0 . , . SGHU H1 [2 . , . , [2 . , .
39] [ − . , − . , [ − . , − .
0] [0 . , . , [0 . , . Table 5:
68% and 95% BCIs intervals of predicted indirect observables: Br ( b → sγ ) , Br ( B s → µ + µ − ) , and Ω h . of the higgsino sector: the distributions for m ˜ χ ± hardly overlap in the H0 case and do not overlapat all in the H1 case. All this may best be done at an e + e − linear collider with high enoughcentre-of-mass energy. Nevertheless, at the LHC a first hint for a non-universal structure maybe obtained from the absence of a kinematic endpoint in the τ + τ − invariant-mass distribution,since in the mSUGRA case we typically have m ˜ τ < m ˜ e R < m ˜ χ < m ˜ e L . Indeed, in themSUGRA fit, ˜ χ → τ ± ˜ τ ∓ typically has about 80–90% branching ratio, followed by ˜ χ → h ˜ χ as the next-important channel, while ˜ χ → e ± ˜ e ∓ R often has a branching ratio below 1%.Before concluding, we recall that in the complete SGHU model in [160], where the thirdgeneration soft terms are computed from two bulk mixing angles, M Q , M U , M D , A t and A b are not independent of each other. Therefore the SGHU distributions in Fig. 4 will be a bitnarrower in the complete model than in the more general version presented here. Conclusions
We investigated whether a sparticle spectrum arising from SGHU can be discriminated againstthe mSUGRA model by LHC measurements. To this end we performed MCMC fits of the twomodels to assumed LHC data for a particular SGHU benchmark point, which is characterizedby GUT-scale degenerate Higgs mass parameters and non-universal third-generation soft terms.54 igure 3:
Marginalized likelihood distributions in 1D for indirect observables predicted in the mSUGRA(orange) and SGHU (red) models; the upper row of plots is for the H0, and the lower row for the H1hypothesis. The green lines indicate the nominal values at the reference point D.
Figure 4:
Marginalized likelihood distributions in 1D for some predicted masses for mSUGRA (orange)and SGHU (red) models in the H0 (upper row) and H1 (lower row) hypotheses. The green lines indicatethe nominal values at the reference point D.
It turned out that the mSUGRA model can fit the anticipated LHC data well; a measure-ment of the ˜ χ , ˜ χ , ˜ e , ˜ g and h masses (with percent-level precision) is not sufficient to discrim-inate the structure of the underlying model. Also the Bayes factor does not allow to favour theSGHU model over mSUGRA. This does not change significantly if information on the heavyHiggs sector is included. However, information on the heavy Higgs sector in combination withimproved B-physics constraints would significantly influence the fits.A decisive model discrimination would be possible through a measurement of the ˜ e R massin e + e − collisions (together with refined measurements of the rest of the spectrum). Besides, ameasurement of the higgsino mass should provide a test of the SGHU condition m = m = m at M GUT . Accurate measurements of the sparticle spectrum in e + e − should also allow to55etermine the neutralino relic density with good precision.Last but not least we note that our analysis is based on assumed LHC measurements ofabsolute masses. It should be possible to improve the fits by including more information, e.g. the positions of kinematic endpoints and event rates. Moreover, a lower limit on the ˜ τ massfrom the absence of a τ + τ − signal would considerably impact the results obtained here. Howwell this can be done should be subject to further investigation. Acknowledgements
We are grateful to Ritesh K. Singh for inspiring discussions about Markov Chains, which trig-gered this analysis. We also thank Michael Rauch for comparisons of the mSUGRA case withSFITTER. 56 ontribution 6MCMC Analysis of the MSSM with arbitrary CP phases
G. Belanger, S. Kraml, A. Pukhov and R.K. Singh
Abstract
We explore the parameter space of the MSSM with explicit CP-violating(CPV) phases by means of a Markov Chain Monte Carlo analysis, im-posing constraints from direct Higgs and SUSY searches at colliders, B-physics, EDM measurements, and the relic density of dark matter. Wefind that over most of the parameter space, large phases are compatiblewith experimental data. We present likelihood maps of the CPV-MSSM,concentrating in particular on quantities relevant for the neutralino relicdensity. Introduction
It was noted early on [176,177] that a neutralino LSP in the MSSM with conserved R-parity is anexcellent cold dark matter candidate. Detailed studies showed that in the MSSM, or constrainedversions thereof, there are several mechanisms that provide the correct rate of neutralino anni-hilation, such that Ω h (cid:39) . : annihilation of a bino LSP into fermion pairs through t -channelsfermion exchange in case of very light sparticles; annihilation of a mixed bino-higgsino orbino-wino LSP into gauge boson pairs through t -channel chargino and neutralino exchange,and into top-quark pairs through s -channel Z exchange; and finally annihilation near a Higgsresonance (the so-called Higgs funnel). Furthermore, coannihilation processes with sparticlesthat are close in mass with the LSP may bring Ω h in the desired range. This way, the measuredrelic density of dark matter is often used to severely constrain the MSSM parameter space.In [168], some of us explored the parameter space of the phenomenological MSSM thatis allowed when requiring that the neutralino LSP constitutes all the dark matter by means of aMarkov Chain Monte Carlo (MCMC) scan.This was done for the case of seven free parameters,where it was assumed that there are no new sources of CP violation beyond the CKM. Herewe go a step further and perform a MCMC analysis of the MSSM parameter space allowing forarbitrary CP phases.The parameters that can have CP phases in the MSSM are the gaugino and higgsino massparameters and the trilinear sfermion-Higgs couplings. Although constrained by electric dipolemoments (EDMs), nonzero phases can significantly influence the phenomenology of SUSYparticles, see e.g. [179] and references therein. They can also have a strong impact on theHiggs sector, inducing scalar-pseudoscalar mixing through loop effects [180–182]. Moreover,CP phases can have a potentially dramatic effect on the relic density of the neutralino [183–187]. This is true not only in the Higgs funnel region: since the couplings of the LSP to othersparticles depend on the phases, so will all the annihilation and coannihilation cross sections, An analogous analysis of the phenomenological MSSM with 25 free parameters was performed in [178]employing a MultiNest algorithm. Setup of the MCMC scans
Table 1 lists the free parameters of the CPV-MSSM together with ranges within which theyare allowed to vary in our scan. We take common masses for the first and second generation ofsfermions to avoid FCNC constraints and assume universality of the gaugino masses at the GUTscale as motivated in the context of models defined at the GUT scale. The trilinear couplings ofthe first and second generation are taken to be zero. For the third generation, mass parametersand trilinear soft terms are treated as independent parameters. In addition, we allow for arbitraryphases of all the gaugino mass parameters and trilinear couplings of the third generation. Thehiggsino mass parameter µ , on the other hand, is taken to be real. This can be done without lossof generality because the physically relevant phases are arg( M i µ ) and arg( A f µ ) .For the numerical analysis, we use micrOMEGAs2.2 [172, 190] linked to CPsuperH2 [191]. The latter gives the CPV Higgs sector, B-physics observables and EDMs. We use thethallium, mercury and electron EDMs d ( T l ) , d ( Hg ) and d ( e − ) ; the neutron EDM is not usedbecause of its big uncertainty stemming from the quark model [191]. To evaluate the limits onthe light Higgs mass, we make use of the HiggsBounds [192] program. For the scan we use thedirected random search MCMC method as described in detail in Ref. [168] (see also referencestherein).We compute the likelihood of a parameter point as the product of likelihoods of all the ob-servables under consideration. The observables considered in our analysis are listed in Table 2along with the shapes of the likelihood functions used. These probability distribution funtions(PDFs) are given as: G ( x, x , σ x ) = exp (cid:20) − ( x − x ) σ x (cid:21) , F ( x, x , σ x ) = 11 + exp[ − ( x − x ) /σ x ] . (1)We use the Gaussian function G for observables for which a measurement is available, andfunction F when there is only an upper or lower bound. Last but not least, we use flat priors forall input parameters, and base the analysis on ten chains with points each. Results
Figure 1 shows the 1D posterior PDFs for some of the most important model parameters like | M | , µ , m H + , tan β , M l , M q . (Here and in the following, dimensionful parameters are in GeV.)Some explanatory comments are in order. First, we observe a slight preference for positive µ ,at the level of 40% minus versus 60% plus sign. A priori this seems in agreement with thepreference of sign( M µ ) = +1 found in [178] caused by the b → sγ constraint (we do not useany constraint on the muon ( g − ). In our case it is, however, mostly due to the fact that wehave six chains that converged in the µ > subspace but only four in the µ < one. Either way,58ymbol stands for General range m H ± mass of H ± [100 , GeV tan β tan β [2 . , µ µ parameter [ − , GeV A t Trilinear stop coupling [0 , GeV Φ t Phase of A t [0 , π ] A b Trilinear sbottom coupling [0 , GeV Φ b Phase of A b [0 , π ] M Gaugino mass, M = M = M / , GeV Φ Phase of M [0 , π ]Φ Phase of M [ − π, π ]Φ Phase of M [0 , π ] M l Common slepton mass for first two generations [500 , GeV M l Mass of left stau [100 , GeV M r Mass of right stau [100 , GeV M q Common squark mass for first two generations [500 , GeV M Q Mass of left stop–sbottom doublet [100 , GeV M u Mass of right stop [100 , GeV M d Mass of right sbottom [100 , GeV m t Top quark mass . ± . GeV [189]Symbol stands for General range Φ µ Phase of µ parameter or π for ± ve value of µA l Trilinear coupling of 1st& 2nd gen. sleptons GeV A q Trilinear coupling of 1st& 2nd gen. squarks GeV
Table 1:
Model parameters and their ranges used in the scan. the preference of one sign over the other is not significant. Second, the heavy Higgs sector ispushed to masses above ca. 500 GeV by B-physics constraints, while EDM constraints push themasses of the first and second generation sfermions to the multi-TeV range. Third, regarding tan β , we observe a preference for small values, caused again by EDM constraints.Correlations between the input parameters can be seen in Fig. 2, which shows the 2D 68%and 95% Bayesian Credibility (BC) regions in the ( µ, | M | ) , ( m H + , | M | ) , (tan β, m H + ) and (Φ , tan β ) planes. CP-conserving (CPC) analogs of the first two plots can be seen in Fig. 3 ofRef. [168]. The correlations between | M | – µ and | M | – m H + are dominantly driven by the relicdensity constraint. The CPV and CPC cases show the same basic features, favouring the mixedbino-higgsino ( | M | ≈ µ ) or the Higgs-funnel regions | M | ≈ M H + / . It is, however, apparentthat allowing for nonzero phases considerably enlarges the parameter space that is compatiblewith a relic density within WMAP bounds. For example, the 68% BC range includes a regionfar from the Higgs funnel where | M | ≈ µ ≈ O (1) TeV and the ˜ χ ± and ˜ χ have a small massdifference with the LSP. This region occurs with much smaller likelihood in the CPC case [168].The impact of the EDM constraints on tan β is apparent from the fourth panel in Fig. 2: when Φ is nonzero, tan β is constrained to very small values, while the large values of tan β are59bservable Limit Likelihood function Ref. Ω h . ± . G ( x, . , . [193] BR( b → sγ ) (3 . ± . × − G ( x, . × − , . × − ) [194, 195] A CP ( b → sγ ) (1 . ± . × − G ( x, . × − , . × − ) [2] BR( B s → µ + µ − ) ≤ . × − F ( x, . × − , − . × − ) [196] R ( B u → τ ν τ ) 1 . ± . G ( x, . , . [194] BR( B d → τ + τ − ) ≤ . × − F ( x, . × − , − . × − ) [2] d ( T l ) e cm ≤ . × − F ( x, . × − , − . × − ) [197] d ( Hg ) e cm ≤ . × − F ( x, . × − , − . × − ) [198] d ( e − ) e cm ≤ . × − F ( x, . × − , − . × − ) [197] R H (Higgs mass) ≤ . F ( x, . , . [192]Mass limits LEP limits or − [199] Table 2:
Observables used in the likelihood calculation.
Figure 1:
1D posterior PDFs for some model parameters; from left to right: | M | , µ , m H + (top row)and tan β , M l , M q (bottom row). Dimensionful parameters are in GeV. allowed only for very small values of Φ .One advantage of the MCMC is that it lets us explore the constraints on the phases in ageneral way, by marginalization over parameters. As expected, we find that the phase that ismost constrained by the EDMs is the relative phase between M and µ . Since we take µ to bereal without loss of generality, this means severe constraints on Φ , as illustrated in Fig. 3. Theother phases are much less constrained. In particular the phases of M and of the trilinear softterms can vary over the full range, Φ ,t,b = [0 , π ] , if the sfermions of the first two generationshave masses of few TeV. Only for Φ there is also some preference for the near-CPC case.Overall, with five phases to vary, the CPC case becomes a point in a 5D parameter space.This has important consequences for the EDMs, since they will be near zero only when all thedominant phases go to zero simultaneously. This means that the EDMs dominantly saturatethe present bounds: they are predicted to be large and potentially observable over most ofthe allowed parameter space. This is illustrated in Fig. 4, which shows the 2D marginalizeddistributions of EDMs at 68% and 95% BC. We see that (i) the EDMs are highly correlated and(ii) the CPC case is just a small corner of the large parameter space we are considering.60 igure 2: Regions of 68% BC (dark grey) and 95% BC (light grey) in the plane µ vs. | M | , m H + vs. | M | , tan β vs. m H + and tan β vs. Φ . Figure 3:
Regions of 68% BC (dark grey) and 95% BC (light grey) of the mercury EDM versus phases(in degrees).
Figure 4:
Regions of 68% BC (dark grey) and 95% BC (light grey) showing the correlations betweenthe EDMs.
Let us now turn to two key quantities determining the dominant annihilation channel of theneutralino LSP: the distance from the (mostly pseudoscalar) Higgs pole, δm H ≡ m h − m ˜ χ ,and the relative mass difference between the lightest and second-lightest neutralino, ∆ χ ≡ ( m ˜ χ − m ˜ χ ) /m ˜ χ . In the CPC case with gaugino mass universality, the latter quantity is a directmeasure of the higgsino fraction of the LSP. The 2D likelihood functions for δm H versus Φ i (with i = 1 , , , t ) are shown in Fig. 5. The analogous distributions for ∆ χ are shown in Fig. 6.We see that for nonzero phases the preferred values of both δm H and ∆ χ can considerably differfrom those in the CPC case. This was already noted in [187] and is confirmed here in a moregeneral way.Finally, Table 3 explicitly lists the 68% and 95% BC intervals for CPV-MSSM parameters,Higgs and sparticle masses, and several low-energy observables. Note, for instance, that thesquark and slepton masses of the first two generations are above 2 (4) TeV at 95% (68%) BC.61 igure 5: Regions of 68% BC (dark grey) and 95% BC (light grey) showing the correlation betweendistance from the h pole ( m h − m ˜ χ ) and the various phases (in degrees). Figure 6:
Regions of 68% BC (dark grey) and 95% BC (light grey) showing the correlation between therelative ˜ χ – ˜ χ mass difference and the various phases (in degrees). The third generation can be much lighter, with a 95% (68%) lower limit of 300–400 GeV(around 800 GeV) for the lighter mass eigenstates ˜ t , ˜ b , and ˜ τ . Neutralinos and charginoscover a large mass range, from about 100 GeV up to ca. 1 TeV. This also holds for the LSP.In turn, the gluino can be rather light, leading to a large pair production cross section at theLHC followed dominantly by decays into third generation quarks—or very heavy, beyond thereach of the LHC. Gaugino–higgsino mixing is sizable over a large part of the parameter space;whether this can lead to observable rates of electroweak ˜ χ ˜ χ ± production at the LHC depends,however, on the neutralino/chargino mass scale, which as said above spans a wide range. Allthese issues will be considered in detail elsewhere [200]. Conclusions
We have presented a first Bayesian analysis of the CPV-MSSM model with parameters definedat the electroweak scale, taking into account constraints from collider searches, B-physics andEDMs and requiring that the neutralino LSP be the dark matter of the Universe with a relicdensity in agreement with WMAP observations. We find that phases can be large if the firsttwo generations of sfermions are above 2 (4) TeV at 95% (68%) BC. In fact only one phase, Φ , is strongly constrained. A large fraction of the parameter space features heavy sparticlesthat are beyond the reach of the LHC. This is just a reflection of the fact that, apart from Ω h ,other measurements do not require a supersymmetric contribution. Clearly improvements onthe experimental determination of the EDMs will play a crucial role in revealing or further con-straining phases. The implications of the phases for LHC phenomenology as well as for darkmatter direct and indirect detection will be presented in an expanded and updated version [200]of this analysis. 62arameters, masses, 68% BCI 68% BCI 95% BCI 95% BCI Remarksobservables min max min max M H +
704 1999 473 1999
Upper limit saturated tan β .
82 17 . .
10 27 . | M |
235 1002 113 1002
Upper limit saturated µ −
750 881 − | A t | Upper limit saturated Φ − . . − .
17 23 . M l Upper limit saturated M q Upper limit saturated M l Upper limit saturated M r Upper limit saturated M q M u M d
989 5008 389 5008
Upper limit saturated m h
114 120 114 123
Lower limit saturated m h
700 1997 466 1998
Upper limit saturated m h
700 1998 466 1998
Upper limit saturated m ˜ χ
70 841 70 935 m ˜ χ
107 925 107 1440
Lower limit saturated m ˜ χ
139 996 139 1758
Lower limit saturated m ˜ χ ±
104 923 104 1440
Lower limit saturated m ˜ χ ±
232 1789 232 1989
Lower limit saturated m ˜ e L,R
Upper limit saturated m ˜ τ
868 3236 335 4297 m ˜ τ Upper limit saturated m ˜ q L,R
Upper limit saturated m ˜ t
867 2444 463 3471 m ˜ t Upper limit saturated m ˜ b
815 2782 378 3973 m ˜ b Upper limit saturated Ω h . . . . Post-dictionBR ( B → X s γ ) × .
44 3 .
78 3 .
23 4 . Post-diction A CP [ B → X s γ ](%) − . . − . . Post-diction R ( B u → τ ν ) 0 . . . . Post-dictionBR ( B s → µµ ) × .
63 3 .
72 3 .
31 4 . Pre-dictionBR ( B d → τ τ ) × .
25 2 .
30 2 .
04 2 . Pre-diction log | d T l | − . − . − . − . | d Hg | − . − . − . − . | d e − | − . − . − . − . R Higgs . . . . c.f. HiggsBounds f H . . . . LSP higgsino fraction m h − m χ −
527 775 − m ˜ χ /m ˜ χ − .
008 1 .
005 0 .
008 1 . Peaks at both ends
Table 3:
The min/max limits of the 68% and 95% BC intervals (BCI) for CPV-MSSM parameters, Higgsand sparticle masses, and various observables. ote added: On completion of this work, we became aware of an improved limit on the mer-cury EDM of d ( Hg ) < . × − e cm [201]. This new limit leads to stronger constraints onthe parameter space, especially on Φ , and will be included in the more detailed report. Acknoledgements
We thank Oliver Brein for discussions on
HiggsBounds .64 trong EWSB ontribution 7Composite Higgs boson search at the LHC
J.R. Espinosa, C. Grojean and M. Mühlleitner
Abstract
In composite Higgs models the Higgs boson emerges as a pseudo-Goldstoneboson from a strongly-interacting sector. While in the Standard Modelthe Higgs sector is uniquely determined by the mass of the Higgs bo-son, in composite Higgs models additional parameters control the Higgsproperties. In consequence the LEP and Tevatron exclusion bounds aremodified and the Higgs boson searches at the LHC are significantlyaffected. The consequences for the LHC Higgs boson search in thecomposite model will be discussed. Introduction
The massive nature of the weak gauge bosons
W, Z requires new degrees of freedom and/ornew dynamics around the TeV scale to ensure unitarity in the scattering of longitudinal gaugebosons W L , Z L . In the Standard Model (SM) unitarity is assured by the introduction of anelementary Higgs boson. The SM Higgs couplings are proportional to the mass of the particleto which it couples, and the only unknown parameter in the SM is the mass of the Higgs boson.Furthermore, the electroweak precision observables and the absence of large flavor-changingneutral currents strongly constrain departures from this minimal Higgs mechanism and rathercall for smooth deformations, at least at low energy (see Ref. [202] for a general discussion).This supports the idea of a light Higgs boson emerging as a pseudo-Goldstone boson from astrongly-coupled sector, the so-called Strongly Interacting Light Higgs (SILH) scenario [203,204]. The low-energy content is identical to the SM with a light, narrow Higgs-like scalar,which appears, however, as a bound state from some strong dynamics [205–211]. A mass gapseparates the Higgs boson from the other usual resonances of the strong sector as a result of theGoldstone nature of the Higgs boson. Since the rates for production and decay, however, candiffer significantly from the SM results we study in the present work how the LHC Higgs bosonsearch channels are affected by the modifications of the composite Higgs boson couplings. Weestimate the experimental sensitivities in the main search channels studied by ATLAS and CMSas well as the luminosities needed for discovery.The effective Lagrangian constructed in [203], which involves higher dimensional oper-ators for the low-energy degrees of freedom, should be seen as an expansion in ξ = ( v/f ) where v = 1 / (cid:112) √ G F ≈ GeV and f is the typical scale of the Goldstone bosons of thestrong sector. It can therefore be used in the vicinity of the SM limit ( ξ → ), whereas thetechnicolor limit ( ξ → ) requires a resummation of the full series in ξ . Explicit models, builtin 5D warped models, provide concrete examples of such a resummation. Here, we will rely ontwo representative 5D models exhibiting different behaviours of the Higgs couplings. In thesemodels the deviations from the SM Higgs couplings are controlled by the parameter ξ = ( v/f ) Parameterization of the Higgs couplings
The effective SILH Lagrangian involves two classes of higher dimensional operators: (i) opera-tors genuinely sensitive to the new strong force, which will affect qualitatively the Higgs bosonphysics and (ii) operators sensitive to the spectrum of the resonances only, which will simplyact as form factors. The effective Lagrangian generically takes the form L SILH = c H f (cid:0) ∂ µ | H | (cid:1) + c T f (cid:16) H † ←→ D µ H (cid:17) − c λf | H | + (cid:18) c y y f f | H | ¯ f L Hf R + h . c . (cid:19) + ic W g m ρ (cid:16) H † σ i ←→ D µ H (cid:17) ( D ν W µν ) i + ic B g (cid:48) m ρ (cid:16) H † ←→ D µ H (cid:17) ( ∂ ν B µν ) + . . . (1)where g, g (cid:48) denote the SM electroweak (EW) gauge couplings, λ the SM Higgs quartic couplingand y f the SM Yukawa coupling to the fermions f L,R . The coefficients c H , c T , ... appearing inEq. 1 are expected to be of order 1 unless protected by some symmetry. The operator c H givesa corrections to the Higgs kinetic term. After rescaling the Higgs field, in order to bring thekinetic term back to its canonical form, the Yukawa interactions read (see Ref. [203] for details) g ξhf ¯ f = g SM hf ¯ f × (1 − ( c y + c H / ξ ) , (2) g ξhV V = g SM hV V × (1 − c H ξ/ , g ξhhV V = g SM hhV V × (1 − c H ξ/ , (3)where V = W, Z , g SM hf ¯ f = m f /v , g ξhW + W − = gM W , g ξhZZ = (cid:112) g + g (cid:48) M Z , g SM hhW + W − = g and g SM hhZZ = ( g + g (cid:48) ) and m f , M W , M Z denote the fermion, W and Z boson masses. Thedominant corrections controlled by the strong operators preserve the Lorentz structure of theSM interactions, while the form factor operators will also introduce couplings with a differentLorentz structure.For our two concrete models studied hereafter we refer to the Holographic Higgs modelsof Refs. [212–214], which are based on a five-dimensional theory in Anti-de-Sitter (AdS) space-time. The bulk gauge symmetry SO (5) × U (1) X × SU (3) is broken down to the SM gaugegroup on the UV boundary and to SO (4) × U (1) X × SU (3) on the IR. In the unitary gauge thisleads to the following Higgs couplings to the gauge fields ( V = W, Z ) in terms of the parameter ξ = ( v/f ) g hV V = g SMhV V (cid:112) − ξ , g hhV V = g SMhhV V (1 − ξ ) . (4)The Higgs couplings to the fermions will depend on the way the SM fermions are embeddedinto representations of the bulk symmetry. In the MCHM4 model [213] with SM fermions67ransforming as spinorial representations of SO (5) , the Higgs fermion interactions are given byMCHM4: g hff = g SMhff (cid:112) − ξ . (5)In the MCHM5 model [214] with SM fermions transforming as fundamental representations of SO (5) , the Higgs fermion couplings take the formMCHM5: g hff = g SMhff − ξ √ − ξ . (6)While the Higgs gauge couplings are always reduced compared to the SM, the Higgs couplingsto fermions behave differently in the two models. In the vicinity of the SM the couplings arereduced, with the reduction being more important for the MCHM5 than for the MCHM4 model.For larger values of ξ , the MCHM5 Higgs fermion couplings raise again and can even becomelarger than in the SM, leading to enhanced gluon fusion Higgs production cross sections. Thelatter will significantly affect the Higgs searches. Constraints from LEP and Tevatron and EW precision data
The ( M H , ξ ) parameter region is constrained from the Higgs searches at LEP and Tevatron.The excluded regions are shown in Fig. 1. For the generation of the plots the program Higgs-Bounds [192] has been used, modified to take into account the latest Tevatron limits.In both composite models the SM Higgs mass LEP limit M H > ∼ . GeV is lowered,since at LEP the most relevant search channel is Higgs-strahlung with subsequent decay into b ¯ b [215, 216]. In both models the production process is suppressed compared to the SM. Sincein MCHM5 at ξ = 0 . the Higgs fermion coupling vanishes, this channel cannot be used inthe area around this ξ value. Constraints are set by Higgs-strahlung production with subsequentdecay into γγ instead [217].At Tevatron, low ξ values are excluded by the Higgs decay into a W pair for Higgs massesaround 160 GeV . The exclusion region quickly shrinks to 0, since the relevant Higgs-strahlungproduction is suppressed compared to the SM for non-vanishing ξ values. In MCHM5, anadditional region M H ∼ − GeV can be excluded for ξ > ∼ . through H → W W [218]where the enhanced Yukawa coupling increases the production in gluon fusion and the
W W branching ratio is still high, before fermionic decays take over close to ξ = 1 . The exclusionis then set by H → τ τ decays [219]. These results should be regarded, however, as roughguidelines. The Tevatron searches combine several search channels from both experiments in asophisticated way. We cannot perform such an analysis at the same level of sophistication.Further constraints arise from the electroweak precision (EWP) data. The oblique param-eters are logarithmically sensitive to the Higgs boson mass [220]. The EWP limits are alsoshown in Fig. 1. In our set-up they are due to the incomplete cancellation between the Higgsand gauge boson contributions to S and T and low ξ values are preferred. The upper bound on ξ is relaxed by a factor of ∼ if one allows for a partial cancellation of the order of 50%. Branching ratios
The partial widths in the composite Higgs models are obtained by rescaling the correspondingHiggs couplings involved in the decay. In the MCHM4 model all couplings are multiplied by Tevatron searches in H → W W decays exclude the SM Higgs boson in the mass range 162 GeV ≤ M H ≤
166 GeV [218]. igure 1: Experimental limits from Higgs searches at LEP (blue/dark gray) and the Tevatron (green/lightgray) in the plane ( M H , ξ ) for MCHM4 (left) and MCHM5 (right). EW precision data prefer low valueof ξ : the red continuous line delineates the region favoured at 99% CL (with a cutoff scale fixed at2.5 TeV) while the region below the red dashed line survives if there is an additional 50% cancellationof the oblique parameters. the same factor √ − ξ so that the branching ratios are the same as in the SM. In the MCHM5model the partial decay width into fermions can be obtained from the corresponding SM widthby, cf. Eq. (6), Γ( H → f ¯ f ) = (1 − ξ ) (1 − ξ ) Γ SM ( H → f ¯ f ) . (7)The Higgs decay width into gluons, mediated by heavy quark loops, reads Γ( H → gg ) = (1 − ξ ) (1 − ξ ) Γ SM ( H → gg ) . (8)The Higgs decay width into massive gauge bosons V = W, Z is given by Γ( H → V V ) = (1 − ξ ) Γ SM ( H → V V ) . (9)The Higgs decay into photons proceeds dominantly via W -boson and top and bottom loops.Since the couplings to gauge bosons and fermions scale differently in MCHM5, the various loopcontributions have to be multiplied with the corresponding Higgs coupling modification factor.As QCD corrections do not involve the Higgs couplings the higher order QCD corrections tothe decays are unaffected and can readily be taken over from the SM.Fig. 2 shows the SM branching ratios and the composite Higgs branching ratios of MCHM5for three representative values of ξ = 0 . , . , . in the mass range favoured by compositeHiggs models between 80 and 200 GeV, which has not been completely excluded by the LEPbounds yet (see Section 3). The branching ratios have been obtained with the program HDE-CAY [221], where the modifications due to the composite nature of the Higgs boson have beenimplemented. For ξ = 0 . the behaviour is almost the same as in the SM, with the Higgs below ∼ M Z decaying dominantly into b ¯ b and a pair of massive gauge bosons, one or two of thembeing virtual. Above the gauge boson threshold, it almost exclusively decays into W W, ZZ .The decays into γγ and Zγ are slightly enhanced compared to the SM though, behaviour whichculminates at ξ = 0 . . Here, due to the specific Higgs fermion coupling in MCHM5, see69 h [ GeV ] -1 -2 -3 - ! + ! - ggcc - ZZWW "" Z " BR(H)SM
80 100 150 200M h [ GeV ] -1 -2 -3 - ! + ! - ggcc - ZZWW "" Z " BR(H)MCHM5 =0.2
80 100 150 200M h [ GeV ] -1 -2 -3 !! Z ! BR(H)MCHM5 " =0.5
80 100 150 200M h [ GeV ] -1 -2 -3 - ! + ! - ggcc - ZZWW "" Z " BR(H)MCHM5 =0.8
Figure 2:
Higgs branching ratios as a function of the Higgs boson mass in the SM ( ξ = 0 , upper left)and MCHM5 with ξ = 0 . (upper right), 0.5 (bottom left) and 0.8 (bottom right). Eq. (6), the decays into fermions and fermion-loop mediated decays into gluons are closed andthe branching ratio into γγ dominates in the low Higgs mass region. This cannot be exploitedfor the LHC searches, however, which rely on this search channel in the low mass region, sincethe gluon fusion production is absent for the same reason and the vector boson fusion process issuppressed by a factor two compared to the SM. At ξ = 0 . the branching ratios into fermionsdominate at low-Higgs mass and are enhanced compared to the SM above the gauge bosonthreshold, which is due to the enhancement factor in the Higgs fermion coupling, while theHiggs couplings to massive gauge bosons are suppressed. LHC production cross sections
The Higgs boson search channels at the LHC can be significantly changed in composite Higgsmodels due to the modified production cross sections and branching ratios. The main char-acteristics of the production cross sections shall be presented here. At the LHC the relevantproduction channels are
Gluon fusion:
The gluon fusion process gg → H [153] constitutes the dominant pro-duction mechanism in the SM. At leading order it is mediated by heavy quark loops. Thenext-to-leading order QCD corrections [118], which enhance the cross section by 50-100%, donot involve Higgs couplings and thus are unaffected by the composite nature of the Higgs bosonin our specific parameterization. The NLO gluon fusion cross section in the composite modelcan hence be obtained from the SM by σ NLO ( gg → H ) = (1 − ξ ) σ SMNLO ( gg → H ) MCHM4 σ NLO ( gg → H ) = (1 − ξ ) (1 − ξ ) σ SMNLO ( gg → H ) MCHM5 . (10) W/Z boson fusion:
Weak boson fusion qq → qq + W ∗ W ∗ /Z ∗ Z ∗ → qqH [222–224] is the70ext important SM Higgs production process. Due to the additional forward jets, which allowfor a strong background reduction, it plays an important role for the Higgs boson search. NLOQCD corrections [113, 225], accounting for a 10% correction, are unaffected by the modifiedcomposite Higgs couplings, so that for our models it is given by σ NLO ( qqH ) = (1 − ξ ) σ SMNLO ( qqH ) for MCHM4 and MCHM5 . (11) Higgs-strahlung:
In the intermediate mass range M H < ∼ M Z Higgs-strahlung off
W, Z bosons q ¯ q → Z ∗ /W ∗ → H + Z/W provides another production mechanism [226, 227]. Thecross section including NLO QCD corrections, which add ∼ % in the SM [113,228], is givenby σ NLO ( V H ) = (1 − ξ ) σ SMNLO ( V H ) for MCHM4 and MCHM5 . (12) Higgs radiation off top quarks:
This production mechanism [229–233] only plays a rolefor Higgs masses < ∼ GeV. NLO QCD corrections increase the cross section at the LHC by ∼ % [234–236], and in the composite Higgs models studied here it is given by σ NLO ( Ht ¯ t ) = (1 − ξ ) σ SMNLO ( Ht ¯ t ) MCHM4 σ NLO ( Ht ¯ t ) = (1 − ξ ) (1 − ξ ) σ SMNLO ( Ht ¯ t ) MCHM5 . (13)While being excluded as discovery channel due to the large background and related uncertain-ties, in MCHM5 it may provide an interesting search channel for large values of ξ near one dueto a significant enhancement factor.Fig. 3 shows the production cross sections as function of M H = 80 ... GeV in the SMand MCHM5 for ξ = 0 . , . and . . For ξ = 0 . the inclusive cross section is considerablyreduced due to reduced couplings in the production cross sections, situation which is even worsefor ξ = 0 . where the gluon fusion and Ht ¯ t cross sections vanish and the others are reduced.For ξ = 0 . the situation is reversed due to the significantly enlarged gluon fusion process. Thecross sections for MCHM4 are not shown separately. They can be obtained from the SM onesby multiplying each with − ξ . Statistical significances
In order to study how the Higgs prospects of discovery will be changed in composite models, weevaluated the statistical significances for the different search channels at the LHC. We referredto the CMS analyses [162]. The results presented hereafter are not significantly changed whenapplying the ATLAS analyses [163]. Assuming that only the signal rates are changed but not thebackgrounds rates, since only Higgs couplings are affected in our models, the significances inMCHM4 and MCHM5 can be obtained by applying a rescaling factor κ to the number of signalevents. Referring to a specific search channel, it is given by taking into account the change inthe production process p and in the subsequent decay into a final state X with respect to theSM, hence κ = σ p BR ( H → X ) σ SMp BR ( H SM → X ) . (14)The number of signal events s is obtained from the SM events s SM by s = κ · s SM , (15)71 H [ GeV ] -1 -2 ! (pp " H+X) [ pb ] s=14TeVM t =175 GeVCTEQ6MSMgg " Hqq " Hqqgg,qq - " Htt - qq - " HW qq - " HZ
80 100 150 200M H [ GeV ] -1 -2 ! (pp " H+X) [ pb ] s=14TeVM t =175 GeVCTEQ6MMCHM5 $ =0.2gg " Hqq " Hqqgg,qq - " Htt - qq - " HWqq - " HZ
80 100 150 200M H [ GeV ] -1 -2 ! (pp " H+X) [ pb ] s=14TeVM t =175 GeVCTEQ6MMCHM5 $ =0.5qq " Hqq qq - " HWqq - " HZ
80 100 150 200M H [ GeV ] -1 -2 ! (pp " H+X) [ pb ] s=14TeVM t =175 GeVCTEQ6MMCHM5 $ =0.8gg " H qq " Hqqgg,qq - " Htt - qq - " HWqq - " HZ Figure 3:
The LHC Higgs boson production cross-sections as a function of the Higgs boson mass in theSM ( ξ = 0 , upper left) and for MCHM5 with ξ = 0 . (upper right), 0.5 (bottom left) and 0.8 (bottomright). The cross sections include NLO QCD corrections and have been obtained by use of the programsHIGLU [237], VV2H [238], V2HV [238], HQQ [238]. where s SM after application of all cuts is taken from the experimental analyses. The signalevents s and the background events after cuts, i.e. b ≡ b SM , are used to calculate the corre-sponding significances in the composite Higgs model. The various channels studied are H → γγ : This channel is crucial for Higgs searches at low masses M H < ∼ GeV.Despite the clean signal, the channel is challenging due to small signal and large backgroundrates. The production is given by the inclusive cross section composed of gluon fusion, vectorboson fusion, Higgs-strahlung and Ht ¯ t production. H → ZZ → l l (cid:48) : The gold-plated channel for Higgs masses above ∼ GeV with theHiggs decaying through ZZ ( ∗ ) in the clean e, e µ and µ final states is based on gluon fusionand vector boson fusion in the production. Since the production cross section is large as wellas the branching ratio into ZZ ( ∗ ) it allows for a precise determination of the Higgs boson massand cross section. H → W W → l ν : Higgs decay into W W with subsequent decay in leptons is the maindiscovery channel in the intermediate region M W < ∼ M H < ∼ M Z . Spin correlations can beexploited to extract the signal from the background. The CMS analyses use gluon and vectorboson fusion to get the signal rates. H → W W → lνjj : Higgs production in vector boson fusion with subsequent decay H → W W → lνjj covers the mass region GeV < ∼ M H < ∼
180 GeV, where the H → ZZ ( ∗ ) branching ratio is largely suppressed. The event topology with two energetic forward jets andsuppressed hadronic activity in the central region can be exploited to extract the signal from thebackground. H → τ τ → l + j + E missT : This channel with the Higgs produced in vector boson fusion,72dds to the difficult Higgs search in the low mass region M H < ∼ GeV. The specific signatureof vector boson fusion production (see above) helps for the extraction of the signal. qqH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 SM120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 % =0.2MCHM4120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 % =0.5MCHM4120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! total significance H ! ZZ !
4l CMS30fb -1 $ =0.8MCHM4120 140 160 180 200M H [ GeV ] Figure 4:
The significances in different channels as a function of the Higgs boson mass in the SM ( ξ = 0 ,upper left) and for MCHM4 with ξ = 0 . (upper right), 0.5 (bottom left) and 0.8 (bottom right). For more details on each search channel and on the significance estimators we used werefer the reader to [239]. In Figs. 4 and 5 we present the SM significance (for comparison) andthe MCHM4 and MCHM5 significances for ξ = 0 . , . , . . The results should be understoodas estimates. They cannot replace experimental analyses. But they can serve as a guideline ofwhat is changed in composite models and where to be careful when it comes to interpretationof experimental results. As can be inferred from Figs. 4, in MCHM4 in all search channelsthe significance is always below the corresponding significance in the SM. With the branchingratios being unchanged, this is due to the production cross sections which are all suppressed bythe universal factor − ξ . The Higgs search will hence be much more difficult. For ξ = 0 . thetotal significance even drops below 5.In MCHM5 the behaviour of the significances is more involved due to the interplay ofmodified production and decay channels. For ξ = 0 . the reduction in production channelscannot be compensated by the enhancement in the branching ratios into γγ and massive gaugebosons, so that the significances are below the SM ones. In total the significance is also belowthe total MCHM4 significance, as gluon fusion production which contributes to the main searchchannels, is more strongly reduced in MCHM5. The situation looks even worse for ξ = 0 . where gluon fusion (and also Ht ¯ t production) is completely erased from the list of productionchannels. Only for low Higgs masses the strong enhancement in the γγ branching ratio can risethe significance above 5, even for M H below the LEP limit, although that has to be confirmedby detailed experimental analyses though. For higher Higgs masses one has to rely on weakboson fusion with H → W W decay. For ξ = 0 . the picture is totally different from MCHM4.73 qH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 SM120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 % =0.2MCHM5120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! H ! ZZ ! total significance CMS30fb -1 $ =0.5MCHM5120 140 160 180 200M H [ GeV ] qqH, H ! WW ! l " jjH ! WW ! " H ! total significance qqH, H !$$! l+jH ! ZZ !
4l CMS30fb -1 % =0.8MCHM5120 140 160 180 200M H [ GeV ] Figure 5:
The significances in different channels as a function of the Higgs boson mass in the SM ( ξ = 0 ,upper left) and for MCHM5 with ξ = 0 . (upper right), 0.5 (bottom left) and 0.8 (bottom right). The production is completely taken over by gluon fusion and leads to large significances inthe massive gauge boson final states. Also γγ final states contribute for M H > ∼ GeV, andprobably for M H > GeV, although this also has to be confirmed by experimental analysesthough. Conclusions
We have shown by focusing on two particular directions in the parameter space of the compositeHiggs model, that the search modes and significances can deviate significantly from the SMexpectations. In the MCHM4 model all couplings are reduced compared to the SM values andhence the Higgs searches deteriorate. In the MCHM5 model, however, the production in gluonfusion is enhanced if the composite scale is low enough. The significances can then be largerthan in the SM case. Once the Higgs boson will show up in the LHC experiments, the studyof the relative importance of the various production and decay channels will thus provide us toa certain extent with information on the dynamics of the Higgs sector and tell us whether theelectroweak symmetry breaking is weak or strong.74 ontribution 8Low-Scale Technicolor at the 10 TeV LHC
K. Black, T. Bose, E. Carrera, S. J. Harper, K. Lane, Y. Maravin, A. Martin and B.C. Smith
Abstract
This report summarizes Low-Scale Technicolor (LSTC) and the workdone by the LSTC-at-LHC group for Les Houches 2009. We studythe reach of the LHC with √ s = 10 TeV for the lightest ρ T , ω T , a T technivectors decaying to W Z , γW , γZ followed by leptonic decays ofthe weak bosons, and to e + e − . For the most part, we restrict ourselvesto luminosities of O (1 fb − ) . The revised LHC run schedule for2010–11 was established as this report was being completed. Introduction
Technicolor (TC) [240–243] was invented to provide a natural and consistent quantum-field-theoretic description of electroweak (EW) symmetry breaking — without elementary scalarfields. Extended technicolor (ETC) [244] was invented to complete that description by includ-ing quark and lepton flavors and their chiral symmetry breaking as interactions of fermions andgauge bosons alone. In particular, from Fig. 1, m q,(cid:96) (cid:39) g ET C (cid:104) ¯ T T (cid:105)
ET C /M ET C , where (cid:104) ¯ T T (cid:105)
ET C is the technifermion condensate renormalized at M ET C . From the beginning, ETC was rec-ognized to have a problem with flavor-changing neutral current interactions, especially thoseinducing K – ¯ K mixing. Masses M ET C of several 100, possibly 1000, TeV are required tosuppress these interactions to an acceptable level. The problem is that this implies m q,(cid:96) of atmost a few MeV if one assumes that, as in QCD, (1) asymptotic freedom sets in quickly abovethe TC scale of a few
100 GeV so that (cid:104) ¯ T T (cid:105)
ET C (cid:39) (cid:104) ¯ T T (cid:105)
T C and (2) (cid:104) ¯ T T (cid:105)
T C can be estimatedby scaling from the quark condensates of QCD. Walking technicolor [245–248] was inventedto cure this problem. The cure is that the QCD-based assumptions may not apply to technicolorafter all. In particular, in walking TC the gauge coupling decreases very slowly, staying large for100s, perhaps 1000s, of TeV and remaining near its critical value for spontaneous chiral sym-metry breaking. Then, the ¯ T T anomalous dimension γ m (cid:39) over this large energy range [249],so that (cid:104) ¯ T T (cid:105)
ET C (cid:29) (cid:104) ¯ T T (cid:105)
T C and reasonable fermion masses result. The important lesson ofwalking technicolor is that QCD-based assumptions for technicolor must, at best, be viewedwith suspicion and used with caution. In particular, all estimates of the precision electroweakparameter S for TC models [251–254] are based on scaling from QCD and, as such, are un-trustworthy [255, 256]. Lattice gauge-theoretic techniques appear to be a promising way to testthe ideas of walking technicolor in a nonperturbative way.A walking TC gauge coupling with γ m (cid:39) for a large energy range occurs if, as in Fig. 1,the critical coupling for chiral symmetry breaking lies just below a value at which the TC β -function vanishes (an infrared fixed point) [257, 258]. This requires a large number of tech- Except for the top quark, which needs an interaction such as topcolor to explain its large mass [250]. N D (cid:29) doublets in the fundamental representa-tion N T C of the TC gauge group, SU ( N T C ) , or by having a few doublets in higher-dimensionalrepresentations [259,260]. In the latter case, constraints on ETC representations [244] almost al-ways imply other technifermions in the fundamental representation as well. In either case, then,there generally are technifermions whose technipion ( π T ) bound states have a decay constant F (cid:28) F π = (246 GeV) . This low scale implies there are, in addition to the π T , technihadrons ρ T , ω T and a T with masses well below a TeV. We refer to this situation as low-scale technicolor(LSTC) [259, 261, 262]. These technivector mesons can be produced as s -channel resonancesin q ¯ q annihilation at the LHC. As we discuss next, they will be extremely narrow, with strikingsignatures visible above manageable backgrounds. ! ! ! ! ! " " ! " " $$ ! $% % % ! ! ! !" ! ! &’ " ! ! !" " Figure 1:
The quark and lepton mass generating mechanism in ETC (left). The β -function in walkingtechnicolor, with the chiral symmetry breaking value of α T C just below an approximate infrared fixedpoint (right).
There are two important consequences of this picture of walking TC. First, to restatewhat we just said, N D > technifermion doublets implies the existence of physical tech-nipions, some of which couple to the lightest technivector mesons. Second, since M π T ∝(cid:104) ¯ T T ¯ T T (cid:105)
ET C ≈ ( (cid:104) ¯ T T (cid:105)
ET C ) , walking TC enhances the masses of technipions much more thanit does other technihadron masses. Thus, it is very likely that the lightest M ρ T < M π T and thatthe two and three- π T decay channels of the light technivectors are closed [259]. This furtherimplies that these technivectors are very narrow, a few GeV or less, because their decay ratesare suppressed by phase space and/or small couplings (see below).A simple phenomenology of LSTC is provided by the Technicolor Straw-Man Model(TCSM) [263–265]. The TCSM’s ground rules and major parameters are these:1. The lightest doublet of technifermions ( T U , T D ) are color- SU (3) C singlets.
2. The decay constant of the lightest doublet’s technipions is F = ( F π = 246 GeV) · sin χ .In the case of N D fundamentals, sin χ ∼ = 1 /N D (cid:28) . In the case of two-scale TC, F π = (cid:112) F + F = 246 GeV with F /F ∼ = tan χ (cid:28) .3. The isospin breaking of ( T U , T D ) is small. Their electric charges are Q U and Q D = Q U − . In the TCSM, the rates for several decay modes of the technivectors to transversely-polarized electroweak gauge bosons ( γ, W ±⊥ , Z ⊥ ) plus a technipion or longitudinal weak Colored technifermions get a large contribution to their mass from SU (3) C gluon exchange. We also assumeimplicitly that, in the case of N D fundamentals, ETC interactions split the doublets substantially. W ± , L ≡ W ± L , Z L ) and for decays to a fermion-antifermion pair depend sensitivelyon Q U + Q D .4. The lightest technihadrons are the pseudoscalars π ± , T ( I = 1) and the vectors ρ ± , T ( I = 1) , ω T ( I = 0) and axial vectors a ± , T ( I = 1) , f T ( I = 0) . Isospin symmetry and quark-modelexperience strongly suggest M ρ T ∼ = M ω T and M a T ∼ = M f T .
5. Since W ± , L are superpositions of all the isovector technipions, the π T are not mass eigen-states. This is parameterized in the TCSM as a simple two-state admixture of W L and thelightest mass-eigenstate π T : | π T (cid:105) = sin χ | W L (cid:105) + cos χ | π T (cid:105) . (1)Thus, technivector decays involving W L , while nominally, strong interactions, are sup-pressed by powers of sin χ .6. The lightest technihadrons, π T , ρ T , ω T and a T , may be studied in isolation , without sig-nificant mixing or other interference from higher-mass states. This is the most importantof the TCSM’s assumptions. It is made to avoid a forest of parameters.7. In addition to these technihadrons and W ± L , Z L , the TCSM involves the transversely-polarized γ , W ±⊥ and Z ⊥ . The principal production process of the technivector mesons athadron and lepton colliders is Drell-Yan, e.g, ¯ qq → γ, Z → ρ T , ω T , a T → X . This givesstrikingly narrow s -channel resonances at M X = M ρ T ,ω T ,a T if M X can be reconstructed.8. Technipion decays are mediated by ETC interactions and, therefore, are expected to beHiggs-like, i.e., π T preferentially decay to the heaviest fermion pairs they can. There isone exception. Something like topcolor-assisted technicolor [250] is required to give thetop quark its large mass. Then, the coupling of π T to top quarks is not proportional to m t ,but more likely to O ( m b ) [250].This TCSM phenomenology was tested at LEP (see, e.g., Refs. [266, 267]) and the Teva-tron [268–270] for some generic values of the parameters. So far there is no compelling evi-dence for TC, but there are also no significant restrictions on the masses and couplings com-monly used in the TCSM search analyses carried out so far: For ρ T → W π T , the limits are M ρ T > ∼ –
250 GeV , M π T > ∼ –
145 GeV when M W + M π T < M ρ T < M π T [270]; for ρ ± T → W Z , they are M ρ T >
400 GeV , M π T >
350 GeV when M ρ T < M W + M π T [269].Both sets of limits use the P YTHIA defaults [271]: sin χ = 1 / , Q U (cid:39) , N T C = 4 , and the ρ T → π T π T coupling scaled from QCD, g ρ T π T π T = (cid:112) π (2 . /N T C ) . On the other hand,the more general idea of LSTC makes little sense if the limit on M ρ T is pushed past ∼
700 GeV .Therefore, we believe that the LHC can discover it or certainly rule it out.In the June 2007 Les Houches summary report [272], several of the current authors usedP
YTHIA [271] together with the the PGS detector simulator [273] to study the reach of the LHCwith √ s = 14 TeV for the LSTC processes q ¯ q → ρ ± T → W ± Z , a ± T → γW ± , ω T → γZ . (2)In all cases, the W and Z decay to e or µ -type leptons. These decay modes were chosen becausethey are not overwhelmed by backgrounds (as is ρ T → W π T which is swamped by t ¯ t at the We assume that the isosinglet π (cid:48) T is too heavy to play a part in LSTC phenomenology. The f T doesn’t eitherbecause it cannot be produced as an s -channel resonance in q ¯ q collisions. See Sect. 5 for a discussion of this assumption on g ρ T π T π T . (cid:96) + (cid:96) − are also quite promising discovery modes.For Les Houches 2007, we concentrated on three TCSM mass points that cover most ofthe reasonable range of LSTC scales; they are listed in Table 1. In all cases, we assumed isospinsymmetry, together with M ρ T = M ω T and M a T = 1 . M ρ T . The near degeneracy of ρ T and a T was motivated by the argument that it makes the low-scale TC contribution to the S -parametersmall (see Ref. [274] and references therein). The P YTHIA defaults listed above were used aswell as M V , = M A , = M ρ T for the LSTC mass parameters controlling the strength of ρ T , ω T , a T decays to a transverse electroweak boson plus π T or W L [264, 265]. The table alsolists the signal cross sections at 14 TeV and, in parentheses, the minimum luminosities for a σ = S/ √ S + B discovery.Case M ρ T = M ω T M a T M π T σ ( W ± Z ) σ ( γW ± ) σ ( γZ ) A 300 330 200 110 (2.4) 168 (2.3) 19.2 (17)B 400 440 275 36.2 (7.2) 64.7 (4.5) 6.2 (46)C 500 550 350 16.0 (15) 30.7 (7.8) 2.8 (97)
Table 1:
The LH 2007 study’s TCSM masses (in GeV) and signal cross sections times e, µ branch-ing ratios (in fb) for pp collisions at √ s = 14 TeV producing the lightest technihadrons. Numbers inparentheses are the luminosities (in fb − ) needed or a σ discovery [272]. In addition to discovering the narrow resonances in these channels, the angular distribu-tions of the two-body final states in the technivector rest frame provide compelling evidence oftheir underlying technicolor origin. Because all the modes involve at least one longitudinally-polarized weak boson, the distributions are dσ (¯ qq → ρ ± T → W ± L Z L ) d cos θ ∝ sin θ, (3) dσ (¯ qq → a ± T , ρ ± T → γW ± L ) d cos θ , dσ (¯ qq → ω T , ρ T → γZ L ) d cos θ ∝ θ . (4)Simulations were presented in the LH 2007 report. While these studies were very preliminary,they indicated that the ρ ± T → W ± Z and a ± T → γW ± distributions easily could be distinguishedfrom background for M (cid:39)
300 GeV with
10 fb − of data and M (cid:39)
400 GeV with 20–
40 fb − .The smaller ω T → γZ signal rates require much more luminosity, e.g.,
40 fb − for M ω T (cid:39)
300 GeV .There are three motivations for the present study. First, for some time to come, the mainoperating c.m. energy of the LHC will, with some luck, be 10, not 14, TeV. This requires thatour studies be repeated and the reach for LSTC signals be estimated for the lower energy —and lower luminosities — expected for the next several years. Second, as noted above, most ofthe 2007 work was carried out using the PGS detector simulator. While adequate for a first lookat LSTC for the LHC, one really wants more substantial studies using the ATLAS and CMS Our luck did not hold. As this document was being completed, a new LHC run plan was adoptedin which the machine would begin an 18–24 month run in 2010 run at √ s = 7 TeV , followed by along shutdown in which it would be prepared for running at the design c.m. energy of
14 TeV . See http://indico.cern.ch/conferenceDisplay.py?confId=83135 . Some justification for our studies maybe derived from the fact that
10 TeV = (cid:112) (7 TeV)(14 TeV) . Finally, twoof us have developed an effective Lagrangian for LSTC [274]. This can be interfaced with suchtools as MadGraph and CalcHEP to generate cross sections for particle production and decayusing P
YTHIA or HERWIG. We present here a selection of first results comparing the parton-level cross sections generated by our Lagrangian with the TCSM as implemented in P
YTHIA .In this paper we report on several more-in-depth studies for some of the classic LSTCdiscovery channels at the LHC, and we add some new ones. The LSTC processes investigatedin this report and the principal results are the following:1. A CMS study of ρ ± T → W ± Z (Bose, Carrera, Maravin).2. A PGS-based study of ω T → γZγ(cid:96) + (cid:96) − (Black, Smith).3. A CMS-based study of ω T , ρ T , a T → e + e − (Harper).4. Comparisons of an effective Lagrangian, L eff , for LSTC with the TCSM in P YTHIA , in-cluding an investigation of the accuracy of the longitudinal gauge boson approximationfor technivector decays (Martin and Lane). The effective Lagrangian implies some strik-ing differences with the TCSM defined in Refs. [263–265] and implemented in P
YTHIA .In particular, the value of g ρ T π T π T is predicted by L eff and it is considerably smallerthan the value (cid:112) π (2 . /N T C ) obtained by scaling from QCD. Thus, the rate for ρ T → W Z predicted by L eff is much smaller than in the TCSM, while the rate for ρ T → γW can be much larger. This is a new result. It is unclear whether it is moreor less credible than the TCSM, but experiment can decide.The mass points and signal cross sections at √ s = 10 TeV (computed from the TCSM inP YTHIA ) are listed in Table 2. Note that ρ T → W π T is forbidden in Case 1a, enhancing the ρ ± T → W Z branching ratio.Case M ρ T ,ω T M a T M π T M V ,...,A σ ( W ± Z ) σ ( γW ± ) σ ( γZ ) σ ( e + e − )
1a 225 250 150 225 230 330 60 1655 (980)1b 225 250 140 225 205 285 45 1485 (980)2a 300 330 200 300 75 105 11 425 (290)2b 300 330 180 300 45 85 7 380 (290)3a 400 440 275 400 22 40 4 130 (90)3b 400 440 250 400 14 35 3 120 (90)
Table 2:
Technihadron masses, LSTC mass parameters (in GeV) and approximate signal cross sec-tions for pp collisions at √ s = 10 TeV (in fb ) for the 2009 Les Houches study. Isospin symme-try is assumed. Other TCSM parameters are sin χ = 1 / , N T C = 4 , Q U = Q D + 1 = 1 , g ρ T π T π T = (cid:112) π (2 . /N T C ) = 4 . , and CTEQ5L parton distribution functions were used.Branching ratios of W and Z to electrons and muons are included. σ ( e + e − ) includes signal plusstandard-model production integrated over approximately M ρ T ,ω T −
25 GeV to M a T + 25 GeV ; thestandard model cross section for this range is in parentheses. This motivation was thwarted to some extent by the collaborations’ requirements for publishing analyses madewith their software and simulation tools. ρ ± T → W ± Z This section summarizes a CMS study of the detector’s reach for ρ ± T → W ± Z → (cid:96) ± ν (cid:96) (cid:96) + (cid:96) − for (cid:96) = e and/or µ as described in the TCSM and encoded in P YTHIA [275]. This studyupdates one carried out for Les Houches 2007 [276], with pp collisions at √ s = 10 TeV andconcentrating on four TCSM mass points not excluded by other experiments and covering arange accessible with an integrated luminosity < ∼ − , namely, the three cases of Table 2 plus M ρ T = 500 GeV . This analysis uses the detailed GEANT
Sources of background are the standard model
W Z production, plus ZZ and W W , Z + γ , W + jets and Z + jets production ( W or Z boson production in association with a pair ofheavy quark jets, referred to as V QQ , is treated separately), and t ¯ t production. The statisticallysignificant instrumental backgrounds come from Z + jets and t ¯ t production. For instance, in anenergetic Z + jet event, the footprint of a jet in the detector can mimic the leptonic decay of a W boson, making it a perfect technicolor candidate event. Massive top quark pair events alsopopulate the invariant mass peaks. To overcome these backgrounds, the analysis puts stringentidentification requirements on final state leptons, enforces constraints on the particle transversemomenta and on invariant quantities such as the mass of the Z boson, making using of theaforementioned data-driven techniques known to have worked in previous experiments.Signal samples are produced with P YTHIA and processed using a detector simulationbased on CMS
GEANT
4. To simulate next-to-leading order predictions, a K -factor of . ± . is applied to all signal cross section values. Most backgrounds are produced with P YTHIA (although, for some processes, MadGraph was used in the generation) and the same selectioncriteria are applied to signal and background simulation samples. Whenever fast simulationis used for the backgrounds, a cross-check with the full detector simulation is performed toensure proper description of detector effects. Next-to-leading order background cross sectionsand K -factors used in the study can be found in [275]. Events are pre-selected using single muon and electron triggers which are efficient andat least leptons with p T >
10 GeV are required. The pair of like-flavored, opposite chargeleptons with invariant mass M (cid:96)(cid:96) closest to the Z nominal mass are assigned as Z decay products.To reject ZZ background, events with two non-overlapping Z candidates that are found within
50 GeV < M (cid:96)(cid:96) <
120 GeV are eliminated. The most energetic lepton in the remaining poolis assigned to the W boson, and the corresponding neutrino assigned transverse energy equal − /E T , the event missing transverse energy. The W Z candidate invariant mass is determined byforcing the known W invariant mass to the lepton-neutrino system while choosing the smallersolution in the calculation of the longitudinal momentum of the neutrino.Electron candidates, which are reconstructed as energy clusters in the electromagneticcalorimeter with a matched pixel track, are required to have p T >
15 GeV , to be consistentwith shape and energy deposition of an electron shower, and to be isolated in order to suppress While P
YTHIA shows the a T → W Z resonance, the /E T resolution in the detector simulation results in itscoalescing with the larger ρ T peak. (GeV) WZ M
100 200 300 400 500 600 700 E v en t s / G e V (GeV) WZ M
100 200 300 400 500 600 700 E v en t s / G e V (M = 225 GeV) T r n fi WZttZ+jetsVQQ 4l fi ZZW+jets
CMS Preliminary (GeV) WZ M
100 200 300 400 500 600 700 E v en t s / G e V (GeV) WZ M
100 200 300 400 500 600 700 E v en t s / G e V (M = 300 GeV) T r (M = 400 GeV) T r (M = 500 GeV) T r n fi WZttZ+jetsVQQ 4l fi ZZW+jets
CMS Preliminary
Figure 2:
W Z invariant mass distributions for the Case 1a signal ( M ρ T = 225 GeV ) and backgroundsamples (left). W Z invariant mass distributions for signal ( M ρ T in the range 300–500 GeV) and back-ground samples (right). The distributions are normalized to an integrated luminosity of − . misidentified jets. Muons are reconstructed using information from the muon detectors and thesilicon tracker. Those assigned to a Z -boson must have p T >
10 GeV , with no track or isolationrequirement due to the low misidentification rate. Tighter selection criteria ( p T >
20 GeV and isolation) are applied to muons from W candidates since a higher misidentification rateis expected. In addition, a quality cut on the impact parameter significance of the muons isapplied.To enhance the signal to background ratio, two sets of further requirements are used inthis study. The first one optimized for early conditions (or for M ρ T = 225 GeV ), and anotherone optimized for higher luminosity scenarios (or for M ρ T >
300 GeV . These requirementsfor early (late) conditions are: p T ( Z ) >
50 (90) GeV , p T ( W ) >
50 (90) GeV , and H T >
130 (160) GeV , where H T is the scalar sum of the transverse momentum of the three chargedleptons in the final state.Figure 2 shows, the W Z invariant mass distributions for the various mass points for − of integrated luminosity. Table 3 lists the number of signal events expected with
200 pb − ofdata within a mass window of . Gaussian standard deviations around the ρ T mass peak. The physics backgrounds,
W Z and ZZ , are estimated from Monte Carlo simulation. Theinstrumental backgrounds fall into two groups, one that includes a genuine Z -boson and onethat does not. The Z + jets background dominates the first group, which also includes Zγ production (found to be negligible), and Zbb production. In the second group t ¯ t productiondominates, followed by W + jets , and QCD multi-jet production (found to be negligible).The Z + jets background (including V QQ ) is estimated using a data-driven technique,the “matrix method”, used successfully in previous experiments. This method makes use of twosamples, a “tight-cut” sample with events passing all the signal selection criteria, and a “loose-cut” sample where events pass all the signal selection requirements except the isolation cuts onthe W ’s charged lepton. Hence, the number of events in each sample are given by N loose = (cid:15) main ) events per 200 pb − events per 200 pb − ρ T (M=225 GeV) . ± .
037 8 . ± .
17 4 . ± . ρ T (M=300 GeV) . ± .
034 3 . ± .
15 1 . ± . ρ T (M=400 GeV) . ± .
046 1 . ± .
50 1 . ± . ρ T (M=500 GeV) . ± .
047 0 . ± .
20 0 . ± . Table 3:
Final efficiencies and number of events for the various selection criteria for
200 pb − of dataat √ s = 10 GeV . The first three cases are 1a, 2a, 3a in Table 2; in the last case M π T = 350 GeV and M V i = M A i = 500 GeV . The quoted uncertainties include statistical and systematic uncertainties(purely from simulation), the latter described later in the text. N lep + N jet and N tight = (cid:15) tight N lep + P fake N jet . Here, N lep and N jet is the number of eventswith the W candidates reconstructed from true leptons and the fake ones from misidentifiedjets, respectively; (cid:15) tight is the efficiency for true leptons to pass the isolation cuts and P fake is the corresponding efficiency for fake leptons. These efficiencies will be extracted from datausing the standard “tag and probe” method, thus minimizing systematic errors due to simulation.Using Monte Carlo simulation, the efficiencies (cid:15) tight for muons and electrons are estimated tobe (93 . ± . and (96 . ± . , respectively, while the rates P fake for misidentified jets are . ± . for electrons and . ± . for muons. The signal and background contributionsare estimated with these measured efficiencies.The t ¯ t and other backgrounds without a genuine Z -boson, which are assumed to dominatethe tails of the Z -boson mass distribution, are estimated using the sideband subtraction method.The final Z -mass distribution, for an integrated luminosity of
200 pb − , is fit to a linear sum of ahistogram and a quadratic function. The “ Z -shaped” histogram is extracted from a combinationof Z + jets and W Z samples with much looser requirements, and the quadratic contributionfrom a combination of t ¯ t and W + jets samples (which are expected to be rather flat).Table 4 presents a summary of the number of background events expected with
200 pb − for the . σ mass window used above for the signal. The uncertainties in the Z + jets , V QQ , t ¯ t , and W + jets backgrounds are taken from the data-driven techniques. Process ρ T (M=225 GeV) ρ T (M=300 GeV) ρ T (M=400 GeV) ρ T (M=500 GeV) W Z . ± . ± .
502 0 . ± . ± .
214 0 . ± . ± .
156 0 . ± . ± . ZZ . ± . ± .
084 0 . ± . ± .
024 0 . ± . ± .
010 0 . ± . ± . Z +jets and V QQ . ± . ± .
506 0 . ± . ± .
064 0 . ± . ± .
020 0 . ± . ± . t ¯ t and W +jets . ± . ± .
247 0 . ± . ± .
104 0 . ± . ± .
065 0 . ± . ± . Total . ± . ± .
76 1 . ± . ± .
25 1 . ± . ± .
17 0 . ± . ± . Table 4:
Summary of final number of background events for 200 pb − of data at √ s = 10 GeV . Sta-tistical and systematic uncertainties (in this order) are also given. Statistical uncertainties include thosefrom data-driven methods for this low luminosity. In the absence of an excess of signal events,
C.L. upper limits can be set on the crosssections. These limits, as functions of integrated luminosity, are summarized in Fig. 3. Thefinal results are presented in Table 5, which include a second set of technicolor parameters that82 -1 Int. Luminosity (pb % C . L . c r o ss s e c t i on li m i t ( pb ) -2 -1
10 1 (M = 225 GeV) T r (M = 300 GeV) T r (M = 400 GeV) T r (M = 500 GeV) T r CMS Preliminary ) -1 Int. Luminosity (pb % C . L . c r o ss s e c t i on li m i t ( pb ) -2 -1
10 1
Figure 3:
95% C.L. limits for σ ( ρ T → W Z ) as a function of integrated luminosity for pp collisions at √ s = 10 GeV . The cross sections include the branching ratio to electrons and muons. The horizontalbands, which indicate the theoretical cross sections (and their associated uncertainty), intersect thelimit curves at approximately the values given in Table 5.Mass values Int. luminosity Int. luminosity Int. luminosityfor C.L limit for
C.L limit for
C.L limit(pb − ) (+ theoretical (- theoreticaluncertainty) (pb − ) uncertainty) (pb − ) M ρ T = 225 GeV , M π T = 150 GeV 400 240 790 M ρ T = 300 GeV , M π T = 200 GeV 440 290 790 M ρ T = 400 GeV , M π T = 275 GeV 1040 710 1800 M ρ T = 500 GeV , M π T = 350 GeV 2050 1450 3310 M ρ T = 225 GeV , M π T = 140 GeV 540 300 1060 M ρ T = 300 GeV , M π T = 180 GeV 1300 800 2550 Table 5:
Integrated luminosity at √ s = 10 GeV needed for exclusion at C.L. The last two columnsindicate the values of integrated luminosity (in fb − ) needed if the theoretical uncertainty in the signalis taken into account. The last two rows show results for different parameter sets for the mass points ρ T = 225 GeV and ρ T = 300 GeV. use lower values for M π T from cases 1b and 2b in Table 2. These limits use the results for
200 pb − given in Table 4. The statistical uncertainty in the total background is scaled withluminosity while the relative systematic uncertainty is kept constant throughout.83s expected from Table 1 (constructed for √ s = 14 TeV ), a σ discovery of technicolorparticles via the ρ T → W Z → leptons process will require well over − of data. ω T → γZ → γ(cid:96) + (cid:96) − The decay ω T → γZ → γ(cid:96) + (cid:96) − may be the discovery channel for ω T at the LHC. This isespecially true if Q U + Q D = 0 , in which case ω T → (cid:96) + (cid:96) − is forbidden (just as in QCD!). Thissection presents a simplified study of ω T → γZ → γµ + µ − using the PGS detector simula-tor [273]. A more in-depth analysis using ATLAS simulation tools for ω T → γZ → γe + e − could not receive collaboration approval for its release in time for this document’s submission.The present PGS-based analysis should be a plausible feasibility study. Another very importantfeature of the γZ mode is its angular distribution. In the approximation that the Z is longitudi-nally polarized, as expected in LSTC, it is θ .Signal and background cross sections were calculated using P YTHIA . The γµ + µ − signalrates are half those in the σ ( γZ ) column of Table 2. The two principal backgrounds are thestandard-model production of γZ and Z + jets where a jet fakes a photon; see the 2007 LesHouches study of LSTC in Ref. [272]. The cross sections for the standard γµ + µ − and Z + jets cross sections are . and , respectively. A parameterized detector simulation with PGS was used to give an estimate of an LHC detec-tor’s response. The parameterization was chosen to correspond to the approximate behavior ofATLAS and CMS. Most notably we assumed a muon identification efficiency of 95%, a photonefficiency of 80%, and a jet to photon misidentification rate of − .The most significant backgrounds are expected from Z events with (1) a photon radiatedfrom the initial ¯ qq or from the Z ’s decay leptons or (2) a quark or gluon jet misidentified as aphoton. To reduce these backgrounds we take advantage of two aspects that differ in signal andbackground kinematics.1. The signal Z -boson will be centrally produced and with typically large transverse mo-menta. In contrast, p T ( Z ) = 0 in lowest order and nonzero p T comes from parton orphoton radiation processes having rapidly falling cross-sections.2. The signal photons should be isolated from the Z or its decay products whereas the radi-ated photons and gluons tend to follow the object which produced them.Therefore, we required the following:1. Two muons of opposite sign, each with p T >
15 GeV and η < . reconstructing a Z -boson within
15 GeV of the nominal Z -mass of . .2. A photon with p T > GeV and η < . .3. The photon and muons have ∆ φ > .4. The photon and Z have ∆ φ > .The efficiencies on the signal and background samples are displayed in Tables 6 and 784ase Z -boson selection photon selection ∆ φ ( γµ ) > φ ( γZ ) >
1a 0.45 ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± ± Table 6:
Cumulative efficiencies for signal event selection in pp → ω T + X , ω T → γZ → γµ + µ − at √ s = 10 TeV . Background Z -boson election photon selection ∆ φ ( γµ ) > ∆ φ ( γZ ) > Zγ ± ± ± ± Z + jets ± ± (7 ± × − (5 . ± × − Table 7:
Cumulative efficiencies for background event selection in pp → ω T + X , ω T → γZ → γµ + µ − at √ s = 10 TeV . (GeV) T ! M220 240 260 280 300 320 340 360 380 400 ) - L d t (f b " Graph T ! M220 240 260 280 300 320 340 360 380 400 ) - L d t ( f b " Graph
Figure 4:
Left: Integrated luminosity of pp collisions at √ s = 10 TeV required for σ evidence (dashed)and σ observation (solid) of ω T → γZ → γµ + µ − as a function of M ω T for LSTC Cases a (blue) and b(black). Right: Integrated luminosity required for 95% C.L. exclusion of Cases a (blue) and b (black). The low branching ratio for ω T → γZ makes this analysis channel significantly morechallenging than the other LSTC processes considered in this report. To evaluate the channel’sdiscovery potential we computed two quantities by counting the events within a
20 GeV windowof the assumed signal mass window: (1) the discovery potential by evaluating the 3 and σ luminosity contours by a simple event counting method; (2) the luminosity required for 95%C.L. exclusion of the signal if none is found. The results are shown in Fig. 4. Depending on themasses the luminosity for σ discovery ranges from a few to
100 fb − . The exclusion contoursare approximate because the rate of Z + jets passing the selection cuts is only approximately85nown. ω T , ρ T , a T → e + e − The neutral states ρ T , ω T and a T all decay to (cid:96) + (cid:96) − (unless Q U + Q D = 0 in which case ω T → (cid:96) + (cid:96) − vanishes.) In the TCSM as implemented in P YTHIA , the ω T signal is generallymuch greater that the ρ T one because of the latter’s larger rates into W π T and W W . In thissection we present an estimate of LHC reach for these technivectors decaying to e + e − based ona CMS study of the Drell-Yan process at √ s = 10 TeV [277]. As we shall see, the a T may bevisible in this mode with only moderate luminosity at M a T < ∼
330 GeV . The presence of thenearby second resonance distinguishes this LSTC signal from Z (cid:48) or G RS searches. (An ATLASstudy of ω T → µ + µ − at √ s = 14 TeV may be found in Ref. [163].)The CMS Collaboration has released public results showing the expected result of an e + e − mass spectrum from
50 GeV to for pp collisions at
10 TeV [277]; this is an updateof a previous study for √ s = 14 TeV [278]. This result is re-interpreted in this report toestimate the sensitivity of the LHC to technicolor using CMS. This is a private interpretationusing information the CMS collaboration has made public and is not an official approved resultof CMS collaboration. The e + e − mass spectrum measurement along with the estimated systematic uncertainties istaken from a preliminary CMS summer 09 result [277]. The parameters for this study are thefollowing: the electron ID efficiency is ± %; e + e − mass resolution is 2%; the uncertainty inthe standard-model Drell-Yan is 11%; the t ¯ t background uncertainty is 16%; the jet backgrounduncertainty is 50%; and a K -factor of 1.35 is used for the Drell-Yan signal and background. Thesystematic uncertainties on the backgrounds are conservative and approximately twice as largeas a similar CDF analysis [279]. Therefore, the possibility that the systematic uncertainties arehalf as large is also considered here. While the CMS Collaboration has made no statement onwhether this reduction is possible, experience at the Tevatron suggests that it will be. The tech-nicolor signal sample is generated using P YTHIA . Both generator level electrons are requiredto satisfy E T >
50 GeV and | η | < . or . < | η | < . corresponding to the kinematicand geometric acceptance of the CMS analysis. As can be seen from Fig. 5 for Case 2a inTable 1, the 2% mass resolution is sufficient to resolve the ω T and a T resonances at 300 and
330 GeV . While the two peaks are distinguishable, the interference effect between the standardmodel and TC signal below the first peak is not visible with this resolution. Figure 5 also showsa sample pseudo-experiment in the presence of technicolor with the predicted standard modelbackgrounds.The technique used to estimate the significance of a technicolor signal is a p -value methodused in the CDF e + e − search described in [279]. This method addresses the “look-elsewhere”effect resulting from the fact that the mass of a new boson resulting from new physics is notknown. First a pseudo-experiment is generated from the expected standard model backgroundmass distribution using a Poisson distributed random number for each bin. Then in a masswindow of ± . times the mass resolution, the Poisson probability, or p -value, of observingthe number of observed events or greater in the absence of new physics is calculated. Theuncertainty on the number of background events is included by averaging the p -values for allpossible background values weighted by a Gaussian with mean and sigma equal to the expected86 (GeV/c ee M
250 300 350 400 450 E ve n t s / G e V / c gen level2% detector res -1 L = 600 pb (cid:242) ) (GeV/c ee M
200 250 300 350 400 450 500 550 E ve n t s / G e V / c pseudo-data SM + TCTC Signal SM Drell-Yan, tW, WW, jet bkgstt -1 L dt = 600 pb (cid:242) = 10 TeVsLHC Experiment,
SM Bkg. Expectations takenfrom CMS-PAS-EXO-09-006
Figure 5:
A comparison of the LSTC signal at generator level and after detector resolution (left) anda pseudo-experiment for Case 2a together with the standard model backgrounds (right); E T ( e ± ) >
50 GeV . The standard model backgrounds are taken from Fig. 2a of [277], scaled by a factor of six toaccount for the luminosity difference [279]. background and its uncertainty. This is done in 1 GeV steps for masses between 200 and1000 GeV. This process is repeated for × pseudo-experiments per luminosity point andthe two smallest p -values in each pseudo-experiment are recorded. The mass windows used tocalculate the p -values are not allowed to overlap to ensure that they do not share any events.Then the process is repeated in the presence of the technicolor signal and the median p -valueis obtained for the signal bins. The fraction of standard-model-only pseudo-experiments whichobserve this p -value or greater is then obtained to determine how often a similar sized signalcan be produced from chance alone.The advantage of this search technique is that it uses very few assumptions and is genericto all new physics types. As there are two peaks, the p -values for both peaks are calculated.Then the fraction of pseudo-experiments generated with standard-model-only templates thathave a p -value < p ω T and another p -value < p a T is determined, where p ω T and p a T are the p -values of the two peaks. This offers some increase in sensitivity compared to using only theleading peak.Limits are then set via a simple Bayesian likelihood method using Poisson statistics. The ± . mass resolution region around each peak correspond to the two bins of the likelihood. Thebackground uncertainty is assumed to be modeled by a truncated Gaussian and that backgrounduncertainty is 100% correlated between the two bins. Tables 8 and 9 show the fraction of standard-model-only pseudo-experiments that have havetwo p -values somewhere in the mass spectrum larger than the median p -value of each peak inthe presence of technicolor for Cases 1a,b and 2a,b respectively. An ω T with mass
225 GeV and M π T = 150 GeV (Case 1a) is discoverable at the σ -level with
200 pb − , while Case 1b87 L dt ( pb − ) Case 1a Case 1a (imp. syst.) Case 1b Case 1b (imp. syst.)50 0.022 (5.5 × − ) 0.017 (3.9 × − ) 0.24 (6.1 × − ) 0.20 (5.0 × − )100 1.0 × − (3.2 × − ) 5.5 × − (9.7 × − ) 0.017 (4.8 × − ) 7.0 × − (1.2 × − )150 1.4 × − (7.1 × − ) 2.7 × − (8.0 × − ) 9.7 × − (3.0 × − ) 3.6 × − (8.0 × − )200 < × − (5.2 × − ) < × − (3.5 × − ) 3.5 × − (2.4 × − ) 7.3 × − (2.9 × − )250 < × − (9.3 × − ) < × − (9.2 × − ) 2.1 × − (2.3 × − ) 2.1 × − (1.2 × − ) Table 8:
The fraction of standard-model-only pseudo-experiments which observe a p -value equal to orlower than the median p -value of the first peak (shown in parentheses) and a second p -value equal to orlower than the median p -value of the second peak of LSTC Cases 1a and 1b. The improved systematicscorrespond to a reduction of background uncertainty by 50% which is a level comparable to that in asimilar CDF analysis [279]. requires ∼
300 pb − . Strong evidence can be obtained for
100 pb − . This puts discovery ofsuch a model well within the expected reach of the first run of the LHC. For √ s = 10 TeV andluminosities of 600–
800 pb − strong evidence can be obtained for Cases 2a,b. Cases 3a,b cannot be distinguished from background at luminosities less than − . Improving the systematicuncertainties gives on average a factor of five increase in significance.In the absence of a signal, limits can be set on the technicolor models. Table 10 showsthe luminosity required at
10 TeV to exclude the various cases considered. Cases 1a and 1bcan be excluded very quickly, requiring 20 and
31 pb − , respectively. Cases 2a and 2b can beexcluded with 170 and
360 pb − of data. Cases 1a,b and 2a and, possibly, 2b could thereforebe excluded by an LHC experiment in 2010–11, however Cases 3a,b require significantly moredata, on the order of an inverse femtobarn. Reducing the systematic uncertainties would reducethe luminosity required to exclude the LSTC models considered here by 10-15% which couldbe important in excluding Cases 2a and 2b by the end of 2011. (cid:82) L dt ( pb − ) Case 2a Case 2a (imp. syst.) Case 2b Case 2b (imp. syst.)400 0.064 (1.4 × − ) 0.042 (6.1 × − ) 0.72 (0.068) 0.72 (0.053)600 8.1 × − (2.2 × − ) 3.0 × − (4.4 × − ) 0.36 (0.039) 0.33 (0.024)800 2.0 × − (6.5 × − ) 3.6 × − (6.2 × − ) 0.089 (0.025) 0.066 (0.012)1000 2.0 × − (1.0 × − ) 1.3 × − (3.8 × − ) 0.067 (0.017) 0.033 (0.0063)1500 3.1 × − (5.1 × − ) 3.5 × − (2.2 × − ) 4.7 × − (9.6 × − ) 7.1 × − (1.9 × − ) Table 9:
The fraction of standard-model-only pseudo-experiments which observe a p -value equal to orlower than the median p -value of the first peak (shown in parentheses) and a second p -value equal to orlower than the median p -value of the second peak of LSTC Cases 2a and 2b. The improved systematicscorrespond to a reduction of background uncertainty by 50% which is a level comparable to that in asimilar CDF analysis [279]. L eff for low-scale technicolor There are three motivations for an effective Lagrangian for LSTC [274]. First, longitudinallypolarized electroweak bosons, W ± L and Z L , play an important role in the TCSM described inSect. 1 and are expected to appear in many of the technivector decays accessible at a hadroncollider. They are treated in the TCSM in the approximation W ± Lµ = ∂ µ Π ± T /M W and Z Lµ = ∂µ Π T /M Z . This is valid when p W (cid:29) M W , but that is not always the case, especially when ρ T → W Z for the lightest ρ T we consider here. Therefore, we want a consistent mathematicaltreatment of longitudinal and transverse weak bosons that a Lagrangian can furnish. This will88odel nominal syst. improved syst.1a 20 201b 31 312a 170 1502b 360 3203a 610 5603b 1120 930 Table 10:
Luminosities (in pb − ) needed at √ s = 10 TeV to exclude ω T → e + e − in the LSTC modelsin Table 1 at 95% C.L. Nominal systematics are on the left. Improved systematics on the right correspondto a reduction of background uncertainty by 50% which is a level comparable to that in a similar CDFanalysis [279]. also allow us to assess the transverse weak boson contribution to the angular distributions inEq. (3). Second, a Lagrangian makes available the versatility of such programs as MadGraphand CalcHEP for generating amplitudes to be used in P YTHIA and HERWIG. Finally, the TCSMdescribes a phenomenology of LSTC expected to be valid only in the limited energy √ ˆ s < ∼ M ρ T ,where the lightest technihadrons may be treated in isolation. An effective Lagrangian, L eff , iswell-suited for this description because it gives warning of its limitation.The hidden local symmetry (HLS) formalism of Bando, et al. [280, 281] was adoptedto construct an L eff describing the technivector mesons, electroweak bosons and technipions ofLSTC. Such an L eff guarantees that production of W L , Z L via annihilation of massless fermionsis well-behaved at all energies in tree approximation. Elastic W L W L scattering still behaves athigh energy as it does in the standard model without a Higgs boson, i.e., the amplitude ∼ ˆ s/F π at ˆ s (cid:29) M ρ T . Of course, this violation of perturbative unitarity signals the strong interactions ofthe underlying technicolor theory. The HLS method also guarantees that the photon is masslessand the electromagnetic current conserved.The gauge symmetry group of L eff is G = SU (2) W ⊗ U (1) Y ⊗ U (2) L ⊗ U (2) R . The firsttwo groups are the standard electroweak gauge symmetries, with primordial couplings g and g (cid:48) and gauge bosons W = ( W , W , W ) and B . The latter two are the “hidden local symmetry”groups. The underlying TC interactions are parity-invariant, so that their zeroth-order couplingsare equal, g L = g R = g T . The assumed equality of the SU (2) L,R and U (1) L,R couplings reflectthe isospin symmetry of TC interactions and the expectation that M ρ T ∼ = M ω T and M a T ∼ = M f T .This symmetry must be broken explicitly if L eff is to allow an appreciable ρ T – ω T splitting. Wehave not done that. The gauge bosons ( L , L ) and ( R , R ) contain the primordial technivectormesons, V , V , A , A ∼ = ρ T , ω T , a T , f T .To describe the lightest π T and to mock up the heavier TC states that contribute most toelectroweak symmetry breaking (see Sect. 1), and to break all the gauge symmetries down toelectromagnetic U (1) , the nonlinear Σ -model fields in L eff are Σ , ξ L , ξ R and ξ M , transformingunder G as Σ → U W Σ U † Y , ξ L → U W U Y ξ L U † L ,ξ M → U L ξ M U † R , ξ R → U R ξ R U † Y . (5) Mixing between ρ T and ω T is limited by the smallness of the T -parameter. D µ Σ = ∂ µ Σ − ig t · W µ Σ + ig (cid:48) Σ t B µ ,D µ ξ L = ∂ µ ξ L − i ( g t · W µ + g (cid:48) y t B µ ) ξ L + ig T ξ L t · L µ ,D µ ξ M = ∂ µ ξ M − ig T ( t · L µ ξ M − ξ M t · R µ ) ,D µ ξ R = ∂ µ ξ R − ig T t · R µ ξ R + ig (cid:48) ξ R ( t + y t ) B µ , (6)where t · L µ = (cid:80) α =0 t α L αµ and t = τ , t = . The hypercharge y = Q U + Q D of the TCSM.The field Σ contains the technipions that get absorbed by the W and Z bosons. They are anisotriplet of F -scale Goldstone bosons, where F = F π cos χ (cid:29) F , and χ was introducedin Sect. 1. It is parameterized as Σ ( x ) = exp (2 i t · π ( x ) /F ) . It is convenient to define Σ = ξ L ξ M ξ R ; then Σ → U W Σ U † Y D µ Σ = ∂ µ Σ − ig t · W µ Σ + ig (cid:48) Σ t B µ . (7)In the unitary gauge ( Σ , ξ L , ξ R → ) this field will be parameterized as Σ = exp (2 it · ˜ π/F ) ,where ˜ π are the isovector and isoscalar technipions π T , π (cid:48) T up to a normalization constant.The complete effective Lagrangian is L eff = L Σ + L WZW + L gauge + L ¯ ff + L M π + L π T ¯ ff . (8)Here, L Σ = F Tr | D µ Σ | + F (cid:110) a Tr | D µ Σ | + b (cid:104) Tr | D µ ξ L | + Tr | D µ ξ R | (cid:105) + c Tr | D µ ξ M | + d Tr( ξ † L D µ ξ L D µ ξ M ξ † M + ξ R D µ ξ † R D µ ξ † M ξ M ) − if g T Tr( D µ ξ M ξ † M D ν ξ M ξ † M t · L µν + ξ † M D µ ξ M ξ † M D ν ξ M t · R µν ) (cid:111) . (9)The dimensionless constants a, b, c, d, f are expected to be O (1) . The first four terms are thoseinvolving only two derivatives and/or gauge fields that are consistent with the symmetries ofTC interactions. The f -term is needed to describe decays of a T . It is one of several possibil-ities and, to minimize the number of free parameters, only one such term is used. The L WZW interaction includes the Wess-Zumino-Witten (WZW) terms [282, 283] implementing the ef-fects of anomalously nonconserved symmetries of the underlying TC theory. They are essentialfor describing the radiative decays of ρ T and ω T as well as π T → γγ . They are described inmore detail in Refs. [274] and [284]. The remaining terms in L eff are the gauge kinetic terms,couplings of quarks and leptons to ( SU (2) ⊗ U (1)) EW gauge bosons, π T mass terms, and thecouplings of π T to quarks and leptons.This Lagrangian describes production and decay of the technivector mesons. In this sec-tion we concentrate on the modes ρ ± T , a ± T → W ± Z and γW ± . The operators describingthe on-shell decays ρ ± T , a ± T → W Z are rather complicated and they are given in Ref. [274],Eqs. (47) and (56). The purely longitudinal process ρ ± T → W L Z L is controlled by the coupling g ρ T π T π T and, as we discuss below, L eff predicts a considerably smaller value of this parameterthan was used in the TCSM. This and the small W, Z momenta make the transverse W and Z contributions to this decay at least as important as the longitudinal ones. The longitudinal- W L ( ρ ± T → γW ± ) = egy F π sin χ M V [ ρ + T µ W − ν + ρ − T µ W + ν ] (cid:101) F µν (cid:39) ey sin χ M V [ ρ + T µν Π − T + ρ − T µν Π + T ] (cid:101) F µν ; (10) L ( a ± T → γW ± ) = − iegF π sin χ M A ( a + T ν W − µ − a − T ν W + µ ) F µν (cid:39) ie sin χ M A ( a + T µν Π − T − a − T µν Π + T ) F µν . (11)Here, F µν is the electromagnetic field strength and (cid:101) F µν = (cid:15) µνλρ F λρ is its dual. The massparameters M V and M A are set equal M ρ T in this study. The coupling g ρ T π T π T and the TCSM mass parameters M V i and M A i are functions of the L eff couplings a, . . . , f and of F π , sin χ and N T C . It is both possible and natural to choose asinputs F π , sin χ , N T C , M ρ T = M ω T , M a T and the mass parameters M V , M A and M A (onlythese enter the technivector decays we study) and to express f , g T and g ρ T π T π T in terms of them.This is what was done in the TCSM in P YTHIA except that there g T is the ρ T → π T π T couplingand was chosen to be ( g T / π ) T CSM = 2 . /N T C ) . We obtain: g T = 16 √ π M A F π sin χN T C M V ( M A + M A ) ,f = (4 πM A F π sin χ ) N T C M V M A ( M A + M A ) ,g ρ T π T π T = M ρ T √ g T ( F π sin χ ) (cid:20) f − M A M A (cid:21) . (12)In the present study we set M V i = M A i = M ρ T . In this case, g T = 8 √ π F π sin χN T C M ρ T , g ρ T π T π T = M ρ T F π sin χ . (13)This expression for g ρ T π T π T (but not for g T ) is what one would expect for a Higgs mechanismorigin for M ρ T with gauge coupling (cid:39) g ρ T π T π T and Goldstone boson decay constant F π sin χ (cid:39) F . It is also reminiscent of the KSRF relation [285, 286]. In QCD, the parameter M V controlling ρ → γπ is
700 MeV , very close to M ρ . The F -scale contribution to the s -parameter vanishes in this limit. g ρ T π T π T Γ( ρ ± T ) B ( W Z ) ρ T B ( γW ) ρ T Γ( a ± T ) B ( W Z ) a T B ( γW ) a T
1a 1.372 46 0.349 0.133 93 0.103 0.0951b 1.372 84 0.191 0.072 113 0.085 0.0782a 1.829 282 0.221 0.033 146 0.124 0.087
Table 11:
The ρ T → π T π T decay constant g ρ T π T π T and total widths (in MeV) and branching ratiosfor ρ ± T and a ± T decays to W ± Z and γW ± for cases 1a,b and 2a. Note that the QCD-inspired value of g ρ T π T π T used in P YTHIA is (cid:112) π (2 . /N T C ) = 4 . for N T C = 4 . Other TCSM parameters usedare sin χ = 1 / , N T C = 4 and Q U = Q D + 1 = 1 (i.e., y = 1 ). Case σ ( W Z ) ρ T σ ( W Z ) a T σ ( γW ) ρ T σ ( γW ) a T
1a 45 (35) 4.3 (30) 1765 (905) 860 (555)1b 25 (35) 3.4 (30) 920 (905) 695 (555)2a 17 (20) 3.7 (17) 280 (245) 575 (160)
Table 12:
Parton-level ρ ± T , a ± T signal cross sections (in fb ) for pp collisions at √ s = 10 TeV for cases1a,b and 2a. Cross sections were calculated using L eff and by integrating over ±
20 GeV about theresonances. Cross sections in parentheses are the underlying standard model rates. Branching ratios of W and Z to electrons and muons are included. Other TCSM parameters used are sin χ = 1 / , N T C = 4 and Q U = Q D + 1 = 1 (i.e., y = 1 ). The important consequence of Eq. (12) is that α ρ T π T π T = g ρ T π T π T / π is proportional to M ρ T . For the M ρ T of low-scale technicolor, α ρ T π T π T is considerably smaller than the defaultvalue . /N T C ) used in the P YTHIA implementation of the TCSM. This greatly reduces thebranching ratios B ( ρ T → W π T , W Z ) and, so long as y is not small, correspondingly enhances B ( ρ T → γπ T , γW ) ; see Table 11. We do not know which value of g ρ T π T π T is more reliable.The KSRF relation g ρππ = M ρ / √ f π works well in QCD. If HLS is more than an accidentaldescription of the low-energy QCD spectrum (see Ref. [287] for a contrary view), that may lendcredence to using the smaller value of g ρ T π T π T here. Still, we must remember the admonition torely with suspicion on QCD for describing walking technicolor. Only experiment can decide.The cross sections for ρ ± T , a ± T → W ± Z and γW ± , followed by W and Z decays toelectrons and muons, for cases 1a (in which ρ T → W π T is forbidden), 1b and 2a are listed inTable 12. The effect of the small g ρ T π T π T on these cross sections compared to the P YTHIA ratesin Table 2 is dramatic.The parton-level invariant mass and angular distributions for these three cases of ρ ± T , a ± T → W ± Z are shown in Figs. 6, 7 and 8. CTEQ5l parton distribution functions were used. Althoughno experimental realism was included in these calculations, comparing with the results of theCMS study in Sect. 3 (see Table 2 and Fig. 2), it seems unlikely that ρ ± T → W Z with suchsmall g ρ T π T π T could be discovered with only 1– − at √ s = 10 TeV . We won’t speculateon what it would take to observe the angular distributions and determine whether or not theyfit the LSTC expectation because no serious studies have been done. However, it is noteworthythat the sideband-subtracted angular distribution (calculated by integrating the standard-modelcontribution over the resonance region and subtracting it from the total cross section) is con-siderably larger than the standard-model one and that it looks much more like sin θ than thestandard model does. It is also clear that, as expected for small g ρ T π T π T , there is substantial92 igure 6: The W ± Z invariant mass (left) and angular (right) distributions calculated from L eff forCase 1a with √ s = 10 TeV , M ρ T = 225 GeV , M a T = 250 GeV and M π T = 150 GeV . The angulardistributions are for the ρ ± T → W ± Z region, and the total (red), standard-model (green), total - SM(blue dashed) and pure sin θ (black dashed) are shown. The standard-model contribution is calculatedover the resonance region. Figure 7:
The W ± Z invariant mass (left) and angular (right) distributions calculated from L eff forCase 1b with √ s = 10 TeV , M ρ T = 225 GeV , M a T = 250 GeV and M π T = 140 GeV . The angulardistribution is for ρ ± T → W ± Z . igure 8: The W ± Z invariant mass (left) and angular (right) distributions calculated from L eff forCase 2a with √ s = 10 TeV , M ρ T = 300 GeV , M a T = 330 GeV and M π T = 200 GeV . The angulardistribution is for ρ ± T → W ± Z . contribution to ρ ± T → W Z from transversely-polarized W or Z , and that this flattens out theangular distributions compared to sin θ . Figure 8 shows that ρ ± T → W L Z L becomes moreimportant as M ρ T increases.The invariant mass and angular distributions of ρ ± T , a ± T → γW ± for cases 1a,b and 2a areshown in Figs. 9, 10 and 11. Thanks to the substantially larger branching ratio for ρ ± T → γW that L eff predicts (for y = O (1) ), both resonances can be seen with quite modest luminosity.Conversely, it appears that − at √ s = 7 TeV would be sufficient to exclude these cases.If the resonances are discovered at the rates shown here, the angular distributions, shown for a ± T → γW , should be measurable as well. The sideband-subtracted distributions are quite closeto the θ expected for a γW L signal. Conclusions and outlook
Low-scale technicolor remains a well-motivated scenario for strong electroweak symmetrybreaking with a walking TC gauge coupling. The Technicolor Straw-Man framework outlinedin Sect. 1 provides the simplest phenomenology of this scenario by assuming that the lightesttechnihadrons — ρ T , ω T , a T and π T — and the electroweak gauge bosons can be treated inisolation. This framework is implemented in P YTHIA . A new effective Lagrangian approachallows direct quantitative tests of some the assumptions on which the TCSM is based, in par-ticular, the dominance of longitudinally-polarized gauge bosons in technivector decay rates andangular distributions.In this report, we used P
YTHIA and various detector simulations, and the effective La-grangian (at the parton level) to study technivector decays to W ± Z , γZ , γW ± and e + e − .At the time of the 2009 Les Houches Summer Study, the initial LHC plan was to run at √ s = 10 TeV , and so all our studies were carried out for this energy and luminosities of O (1 fb − ) . As the report was being written, the LHC run plan for 2010-11 changed to run-94 igure 9: The γW ± invariant mass (left) and angular (right) distributions calculated from L eff forCase 1a with √ s = 10 TeV , M ρ T = 225 GeV , M a T = 250 GeV and M π T = 150 GeV . The angu-lar distributions are for the a ± T → γW ± region, and the total (red), standard-model (green), total - SM(blue dashed) and pure sin θ (black dashed) are shown. The standard-model contribution is calculatedover the resonance region. Figure 10:
The γW ± invariant mass (left) and angular (right) distributions calculated from L eff forCase 1b with √ s = 10 TeV , M ρ T = 225 GeV , M a T = 250 GeV and M π T = 140 GeV . The angulardistribution is for a ± T → γW ± . igure 11: The γW ± invariant mass (left) and angular (right) distributions calculated from L eff forCase 2a with √ s = 10 TeV , M ρ T = 300 GeV , M a T = 330 GeV and M π T = 200 GeV . The angulardistribution is for a ± T → γW ± . ning at with the aim of collecting about − of data. The reach of LHC experimentsat for the resonant processes discussed here may be estimated from our results by usingthe parton-parton luminosities and their ratios in Ref. [288]. Overall, the first run of the LHCshould be able to set some useful new limits on low-scale technicolor. We reiterate what wesaid two years ago: With sufficient luminosity, generally in the range of 5–
40 fb − , the LHC atits design energy of
14 TeV can discover or rule out low-scale technicolor in the channels dis-cussed here; with more luminosity angular distributions can be measured to determine whethertechnicolor is the underlying dynamics of discovered resonances. Thus, by the time of the nextLes Houches Summer Study, we all hope that we can return to more in-depth studies of LHCreach at
14 TeV . We conclude as we did two years ago: the main goal of our Les Houches work,as it is for the other “Beyond the Standard Model” scenarios investigated for Les Houches 2009,is to motivate the ATLAS and CMS collaborations to broaden the scope of their searches forthe origin and dynamics of electroweak symmetry breaking.“Faith” is a fine inventionWhen Gentlemen can see —But
Microscopes are prudentIn an Emergency.— Emily Dickinson, 1860
Acknowledgements
We thank the organizers and conveners of the Les Houches workshop, “Physics at TeV Col-liders”, for a most stimulating meeting and for their encouragement in preparing this work.96e benefited from Conor Henderson’s participation in our group at Les Houches. We thankmany other participants, too numerous to name, for spirited discussions. Lane is indebted toLaboratoire d’Annecy-le-Vieux de Physique des Particules (LAPP) and Laboratoire d’Annecy-le-Vieux de Physique Theorique (LAPTH) for generous hospitality and support. He thanksLouis Helary and Nicolas Berger of LAPP for many illuminating discussions. This researchwas supported by the U.S. Department of Energy under Grants DE-FG02-91ER40654 (Blackand Smith), DE-FG02-91ER40676 (Bose and Lane), and Fermilab operated by Fermi ResearchAlliance, LLC under contract number DE-AC02-07CH11359 by the U.S. Department of Energy(Martin). 97 igh mass resonances ontribution 9LHC studies inspired by warped extra dimensions
K. Agashe, L. Basso, G. Brooijmans, S.P. Das, H. Gray, M. Guchait, J. Jackson,M. Karagöz, S.J. Lee, R. Rosenfeld, C. Shepherd-Themistocleous and M. Vos
Abstract
The framework of a warped extra dimension with the Standard Model(SM) fields propagating in it is a very well-motivated extension of theSM since it can address both the Planck–Weak and flavor hierarchyproblems of the SM. We consider processes at the large hadron collider(LHC) inspired by signals for new particles in this framework. Ourstudies include identification of boosted top quarks and
W/Z , produc-tion of a particle called radion with Higgs-like properties and effects offlavor violating tc Higgs coupling. Introduction
The framework of a warped extra dimension à la Randall-Sundrum (RS1) model [289], but withall the SM fields propagating in it [290–294] is a very-well motivated extension of the StandardModel (SM): for a review and further references, see Ref. [295]. Such a framework can addressboth the Planck–Weak and the flavor hierarchy problems of the SM, the latter without resultingin (at least a severe) flavor problem. The versions of this framework with a grand unifiedgauge symmetry in the bulk can naturally lead to precision unification of the three SM gaugecouplings [296] and a candidate for the dark matter of the universe (the latter from requiringlongevity of the proton) [297, 298]. The new particles in this framework are Kaluza-Klein (KK)excitations of all SM fields with masses at ∼ TeV scale. In addition, there is a particle, denotedby the “radion”, which is roughly the degree of freedom corresponding to the fluctuations ofthe size of extra dimension, and typically has a mass at the weak scale. In this write-up, wesummarize some of the signals at the large hadron collider (LHC) for these new particles. Someof these studies can be useful in other contexts as well. Review of Warped Extra Dimension
The framework consists of a slice of anti-de Sitter space in five dimensions (AdS ), where (dueto the warped geometry) the effective 4D mass scale is dependent on position in the extra dimen-sion. The 4D graviton, i.e., the zero-mode of the 5D graviton, is automatically localized at oneend of the extra dimension (called the Planck/UV brane). If the Higgs sector is localized at theother end (in fact with SM Higgs originating as 5th component of a 5D gauge field ( A ) it is au-tomatically so [212]), then the warped geometry naturally generates the Planck–Weak hierarchy.Specifically, TeV ∼ ¯ M P e − kπr c , where ¯ M P is the reduced 4D Planck scale, k is the AdS curva-ture scale and r c is the proper size of the extra dimension. The crucial point is that the requiredmodest size of the radius (in units of the curvature radius), i.e., kr c ∼ /π log (cid:0) ¯ M P / TeV (cid:1) ∼ can be stabilized (i.e., the radion given a mass) with only a corresponding modest tuning in99he fundamental or 5D parameters of the theory [299, 300]. Remarkably, the correspondencebetween AdS and 4D conformal field theories (CFT) [301–303] suggests that the scenario withwarped extra dimension is dual to the idea of a composite Higgs in 4D [212, 304, 305]. It was realized that with SM fermions propagating in the extra dimension, we can also accountfor the hierarchy between quark and lepton masses and mixing angles (flavor hierarchy) asfollows [293, 294]: the basic idea is that the 4D Yukawa coupling are given by the product ofthe 5D Yukawa and the overlap of the profiles in the extra dimension of the SM fermions (whichare the zero-modes of the 5D fermions) with that of the Higgs. The light SM fermions can belocalized near the Planck brane, resulting in a small overlap with the TeV-brane localized SMHiggs, while the top quark is localized near the TeV brane with a large overlap with the Higgs.The crucial point is that such vastly different profiles for zero-mode fermions can be realizedwith small variations in the 5D mass parameters of fermions. Thus we can obtain hierarchicalSM Yukawa couplings without any large hierarchies in the parameters of the 5D theory, i.e. the5D Yukawas and the 5D masses.With SM fermions emerging as zero-modes of 5D fermions, so must the SM gauge fields.Hence, this scenario can be dubbed “SM in the (warped) bulk”. Due to the different profilesof the SM fermions in the extra dimension, flavor changing neutral currents (FCNC) are gener-ated by their non-universal couplings to gauge KK states. However, these contributions to theFCNC’s are suppressed due to an analog of the Glashow–Iliopoulos–Maiani (GIM) mechanismof the SM, i.e. RS–GIM, which is “built-in” [294, 306, 307]. The point is that all
KK modes(whether gauge, graviton or fermion) are localized near the TeV or IR brane (just like the Higgs)so that non-universalities in their couplings to SM fermions are of same size as couplings to theHiggs. In spite of this RS–GIM suppression, the lower limit on the KK mass scale can be − TeV [308–310] although these constraints can be ameliorated by addition of 5D flavor symme-tries [313–320]. Finally, various custodial symmetries [321, 322] can be incorporated such thatthe constraints from the various (flavor-preserving) electroweak precision tests (EWPT) can besatisfied for a few TeV KK scale [323, 324]. The bottom line is that a few TeV mass scale forthe KK gauge bosons can be consistent with both electroweak and flavor precision tests. Clearly, the light fermions have a small couplings to all KK’s (including graviton) based simplyon the overlaps of the corresponding profiles, while the top quark and Higgs have a large cou-pling to the KK’s. To repeat, light SM fermions are localized near the Planck brane and photon,gluon and transverse
W/Z have flat profiles, whereas all KK’s, Higgs (including longitudinal
W/Z ) and top quark are localized near the TeV brane. Schematically, neglecting effects relatedto electroweak symmetry breaking (EWSB), we find the following ratio of RS1 to SM gaugecouplings: g q ¯ q,l ¯ l A (1) RS g SM (cid:39) − ζ − ≈ − g Q ¯ Q A (1) RS g SM , g t R ¯ t R A (1) RS g SM (cid:39) to ζ ( ≈ See Refs [311] and [312] for “latest” constraints from lepton and quark flavor violation, respectively,i.e.,including variations of the minimal framework.
HHA (1) RS g SM (cid:39) ζ ≈ H = h, W L , Z L ) g A (0) A (0) A (1) RS g SM ∼ (1)Here q = u, d, s, c, b R , l = all leptons, Q = ( t, b ) L , and A (0) ( A (1) ) correspond to zero (firstKK) states of the gauge fields. Also, g xyz RS , g SM stands for the RS1 and the three SM (i.e., 4D)gauge couplings respectively. Note that H includes both the physical Higgs ( h ) and un physicalHiggs, i.e., longitudinal W/Z by the equivalence theorem (the derivative involved in this cou-pling is similar for RS1 and SM cases and hence is not shown for simplicity). Finally, theparameter ξ is related to the Planck–Weak hierarchy: ζ ≡ √ kπr c .We also present the couplings of the KK graviton to the SM particles. These couplingsinvolve derivatives (for the case of all SM particles), but (apart from a factor from the overlapof the profiles) it turns out that this energy-momentum dependence is compensated (or madedimensionless) by the ¯ M P e − kπr c ∼ TeV scale, instead of the ¯ M P -suppressed coupling to theSM graviton. Again, schematically: g q ¯ q,l ¯ l G (1) RS ∼ E ¯ M P e − kπr c × D Yukawa g A (0) A (0) G (1) RS ∼ kπr c E ¯ M P e − kπr c g Q ¯ Q A (1) RS , g t R ¯ t R G (1) RS ∼ (cid:18) kπr c to (cid:19) E ¯ M P e − kπr c g HHG (1) RS ∼ E ¯ M P e − kπr c (2)Here, G (1) is the KK graviton and the tensor structure of the couplings is not shown for simplic-ity. The unperturbed metric is written as: ds = (cid:18) Rz (cid:19) (cid:0) η µν dx µ dx ν − dz (cid:1) , (3)where z refers to the coordinate in the 5th dimension restricted to R < z < R (cid:48) , and R is theAdS curvature. The radion is related to the scalar perturbation of the metric, which at leadingorder is given by: δg MN = − F (cid:18) Rz (cid:19) (cid:18) η µν
00 2 (cid:19) (4)where F ( x, z ) is the 5D radion field.The linear radion couplings are determined by the modification of the action due to thelinear perturbation of the metric, which by the definition of the energy-momentum tensor isgiven by: δS = − (cid:90) d x √ gT MN δg MN = (cid:90) d x √ gF ( T rT MN − g T ) (5)101he canonically normalized scalar radion field in 4D is related to F ( x, z ) by: r ( x ) = Λ r (cid:18) R (cid:48) z (cid:19) F ( x, z ) (6)where Λ r = √ /R (cid:48) . For fields that are strongly localized in the infrared brane, such as theHiggs boson and the top quark, the coupling to the radion is given by the usual term L = r ( x )Λ r T µµ (7)For the top quark one has T ( t ) µν = i ¯ tγ µ ∂ ν t − η µν ¯ t ( iγ α ∂ α − m ) t (8)which implies L rtt = r Λ r m t ¯ tt. (9)However, for the Higgs boson the situation is complicated because of two factors: spon-taneous symmetry breaking and the fact that the energy-momentum tensor of a scalar field mustbe modified in order for its trace to vanish in the zero-mass limit, as it is required by conformalinvariance [329].For a Higgs lagrangian (after symmetry breaking) L h = 12 ( ∂ µ h ) − λ (cid:18) ( h + v ) − v (cid:19) (10)the modified energy-momentum tensor Θ µν reads: Θ µν = ∂ µ h∂ ν h − η µν L h + ξ (cid:0) η µν ∂ λ ∂ λ − ∂ µ ∂ ν (cid:1) (cid:18) ( h + v ) (cid:19) (11)which leads to Θ µµ = − (1 − ξ )( ∂ µ h ) + (1 − ξ )( λh + 4 λvh ) + (4 − ξ ) λv h − ξλv h. (12)Therefore, for ξ = 1 / , one gets Θ µµ = − λv h − λv h (13)where the first term of the trace of the modified energy-momentum tensor is proportional to theHiggs mass whereas the second term will induce a mixing between the radion and the Higgsboson.Radion phenomenology is very sensitive to the values of ξ . The ξ term for a general scalarfield φ can be written as a coupling to the Ricci scalar R as L ξ = ξRφ (14) Λ r ≈ KK scale, which can be varied by O (1) number. But canonical value is given by the above equation. ξ = 0 . Even if the Higgs is an approximate Goldstone boson, ξ should be small. Sincein this note we will be interested in the case where the radion mass is at least twice the Higgsboson mass, we will neglect the possibility of Higgs–radion mixing. In this case it follows that L rhh = r Λ r (cid:0) ( ∂ µ h ) − m h h (cid:1) (15)where the Higgs mass is m h = 2 λv .The leading contribution in the radion interaction with massive gauge bosons W ± and Z is given by L rV V = − r Λ r (cid:0) M W W + M Z Z (cid:1) (16)but there are model dependent corrections that we include in our analyses.Usually the coupling of the radion to massless gauge bosons vanishes at tree level. At1-loop it arises due to two contributions: the trace anomaly, which is related to the beta func-tion, and the top quark triangle diagram. However, in the warped scenario, there are two maindifferences: a tree level bulk contribution from radion and gauge bosons wave functions anda modification in the beta function term to take into account that only particles in the infraredbrane contribute to the running. The final result for this coupling is: L rAA = r r ln( R (cid:48) /R ) (cid:16) − πα (cid:16) τ (0) UV + τ (0) IR (cid:17) + (17) α π (cid:18) − − F ( τ w ) − F / ( τ t ) (cid:19) ln( R (cid:48) /R ) (cid:19) F µν F µν for photons and L rgg = r r ln( R (cid:48) /R ) (cid:16) − πα s (cid:16) τ (0) UV + τ (0) IR (cid:17) + (18) α s π (cid:18) − F / ( τ t ) (cid:19) ln( R (cid:48) /R ) (cid:19) G aµν G µνa for gluons where τ x = 4 m x /m r and the functions F , / ( τ ) vanishes when τ < . The parame-ters τ (0) UV and τ (0) UV are related to the Planck and TeV-brane induced kinetic terms. As indicated above, masses below about TeV for gauge KK particles are strongly disfavoredby precision tests, whereas masses for other KK particles are expected (in the general frame-work) to be of similar size to gauge KK mass and hence are (in turn) also constrained to beabove TeV. However, direct constraints on masses of other (than gauge) KK particles can beweaker. Radion mass can vary from ∼ GeV to ∼ TeV. In minimal models, KK gravitonis actually about . heavier than gauge KK modes, i.e., at least TeV.As far as KK fermions are concerned, in minimal models, they have typically massessame as (or slightly heavier than) gauge KK and hence are constrained to be heavier than TeV (in turn, based on masses of gauge KK required to satisfy precision tests). However, themasses of the KK excitations of top/bottom (and their other gauge-group partners) in some non-minimal (but well-motivated) models (where the 5D gauge symmetry is extended beyond thatin the SM) can be (much) smaller than gauge KK modes, possibly ∼ GeV.103
KK signals at the LHC
Based on these KK couplings and masses, we are faced with the following challenges in obtain-ing signals at the LHC from direct production of the KK modes, namely,(i) Cross-section for production of these states is suppressed to begin with due to a smallcoupling to the protons’ constituents, and due to the large mass of the new particles;(ii) Decays to “golden” channels (leptons, photons) are suppressed. Instead, the decays aredominated by top quark and Higgs (including longitudinal
W/Z );(iii) These resonances tend to be quite broad due to the enhanced couplings to top quark/Higgs.(iv) The SM particles, namely, top quarks/Higgs/
W/Z gauge bosons, produced in the decaysof the heavy KK particles are highly boosted, resulting in a high degree of collimationof the SM particles’ decay products. Hence, conventional methods for identifying topquark/Higgs/
W/Z might no longer work for such a situation.However, such challenges also present research opportunities – for example, several tech-niques to identify highly boosted top quark/Higgs/
W/Z have been developed [66–71, 73, 74,330–337]. Direct KK effects
Next, we summarize decay channels and production cross-sections for the KK particles: formore details, see corresponding references given in each title and for an overview, see Ref. [338].Based on the above discussion, note that the polarization of
W/Z ’s in these decay channels isdominantly longitudinal . Kaluza Klein partners of the gluon offer a particulary interesting phenomenology at the LHC.The cross-section of such coloured states can exceed that of typical electro-weak ( Z (cid:48) ) reso-nances by one or even several orders of magnitude. However, these states cannot be observedthrough the golden di-lepton resonance searches and discovery is only possible in the morechallenging hadronic final states.In this contribution, the focus is on the basic RS setup of Ref. [345] . In this model the KKgluon displays strongly enhanced couplings to (right-handed) top quarks. The most promisingsignature of the KK gluon is resonant t ¯ t production on top of the Standard Model t ¯ t continuum.The LHC (14 TeV) production rate of the pp → g KK → t ¯ t process ranges from nearly 30 pbfor a 1 TeV resonance to approximately 3 pb for a 3 TeV resonance.The KK gluon of the basic RS setup has a number of features that do not satisfy the usualassumptions of model-independent narrow resonance searches. A KK gluon search must takeinto acount the following:– The width of the KK gluon, 17 % of the mass in the basic RS setup, is not negligiblecompared to the experimental mass resolution. The model-independent limit for narrow resonances derived in the large majority of published t ¯ t resonance searches therefore doesnot apply. An experimental strategy must be developed to deal with the width explicitly. In Ref. [344] many different parameter sets for the KK gluon, each with a quite different phenomenology, arediscussed. mass (GeV)tt 400 600 800 1000 1200 1400 1600 / G e V - e v en t s /f b Figure 1:
The t ¯ t invariant mass distribution: Standard Model continuum (shaded histogram), the sumof SM and resonant production (dashed line) and the full interference of SM and resonant production(continuous line). – The interplay between the width of the resonance and the parton luminosity function leadsto a significant skew of the mass distribution of the pp → g KK → t ¯ t process. Especiallyfor large KK gluon mass a long tail towards lower mass develops. It is therefore non-trivial to relate an excess of events in a mass window to a total cross-section.– The interference between the resonant production and Standard Model t ¯ t production canbe significant. Figure 1 shows the difference between the full interference (continuousline) and the sum of signal and background processes (dashed line) for a generic, spin-1colour octet with a mass of 1 TeV and the couplings of the KK gluon implemented inMadGraph [7]. The interference leads to a pronounced reduction of the production ratefor M t ¯ t ∼ M g KK / .Therefore, while the KK gluon could be rather abundantly produced at the LHC, a com-plete experimental strategy for this type of broad coloured resonances is not yet fully developed(see, however, Contribution 13 in these proceedings). The dominant decay channels are into t ¯ t , W W , ZZ , hh . For a TeV KK graviton, each ofthese cross-sections can be ∼ O (10 fb ) with a total decay width of ∼ O (100 GeV ) . W (cid:48) [350] It turns out that in addition to KK W + L , these models also have a KK W + R (with no correspondingzero-mode), due to the custodial (i.e., extended 5D gauge) symmetry. These two KK states mixafter EWSB and the mass eigenstates are generically denoted by W (cid:48) . The dominant decaymodes for W (cid:48) are into W Z and
W h . For each W (cid:48) and with a mass of 2 TeV, the cross-sectionis ∼ O (10 fb) with a total decay width of ∼ O (100 GeV) . In some models, W (cid:48) decays to t ¯ b —giving boosted top and bottom — can also have similar cross-section. Interestingly, the processKK gluon → t ¯ t — with KK gluon mass being similar to W (cid:48) — can be a significant background105o this channel since a highly boosted top quark can fake a bottom quark: techniques similar tothe ones used to identify highly boosted tops can now be applied to veto this possibility! Z (cid:48) [351] There are actually three neutral KK states: KK Z , KK photon and a KK mode of an extra U (1) (again, with no corresponding zero-mode). These states mix after EWSB and the masseigenstates are generically denoted by Z (cid:48) . The dominant decay modes are to t ¯ t , W W and Zh ,each with a cross-section of ∼ O (10 fb) for a TeV Z (cid:48) with a total decay width of ∼ GeV. However, the t ¯ t channel can be swamped by KK gluon → t ¯ t if the Z (cid:48) and KK gluon havesimilar mass. The KK fermions in the minimal model being TeV or heavier, even single production of theseparticles can be very small (pair production is even smaller).
As mentioned above, in non-minimal models, KK partners of top/bottom can be light so thattheir production (both pair and single, the latter perhaps in association with SM particles) canbe significant. As these particles are “top-like" with respect to their production at the LHC,the yields can be sizeable. For example, the pair production cross-section of a KK bottomwith mass of
GeV is ∼ pb at √ s = 10 TeV. These particles decay into t/b + W/Z/h ,where the
W/Z can be boosted at the LHC (even for fermionic KK partners with masses aslow as ∼ GeV). Some of these light KK fermions can have “exotic" electric charges – forexample, / and / . This makes them appealing with respect to a generic b (cid:48) /t (cid:48) from, forexample, a minimal extension to SM generations [355]. Recently, Tevatron experiments haveplaced limits on such KK fermions [356]. Various search strategies for KK fermions are beingdeveloped at the LHC [352–354] (see also Contribution 11 in these proceedings).In addition, the other heavier (spin-1 or 2) KK modes can decay into these light KKfermions, resulting in perhaps more distinctive final states for the heavy KK’s than the pairs of W/Z or top quarks that have been studied so far – for such a study for KK gluon, see Ref. [357].
Radion production at the LHC could be substantial due to the fact that the branching fractionof the radion to two gluons could be enhanced by as much as a factor of 10 (for Λ r = 1 TeV)in comparison with the Higgs branching fraction to gluons. The enhancement is due to thefact that the radion couples to massless gauge bosons through the conformal anomaly, which israther large for QCD. As a bona fide dilaton, the radion couples to the energy-momentum tensorof the theory. Hence, its couplings are proportional to masses of particles, in much the sameway as the usual Higgs boson. As mentioned above, radion mass is a free parameters of thetheory, varying from ∼ GeV to ∼ TeV, which means that dominant decay channels aredetermined by radion mass. For radion mass lighter than M W , r → γγ is a promising channel,which can be also dramatically enhanced in the presence of Higgs–radion mixing. For largerradion mass, W W , hh , ZZ , t ¯ t channels are the dominant channels, which can pose a challengefor detecting highly boosted signals. 106 (R) (GeV)
400 500 600 700 800 900 1000 W + W -) ( pb ) ! R ! ( gg " KK R, CTEQ6M, 10 TeV
Figure 2: σ ( gg → r → W W ) cross section at the LHC for 10 TeV center of mass energy. C r o ss s e c t i on ( f b ) M R R->HH 14 TeV10 TeV
Figure 3: σ ( pp → r → HH ) cross section at the LHC for 14 and 10 TeV center of mass energies. Above a radion mass of 400 GeV or so, where decay products of radions can start to beboosted, the branching fractions of radion into SM particles are reasonably flat. Dependingon model parameters, the
W W channel can be the most dominant channel with a branchingfraction of about 50%. Figure 2 shows
W W cross section of radions as a function of radionmass at 10 TeV LHC center of mass energy using the CalcHEP implementation of Ref. [328].It can be seen that even for a high value of Λ r at 3 TeV, the cross sections can be as high asa fraction of a picobarn. Reach prospects improve when a value of 2 TeV for Λ r is chosen,as allowed by the EWPT results. The largest yield in W W channel would come from fullyhadronic decays of the W boson, however, this channel may suffer largely from QCD dijetproduction at the LHC. Looking at the semi-leptonic channel, as was done for W W scatteringsearches at ATLAS [66, 163, 337], may provide a way to observe radion production in
W W channel. For example, for a radion mass of 600 GeV, the σ ( r → W had W lep ) is ∼ O (10 fb) , for Λ r = HH channel for the samesettings as before in the W W channel. 107
Identification of boosted objects from KK direct production
Motivated by above discussion of signals for KK particles in warped extra dimensional frame-work, we study in this section the identification of boosted SM particles which decay. W and Z decay products The identification of W and Z decays products from the models discussed will be experimen-tally challenging due to the boosted nature of the decaying system. For available LHC energies,the angular separation in the lab frame of the decay products will be of the order . .For decays to e , µ , ν , this hampers traditional reconstruction techniques which rely onisolated leptons in order to reject jet backgrounds and to clean fake E/ T . For hadronic decays,the two decaying quarks will merge into one collimated jet. It is possible that by exploring jetsubstructure, these will be identifiable with backgrounds under control. Studies in that directionhave already been performed and discussed elsewhere (see, e.g., Ref. [163]), thus here we onlyconcentrate on the leptonic decays. Z The main challenge in identifying boosted Z → e + e − will be the merging of electromagneticclusters. The granularity of typical LHC calorimetry is such that this will be an algorithmicrather than a physical issue. In particular, algorithms designed to recover energy lost due toBrehmstrahlung radiation may be detrimental to boosted Z identification.The results of a toy Monte Carlo simulation of boosted Z → e + e − , assuming a efficiency to identify a single electron, are shown in Fig. 4. Within typical LHC detector ac-ceptance (electron acceptance is taken to be in the region | η | < . ), identification ispossible for centrally boosted Z ’s with high relativistic γ . At high energy, the energy resolutionis dominated by the constant term, and as such resolutions of the order − can be expected.Existing background rejection methods, such as the jet fake rate, developed for non-boosteddecays of heavy neutral particles to di-lepton pairs will be equally applicable to the boostedreconstruction scenario. Figure 4:
Toy Monte Carlo simulation of boosted Z → e + e − identif ication In the µ + µ − channel, angular separation is not an issue, however the momentum mea-surement will be affected by the low-curvature tracks. CMS and ATLAS expect a momentum108esolution of order for TeV muons [358, 359]. W Where a W decays leptonically, W → eν e , W → µν µ is also of interest to the models discussed.Such a decay leads to significant E/ T , correlated with the electron (muon) direction. This allowsthe W mass to be reconstructed in the collinear approximation, where the neutrino three-vectoris defined as (cid:126)p ν e = ( (cid:54) E x , (cid:54) E y , (cid:112) (cid:54) E x + (cid:54) E y (cid:112) p x,e + p y,e p z,e ) , (19)where (cid:126)p e is the electron momentum. The neutrino four-vector is defined as p ν e µ = ( (cid:126)p ν e , | (cid:126)p ν e | ) .Plotting the electron-neutrino invariant mass against the angle in φ between the electronand E/ T provides a powerful discriminant between signal and background, as shown in Fig. 5 forevents simulated with Pythia [271] and PGS [360]. The signal is a excited quark, whichcan be taken as producing a generic boosted W with momentum near
500 GeV . A cut in the2D plane of ∆ φ < M W,col /c with c = 100 yields a boosted W identification efficiency of and a t ¯ t rejection of . Further study and tuning is needed with full detector simulation, butit appears that powerful signal selection and background rejection is possible (Fig. 6). (a) (b)(c) (d) Figure 5:
Discriminating boosted W ± s from background for signal (a), W + Jets (b), t ¯ t (c) and Z → e + e − (d) igure 6: Acception / rejection efficiencies for signal and background, varying the constant term c in the2D-plane cut. p T top reconstruction CDF and D0 have performed extensive t ¯ t resonance searches [361, 362] and a δσδM t ¯ t measure-ment [363]. No deviations from the Standard Model prediction have been observed and limitsare derived for several models.At the Tevatron, the large majority of t ¯ t pairs are produced essentially at rest. The t ¯ t pairwith the largest invariant mass is registered with approximately 1 TeV. At 14 TeV, in 20 % of t ¯ t produced, one of the top quarks has a transverse momentum greater than 200 GeV . The LHCwill be able to explore the t ¯ t mass spectrum into the several TeV regime.The reconstruction of highly boosted top quarks is an experimental challenge. The topquark decay products are collimated in a narrow cone. The hadronic decay products oftencannot be individually resolved by jet algorithms. The isolation of the leptons from W -decayis broken by the neighbouring b-jet. A number of references in the literature [66–71, 73, 74,330–337] have addressed this issue proposing a new approach, where top decays (and similarlyW/Higgs decays) are reconstructed as a single jet. A number of techniques has been developedthat allow to identify (tag) these top mono-jets.Recent CMS [366, 367] and ATLAS [368–371] studies have implemented these ideasand established their performance on fully simulated signal and background events. Thesetechniques are indeed found to offer greatly improved top quark reconstruction efficiency, whilemaintaining an adequate reduction of non- t ¯ t backgrounds (primarily W+jets and QCD di-jetproduction). Thus, the sensitivity of t ¯ t resonance searches is improved with respect to thatobtained with classical reconstruction techniques. Indirect KK effects
In addition to signals from the direct production of the KK particles at the LHC, there can alsobe effects of these KK particles on the properties of the SM particles themselves. Estimate obtained using MC@NLO [364, 365] .1 Flavor-violating Higgs/Radion couplings
Higgs flavor-violation induced by KK particles in the warped extra-dimensional framework isdiscussed in Refs. [372,373]. Estimates are BR ( t → ch ) ∼ − and BR ( h → tc ) ∼ × − :see Ref. [373] for more detailed numbers. Note that the radion also decays to tc (very similarlyto Higgs): see Ref. [374]. Reference [375] claims LHC sensitivity of × − for BR ( t → cH ) (obviously for m h < m t ) (see also Ref. [376] for more details). A reference for a study of h → tc — obviously for m h > m t — at a similar level of detail could not be found.We suggest performing a detailed study of LHC sensitivity for tch coupling for the case m h > m t , i.e., when t → ch is not allowed. One method is via Higgs decays : the Higgs can beproduced via gluon fusion or by W W fusion (in the latter case, we can tag forward jets). SeeRefs. [373] and [374] for first steps toward this goal (including some analysis of background).[Reference [377] studied flavor-violating Higgs decays to top in a different framework (2-Higgsdoublet model), but without any analysis of background].Another option is to use the tch coupling to produce the Higgs, for example, gc → th (see Ref. [376] for a study of this channel, but using h → b ¯ b , whereas here we would like to use h → W W/ZZ since we have m h > m t ).In both directions mentioned above, a starting point might be to use existing studies ofrelated channels in SM (or its extensions) in order to see how background was reduced – forexample, gb → tH + in -Higgs doublet models vs. gc → th here or single top production inSM vs. gg → h → tc here.Finally, a leptonic (and thus cleaner) channel: BR ( h → µτ ) can be large in this frame-work (see Ref. [373]) which might be within the LHC reach (see Ref. [378]). t → ch at LHC Even though the LHC sensitivity for observing the flavor violating decay of top quark, t → ch (when m h < m t ), has been studied in detail in Ref. [376], we think it is useful to re-visit thisanalysis which is the goal of this section. Specifically, we focus on the following new aspects:i) optimizing cuts to improve the sensitivity, (ii) tagging charm quark, motivated by the factthat since typically t → ch dominates over t → uh , the signal under consideration contains acharm quark and (iii) considering m h = 160 GeV so that h → b ¯ b is very small and h → W W dominates (note that only the cases m h = 110 , GeV were studied in Ref. [376] such thatthe dominant decay mode h → b ¯ b was used).The signal at the LHC arises from pp → t ¯ t → bW ch → bW cW W leading to (cid:96)bc jE/ T events, where (cid:96) = e or µ and j = u, d, c, s is from W decays. We allowed all the three W -bosons decay into all possible channel. The effective cross section for this signal topology canbe expressed as, C (cid:96)bc j = σ SM ( pp → t ¯ t ) BR ( t → ch ) BR ( t → bW ) BR ( h → W W ) , (20)where W stands for W ± . We consider m t =175 GeV, m h = 160 GeV and BR( t → ch )= − . TheSM backgrounds with the similar signal topology arises from many reducible and irreduciblesources. However, for the present study we considered the dominant two, namely, t ¯ t and t ¯ tb ¯ b .In our signal simulation we used the PYTHIA v6.408 event generator [271]. The
SLHA [379]input is used to provide the flavor violating branching ratios of the top quark. for generatingparton level SM backgrounds, we used
MadGraph/MadEvent v4.4.15 [7], and we later fedthem to
PYTHIA for showering. The backgrounds events were generated with the following111reselection kinematical cuts: p j,bT > ∼ GeV; η j,b < ∼ . ; ∆ R ( jj, bb, bj ) > ∼ . . We set the renor-malization and factorization scale to Q = √ ˆ s and used CTEQ5L for the parton distributionfunctions (PDF). All the masses and mass parameters are given in GeV.We simulate our signal and backgrounds at the LHC for 14 TeV center of mass energybased on the following assumptions:– The ATLAS [24] calorimeter coverage is | η | < . ;– The segmentation is ∆ η × ∆ φ = . × . (i.e., approximately ∆ R = 0 . ) whichresembles the ATLAS detector;– The toy calorimeter, PYCELL , provided in
PYTHIA for the jet reconstruction. The totalenergy of jets and leptons are smeared according to Gaussian distribution. The energyresolution is taken as ∆ E j,(cid:96) E j,(cid:96) = 50% (cid:112) E j,(cid:96) ⊕
3% ; (21)We reconstructed the missing energy ( E/ T ) from smeared observed particles. We have notincluded any real detector effects in our simulations;– The showering scales are the following: for ISR and FSR we multiplied the hard scatter-ing scale, Q , which we set as f × ˆ s , where f =4.0 ;– A cone algorithm with ∆R(j , j) = (cid:112) ∆ η + ∆ φ ≥ . has been used for jet finding ;– The E cellT , min ≥ . is considered to be a potential candidate for jet initiator. The cell with E cellT , min ≥ . is treated as a part of the would be jet and minimum summed E jT , min ≥ . is accepted as a jet and the jets are ordered in E T ;– Leptons ( (cid:96) = e , µ ) are selected with E (cid:96) T ≥ . and | η (cid:96) | ≤ . ;– We have implemented jet and lepton ( (cid:96) = e or µ ) isolation using the following criteria:if there is a jet within the vicinity of the partonic lepton ( (cid:96) p ) with ∆R(j − (cid:96) p ) ≥ . and . ≤ E jT / E (cid:96) p T ≤ . , the jet is removed from the list of jets and treated as a lepton, elsethe lepton is removed from the list of leptons;– b-tagging: A jet with | η j | ≤ . matched with a b − flavored hadron B , i.e., with ∆ R ( j, B ) < . , is considered to be b-taggable . We imposed the b tagging in thesetaggable jets with probability (cid:15) b =0.50;– c-tagging: A jet with | η j | ≤ . matched (similar to B-Hadron) with a C − flavored hadron C − hadron (e.g., D-meson, Λ c -baryons), i.e., with ∆ R ( j, C − hadron ) < . , is con-sidered to be c-taggable . We imposed the c tagging in these taggable jets with probability (cid:15) c =0.10;– b-mis-tagging: Jets other than b -taggable/tagged and c -taggable/tagged are matched withthe light flavor parton (q = u,d,s,g and τ with minimum ∆ R ( j − q ) and ≤ E jT ≥ and η j ≤ . then the jet is treated as a mis-taggable jets with the flavor similar to thematched parton, q . If a jet does not match with any parton in the event we considerthis jet as a gluon-jet originating from the secondary radiation. The jets are mis-taggedby generating random numbers according to the flavors; we considered (cid:15) u,d,s,g =0.0025following the recent ATLAS analysis [163, 380] and [381] . It is important to note that themis-tagging rate can be known precisely once we have the real LHC data. Unlike jet-lepton matching we considered only the minimum ∆ R ( j, B ) and not the E T ratios. The τ -lepton is considered to be a parton in our analysis with nearly zero mis-tagging probability. b ≥ N c ≥ N tot ≥ Process EvtSim 1 2 3 1 2 3 1 2 3 m h =160 100000 .4267 .0063 .0007 .1090 .0049 .0001 .4984 .0641 .0051 t ¯ t t ¯ tb ¯ b Table 1:
The Individual efficiencies for purely b -tagged ( N b ), purely c -tagged ( N c ) and with the inclusionof low flavor mis-tagged ( N tot ) at LHC. EvtSim stands for number of event simulated. Process EvtSim C1 C2 C3 C4a C4b C4 m h =160 100000 .719 .357 .790 .427 .109 .064 t ¯ t t ¯ tb ¯ b Table 2:
The individual efficiencies of various kinematical selections for signal and backgrounds at LHC.EvtSim stands for the number of event simulated. See text for the numerical values of the kinematicalselections.
We need to retain as many signal events as possible and at the same time suppress thebackgrounds to a large extent by applying different kinematical selection. In doing so we intro-duce the following kinematical selection:– C1: N jet ≥ , E j =1 − T > . and | η j =1 − | < . ;– C2: N lepton ≥ , E (cid:96)T > . and | η (cid:96) | < . ;– C3: E/ T > where E/ T is calculated from all visible particles;– C4a: N b − tag ≥ ; | η b − jet | < . , ∆ R ( j, B ) ≤ . ;– C4b: N c − tag ≥ ; | η c − jet | < . , ∆ R ( j, C − hadron ) ≤ . ;– C4(with Mis-tagging from the light quarks and gluon) : N tot = N ( b + c ) − tag + q − mistag ≥ .The individual efficiencies for N b − tag , N c − tag and N tot are given in Table 1. As expected, N c − tag efficiencies for Signal is larger than t ¯ t and t ¯ tb ¯ b . We have also shown the individualefficiencies for number of jets, number of lepton and missing energy ( E/ T ) in Table 2.To ensure the flavor violating decay of top quark we reconstructed the W -boson, Higgsboson and top quark masses. In order to suppress the huge backgrounds, before mass recon-struction, at the first step we applied the basic acceptance cuts, JLM = C1 ⊗ C2 ⊗ C3. Werequired one c -tagged (C4b) events to suppress more background and can be seen from Table 3.Finally, we consider events with at least two tagged jet (C4) for the mass reconstruction.We show the cumulative number of events after applying some combined selections inTable 3. The combined selections are the following:– JLM: C1 ⊗ C2 ⊗ C3;– JLMNc: JLM ⊗ C4b;– W -reco: JLMNc ⊗ C4; 113rocess RawEvt JLM JLMNc W -reco W h t m h =160 8000. 1190. 138.5 66.6 63.2 47.4 6.93 t ¯ t t ¯ tb ¯ b Table 3:
The cumulative events for Signal and Backgrounds survived after different combination ofselection criterion at the LHC for 100 f b − integrated luminosity. RawEvt stands for the number ofevents produced in reality. W -reco stands number of events for the combined selections: C1 ⊗ C2 ⊗ C3 ⊗ C4b ⊗ C4. W , h and t represent the selection on the reconstructed masses of W , Higgsboson and top quark, see text for details. m W (GeV) d N / d m W / G e V Sx1000t t – t t – b b – x10 Figure 7:
The reconstructed W-boson mass ( m W ) for signal and backgrounds. The distribution is nor-malized to t , see the last column in Table 3. Signal and t ¯ tb ¯ b are scaled with 1000 and 10 respectively. – W : W -reco ⊗ m W ±
70 GeV;– h : W ⊗ m h ±
100 GeV;– t : h ⊗ m t ±
100 GeVWe calculated all the possible di-jet invariant mass ( m jj ) without considering the pure b -tagged jets ( since BR( W → ¯ bu ( c ) ) approximately O (10 − ) ). The pair of jets for recon-structing each m W were selected by minimizing | m j j − m j j | . The reconstruction of m h isthen straightforward, i.e., m h = m j j j j . Furthermore, we reconstruct the top quark mass toensure the flavor violating decay. We consider the remaining jets, without pure b -tagged jets(to ensure the flavor violating decay), combined with the selected four jets ( m h candidates); byminimizing | m j j j j j − m t | . After mass reconstructions, we applied t selection and showthe reconstructed masses for W , H and top in Fig. 7, 8 and 9 respectively . It can be seen fromTable.3 that the number of signal (total background) event is approximately 7(9000).Our preliminary analysis shows that the number of signal and total background events We scaled the signal ( t ¯ tb ¯ b ) distribution by 1000 (10) in all the figures. h (GeV) d N / d m h / G e V Sx1000t t – t t – b b – x10 Figure 8:
The reconstructed Higgs boson mass ( m h ) for signal and backgrounds. The distributionis normalized to t , see the last column in Table 3. Signal and t ¯ tb ¯ b are scaled with 1000 and 10respectively. m t (GeV) d N / d m t / G e V Sx1000t t – t t – b b – x10 Figure 9:
The reconstructed top quark mass ( m t ) for signal and backgrounds. The distribution is nor-malized to t , see the last column in Table 3. Signal and t ¯ tb ¯ b are scaled with 1000 and 10 respectively. t → ch . With luminosity upgrade,one could get more signal events, however, the backgrounds will also be large. One has todesign more clever selection to reject t ¯ t backgrounds. Intuitively, a slightly different approachwhile reconstructing the top mass might be useful, for example, considering explicit c -jet (i.e., j candidate jet in our present analysis). Of course, an analysis by LHC experimental groupsof sensitivity for t → ch is desirable. We are aware that ATLAS group is already undertakingsuch a study. Summary
In this note, we have given an overview of LHC signals for the very well-motivated frameworkof SM particles propagating in a warped extra dimension. We have also presented some new results of (or directions for) such studies, for example, identification of boosted
W/Z ’s andtop quarks, t → ch and radion production. It is worth pointing out that some of these studiesmight also be relevant in searching for other types of new physics, for example, other modelsbeyond the SM can also contain heavy particles decaying into top quarks, W/Z or can give riseto sizable flavor-violating tc Higgs coupling.
Acknowledgements
The work of K.A. is supported by the NSF grant No. PHY-0652363.116 ontribution 10 Z (cid:48) discovery potential at the LHC in the minimal B − L model L. Basso, A. Belyaev, S. Moretti, G.M. Pruna and C.H. Shepherd-Themistocleous
Abstract
We present the Large Hadron Collider (LHC) discovery potential in the Z (cid:48) sector of a U (1) B − L enlarged Standard Model for √ s = 7 and TeV centre-of-mass (CM) energies, considering both the Z (cid:48) B − L → e + e − and Z (cid:48) B − L → µ + µ − decay channels. Electrons provide a higher sensitivity to smaller cou-plings at small Z (cid:48) B − L masses than do muons. The resolutions achievable mayallow the Z (cid:48) B − L width to be measured at smaller masses in the case of elec-trons in the final state. The run of the LHC at √ s = 7 TeV, assuming at most (cid:82)
L ∼ fb − , will be able to give similar results to those that will be availablesoon at the Tevatron in the lower mass region, and to extend them for a heavier M Z (cid:48) . A run at TeV is needed to fully probe the parameter space. If noevidence is found in any energy configuration,
C.L. limits can be deter-mined, and, given their better resolution, the limits from electrons will alwaysbe more stringent than those from muons. Introduction
The evidence for non vanishing (although very small) neutrino masses is so far possibly theonly hint for new physics beyond the Standard Model (SM) [382, 383]. It is noteworthy thatthe accidental U (1) B − L global symmetry is not anomalous in the SM with massless neutrinosthough its origin is not well understood. It thus becomes appealing to extend the SM to simul-taneously explain the existence of both (i.e., neutrino masses and the B − L global symmetry)by gauging the U (1) B − L group thereby generating a Z (cid:48) state. This requires that the fermion andscalar spectra are enlarged to account for gauge anomaly cancellations. The results of directsearches constrain how this may be done [384–387]. Minimally, this requires the addition ofa scalar singlet and three right-handed neutrinos, one per generation [388–390], which couldtrigger the see-saw mechanism explaining the smallness of the SM neutrino masses [391–395].Within this model, the masses of the heavy neutrinos are such that their discovery falls withinthe reach of the LHC over a large portion of parameter space [34, 396].In general, studies of this model focus on a specific non-disfavoured point in the parameterspace and do not preform a systematic analysis of the entire space. The Z (cid:48) B − L boson is also notalways considered as a traditional benchmark for generic collider reach studies [162, 397–401]or in data analyses [386, 387]. We have therefore performed a (parton level) discovery potentialstudy for the LHC in the Z (cid:48) sector of the B − L model. In the light of the LHC plan of actionover the next few years [402], we consider the CM energies of and TeV, with integratedluminosities up to fb − for TeV, and up to fb − for TeV. We also include a comparisonwith the Tevatron reach for its expected fb − of integrated luminosity. We chose to study the117i-lepton channel (both electrons and muons), the cleanest and most sensitive Z (cid:48) boson decaychannel in our model at colliders.This work is organised as follows. Section 2 describes the B − L model under consider-ation. Section 3 illustrates the computational techniques adopted. The results are presented inSection 4 for the Z (cid:48) boson sector and finally the conclusions are given in Section 5. The Model
The model under study is the so-called “pure” or “minimal” B − L model (see Ref. [34, 390]for conventions and references) since it has vanishing mixing between the U (1) Y and U (1) B − L gauge groups. In the rest of this paper we refer to this model simply as the “ B − L model”.This work focuses on the extended gauge sector of the model, whose Abelian Lagrangian canbe written as follows: L Abel
Y M = − F µν F µν − F (cid:48) µν F (cid:48) µν , (1)where F µν = ∂ µ B ν − ∂ ν B µ , (2) F (cid:48) µν = ∂ µ B (cid:48) ν − ∂ ν B (cid:48) µ . (3)In this field basis, the covariant derivative is: D µ ≡ ∂ µ + ig S T α G αµ + igT a W aµ + ig Y B µ + i ( (cid:101) gY + g (cid:48) Y B − L ) B (cid:48) µ . (4)The “pure” or “minimal” B − L model is defined by the condition (cid:101) g = 0 , that implies no mixingbetween the Z (cid:48) B − L and SM Z gauge bosons.The fermionic Lagrangian (where k is the generation index) is given by L f = (cid:88) k =1 (cid:16) iq kL γ µ D µ q kL + iu kR γ µ D µ u kR + id kR γ µ D µ d kR ++ il kL γ µ D µ l kL + ie kR γ µ D µ e kR + iν kR γ µ D µ ν kR (cid:17) , (5)where the fields’ charges are the usual SM and B − L ones (in particular, B − L = 1 / for quarksand − for leptons with no distinction between generations, hence ensuring universality). The B − L charge assignments of the fields as well as the introduction of new fermionic right-handedheavy neutrinos ( ν R ’s) and a scalar Higgs field ( χ , with charge +2 under B − L ) are designedto eliminate the triangular B − L gauge anomalies and to ensure the gauge invariance of thetheory, respectively. Therefore, a B − L gauge extension of the SM gauge group broken atthe TeV scale requires at least one new scalar field and three new fermionic fields which arecharged with respect to the B − L group.An important feature of the Z (cid:48) gauge boson in the B − L model is the chiral structure ofits couplings to fermions: since the B − L charges do not distinguish between left-handed andright-handed fermions, the B − L neutral current is purely vector-like, with a vanishing axialpart . As a consequence, we do not study the asymmetries of the decay products stemming That is, g VZ (cid:48) = g LZ (cid:48) + g RZ (cid:48) , g AZ (cid:48) = g RZ (cid:48) − g LZ (cid:48) , hence g RZ (cid:48) = g LZ (cid:48) . Z (cid:48) B − L bosons, given that their distribution is trivial in the peak region which is studiedhere. However, asymmetries do become important in the interference region, especially justbefore the Z (cid:48) boson peak, where the Z − Z (cid:48) interference will effectively provide an asymmetricdistribution somewhat milder than the case in which there is no Z (cid:48) boson. This is a powerfulmethod of discovery and identification of a Z (cid:48) and it will be reported on separately [403].The scalar and Yukawa sectors of the model play no relevant role in this analysis, thereforewe refer to Ref. [404] for a more detailed overview of the model . Computational details
The study we present in this paper has been performed using the CalcHEP package [407]. Themodel under discussion has previously been implemented in this package using the LanHEPtool [408], as discussed in Ref. [34].The process we are interested in is di-lepton production. We define our signal as pp → γ, Z, Z (cid:48) B − L → (cid:96) + (cid:96) − ( (cid:96) = e, µ ), i.e., all possible sources together with their mutual interfer-ences, and the background as pp → γ, Z → (cid:96) + (cid:96) − ( (cid:96) = e, µ ), i.e., SM Drell-Yan production(including interference). No other sources of background, such as W W , ZZ , W Z or tt , havebeen taken into account. These can be suppressed or/and are insignificant [162]. For boththe signal and background, we have assumed standard acceptance cuts (for both electrons andmuons) at the LHC: p lT >
10 GeV , | η l | < . l = e, µ ) , (6)and we apply the following requirements on the di-lepton invariant mass, M ll , depending onwhether we are considering electrons or muons. We distinguish two different scenarios: an“early” one (for √ s = 7 TeV) and an “improved” one (for √ s = 14 TeV), and, in computingthe signal significances, we will select a window as large as either the width of the Z (cid:48) B − L bosonor twice the di-lepton mass resolution , whichever is the largest. The windows in the invariantmass distributions respectively are, for the “early scenario”electrons: | M ee − M Z (cid:48) | < max (cid:18) Γ Z (cid:48) , (cid:18) . M Z (cid:48) GeV (cid:19)
GeV (cid:19) , (7)muons: | M µµ − M Z (cid:48) | < max (cid:18) Γ Z (cid:48) , (cid:18) . M Z (cid:48) GeV (cid:19)
GeV (cid:19) , (8)and for the “improved scenario”electrons: | M ee − M Z (cid:48) | < max (cid:18) Γ Z (cid:48) , (cid:18) . M Z (cid:48) GeV (cid:19)
GeV (cid:19) , (9)muons: | M µµ − M Z (cid:48) | < max (cid:18) Γ Z (cid:48) , (cid:18) . M Z (cid:48) GeV (cid:19)
GeV (cid:19) . (10)Our choice reflects the fact that what we will observe is in fact the convolution between theGaussian detector resolution and the Breit-Wigner shape of the peak, and such a convolution Although they do not modify the Z (cid:48) boson properties significantly, for completeness we state the chosen heavyneutrino and the scalar masses and the scalar mixing angle: m ν h = m ν h = m ν h = 200 GeV (value that can leadto interesting phenomenology [34]), m h = 125 GeV, m h = 450 GeV and α = 0 . (allowed by a preliminarystudy on the unitarity bound [405], as well as on the triviality bound [406] of the scalar sector). We take the CMS di-electron and di-muon mass resolutions [359] as representative of a typical LHC environ-ment. ATLAS resolutions [358] do not differ substantially. p lT >
18 GeV , | η l | < l = e, µ ) , (11)and the following requirements on the di-lepton invariant mass, M ll , depending on whether weare considering electrons or muons :electron: | M ee − M Z (cid:48) | < max (cid:32) Γ Z (cid:48) , (cid:32) . (cid:114) M Z (cid:48) GeV GeV + 0 . M Z (cid:48) GeV (cid:33)
GeV (cid:33) , (12)muons: | M µµ − M Z (cid:48) | < max (cid:32) Γ Z (cid:48) , (cid:32) . (cid:18) M Z (cid:48) (cid:19) (cid:33) GeV (cid:33) . (13)In our analysis we also use a definition of the signal significance σ , as follows. In theregion where the number of both signal ( s ) and background ( b ) events is “large” (here taken tobe bigger than 20), we use a definition of significance based on Gaussian statistics: σ ≡ s / √ b . (14)Otherwise, in case of smaller statistics, we used the Bityukov algorithm [410], which basicallyuses the Poisson ‘true’ distribution instead of the approximate Gaussian one.Finally, as in [34, 411], we used CTEQ6L [412] as the default Parton Distribution Func-tions (PDFs), evaluated at the scale Q = M ll . Only the irreducible SM Drell-Yan backgroundhas been considered. Reducible backgrounds, ISR, photon-to-electron conversion etc. wereneglected. Z (cid:48) Boson Sector: Results
In this section we determine the discovery potential and we present exclusion plots for the LHC.We use centre-of-mass (CM) energies of and TeV and relevant integrated luminosities.The experimental constraints come from LEP and the Tevatron. For the B − L model, themost recent limit from LEP [385] is: M Z (cid:48) g (cid:48) ≥ . (15)The most recent limits from the Tevatron for the Z (cid:48) B − L boson (from the CDF analyses ofRef. [386, 387] using . fb − and . fb − of data for electrons and muons in the final state,respectively), are shown in table 1 (for selected masses and couplings).The production cross sections for the process pp ( p ) → Z (cid:48) B − L for g (cid:48) = 0 . are shown inFig. 1. Note that although at the Tevatron the production cross section is smaller than at theLHC, the integrated luminosity considered here for the LHC at √ s = 7 TeV (i.e. 1 fb − ) issmaller than for the Tevatron (i.e. 10 fb − ). We take the CDF di-electron and di-muon mass resolution [409] as respresentative of a typical Tevatronenvironment. p → e + e − pp → µ + µ − g (cid:48) M Z (cid:48) (GeV) g (cid:48) M Z (cid:48) (GeV)0.042 600 0.06 6000.086 700 0.1 7500.115 800 0.123 8000.19 900 0.2 9000.3 1000 0.3 1000- - 0.5 1195 Table 1:
Lower bounds on the Z (cid:48) mass for selected g (cid:48) values in the B − L model, at C.L., bycomparing the collected data of Ref. [386, 387] with our theoretical prediction for pp → Z (cid:48) B − L → e + e − ( µ + µ − ) at the Tevatron. LHC ( ! s=7 TeV)LHC ( ! s=10 TeV)LHC ( ! s=14 TeV)TevatronM Z' (TeV) " ( p p ( pb a r ) Z ' ) ( pb ) -5 -4 -3 -2 -1 Figure 1:
Cross sections for pp ( p ) → Z (cid:48) B − L at the Tevatron and at the LHC (for √ s = 7 , and TeV) for g (cid:48) = 0 . . √ s = 7 TeV
Initial LHC running will be at a CM energy of TeV, where the total integrated luminosity islikely to be of the order of fb − . Figure 2 shows the discovery potential under these conditions.In the same figure we also include for comparison the Tevatron discovery potential at the inte-grated luminosities used for the latest published analyses ( . fb − [386] and . fb − [387] forelectrons and muons, respectively) as well as the expected reaches at L = 10 fb − . Ref. [404]where a comparision to Tevatron data is shown, one can see that our parton level simulationreproduces experimental conditions reasonably well.At this stage of the LHC, the Tevatron will still be competitive, especially in the lowermass region where the LHC requires fb − to be sensitive to the same couplings as the Tevatron.The LHC will be able to probe the Z (cid:48) B − L for values of the coupling down to − · − (for121lectrons and muons respectively), while the Tevatron can be sensitive down to − · − . -2 -1 ElectronsTevatron, ! s=1.96 TeV5 " , L = 10 fb -1 " , L = 10 fb -1 " , L = 2.5 fb -1 " , L = 2.5 fb -1 LHC, ! s=7 TeV5 " , L = 100 pb -1 " , L = 100 pb -1 " , L = 1 fb -1 " , L = 1 fb -1 M Z' (TeV) g ' Figure 2:
Significance contour levels plotted against g (cid:48) and M Z (cid:48) at the LHC for √ s = 7 TeV and . − fb − and at the Tevatron ( √ s = 1 . TeV) for (left, electrons) . − fb − and (right, muons) . − fb − of integrated luminosity. The shaded areas correspond to the region of parameter space excludedexperimentally in accordance with Eq. (15) (LEP bounds, in black) and table 1 (Tevatron bounds, in red). Figure 3:
Integrated luminosity required for observations at 3 σ and 5 σ vs M Z (cid:48) for selected values of g (cid:48) for electrons (left) at the LHC for √ s = 7 TeV and (right) at the Tevatron ( √ s = 1 . TeV). Onlycombinations of masses and couplings not yet excluded are shown. Similar plots for muons in the finalstate are in Ref. [404].
Figure 3 shows the integrated luminosity required for σ evidence and σ discovery as122 function of the Z (cid:48) B − L boson mass for selected values of the coupling for the electron fi-nal state, both at the LHC and at the Tevatron. We now fix some values for the coupling( g (cid:48) = 0 . , . , . for the LHC analysis, g (cid:48) = 0 . , . for the Tevatron) and we see whatluminosity is required for discovery at each machine in the case of electrons in the final state.For muons in the final state, see Ref. [404] for a similar analysis. For g (cid:48) = 0 . the LHC re-quires . fb − to be sensitive at σ , while the Tevatron requires fb − . For the same value ofthe coupling, the Tevatron can discover the Z (cid:48) B − L boson up to M Z (cid:48) = 825 GeV, with fb − of data. The LHC can extend the Tevatron reach up to M Z (cid:48) = 925 GeV for g (cid:48) = 0 . . For g (cid:48) = 0 . , a discovery can be made, chiefly with electrons, requiring . fb − , for massesup to GeV at the LHC(Tevatron). Both machines will be sensitive at σ with muchlower integrated luminosities, requiring roughly . − . fb − to probe the Z (cid:48) B − L at the LHCand . fb − at the Tevatron, for g (cid:48) = 0 . − . . Finally, larger values of the coupling, such as g (cid:48) = 0 . , can be probed only at the LHC, which provides sensitivity at σ for masses up to . TeV. The lower masses kinematically accessible at the Tevatron limit the coupling that maybe probed while satisfying the LEP constraints.If no evidence for a signal is found at this energy and luminosity configuration of theLHC,
C.L. exclusion limits can be derived. We present here exclusion plots for the LHCas well as the expected exclusions at the Tevatron for (cid:82) L = 10 fb − . We start by looking at the C.L. limits presented in Fig. 4 for the Tevatron and for this stage of the LHC (for fb − and fb − of integrated luminosities, respectively).One can see that the different resolutions imply that the limits derived using electrons arealways more stringent than those derived using muons in excluding the Z (cid:48) B − L boson. As for thediscovery reach, the Tevatron is also competitive in setting limits, especially in the lower massregion. In particular, using electrons at the Tevatron for fb − , the Z (cid:48) B − L can be excluded forvalues of the coupling down to . ( . for muons) for M Z (cid:48) = 600 GeV. For the LHC to setthe same exclusion limit the same mass, fb − of integrated luminosity is required, allowing theexclusion of g (cid:48) > . . using electron(muons) in the final state. For the same integratedluminosity, the LHC has much more scope in excluding a heavier Z (cid:48) B − L boson, for M Z (cid:48) > GeV. For a coupling of . , the Z (cid:48) B − L boson can be excluded up to . . TeV at the LHCconsidering electrons(muons) for fb − , and up to GeV at the Tevatron for fb − ofdata. For g (cid:48) = 0 . , the LHC when looking at muons will require pb − to start improvingthe current available limits, while with pb − it can set limits on g (cid:48) = 0 . , out of the reachof Tevatron. It will ultimately be able to exclude Z (cid:48) B − L up to M Z (cid:48) = 1 . TeV for fb − (bothwith electrons and muons). √ s = 14 TeV
We consider here the performance at the design centre of mass energy of √ s = 14 TeV for anintegrated luminosity (cid:82) L = 100 fb − . We will present plots for the discovery potential onlyfor the Z (cid:48) B − L → e + e − channel. Similar plots for the muon channel can be found in Ref. [404].As before, exclusion plots will be presented for both electrons and muons in the final state.Figure 5 (left) shows the discovery potential for the Z (cid:48) B − L boson under these conditions,while Fig. 5 (right) shows the integrated luminosity required for σ evidence as well as for σ discovery as a function of the Z (cid:48) B − L boson mass for selected values of the coupling at √ s = 14 TeV. We consider the integrated luminosity in the range between pb − up to fb − . Aftera number of years of data analysis, the performance of the detector will be well understood.123 .75 1 1.25 1.510 -2 -1 95% C.L.LHC: ! s=7 TeVTevatron (L = 10 fb -1 )ElectronsMuonsLHC (L = 100 pb -1 )ElectronsMuons LHC (L = 1 fb -1 ) M Z' (TeV) g ' g' =0.1Tevatron, 95% C.L.Electrons Muons g' =0.1g' =0.05 g' =0.05 M Z' (TeV) I n t . Lu m . (f b - ) Figure 4: (top) Contour levels for
C.L. exclusion in the ( g (cid:48) , M Z (cid:48) ) plane at the LHC for selectedintegrated luminosities, and in the (integrated luminosity, M Z (cid:48) ) plane for selected values of g (cid:48) (in whichonly the allowed combination of masses and couplings are shown), for (bottom left) the LHC at √ s = 7 TeV and (bottom right) the Tevatron ( √ s = 1 . TeV), for both electrons and muons. The shaded areasand the allowed ( M Z (cid:48) , g (cid:48) ) shown are in accordance with Eq. (15) (LEP bounds, in black) and table 1(Tevatron bounds, in red for electrons and in green for muons). We therefore use the resolutions for both electrons and muons quoted in Eqs. (9) and (10),respectively.From Fig. 5 (left), we can see that the LHC at √ s = 14 TeV will start probing a com-pletely new region of the parameter space for (cid:82)
L ≥ fb − . For (cid:82) L ≥ fb − Z (cid:48) B − L gaugeboson can be discovered up to masses of TeV and for couplings as small as . . At (cid:82) L = 100 fb − , the coupling can be probed down to values of · − . The mass region that can be covered124 -2 -1 ! , g' =0.23 ! , g' =0.2Electrons5 ! , g' =0.13 ! , g' =0.15 ! , g' =0.053 ! , g' =0.055 ! , g' =0.0253 ! , g' =0.025 -2 -1 M Z' (TeV) I n t . Lu m . (f b - ) Figure 5: (left) Significance contour levels in the ( g (cid:48) , M Z (cid:48) ) plane for several integrated luminosities and(right) in the (integrated luminosity, M Z (cid:48) ) plane for selected values of g (cid:48) at the LHC for √ s = 14 TeVfor electrons. The shaded areas and the allowed ( M Z (cid:48) , g (cid:48) ) shown are in accordance with Eq. (15) (LEPbounds, in black) and table 1 (Tevatron bounds, in red). Similar plots for muons can be found in [404]. extends towards TeV.As before, Fig. 5 (right) shows the integrated luminosity required for σ evidence(discovery)of the Z (cid:48) B − L boson as a function of its mass, for selected values of the coupling. We explore lu-minosities in the range from pb − to fb − . However, only the configuration with g (cid:48) = 0 . can be probed with very low luminosity, and pb − is required to enable sensitivity (at σ ) to both . and . values of the coupling. It is worth emphasising here that the firstcouplings that will start to be probed at the LHC are those around g (cid:48) = 0 . , since the currentexperimental constraints are looser for this value of the coupling.At a given mass, the superior resolution in the electron w.r.t. the muon channel resultsin greater sensitivity to smaller couplings. For M Z (cid:48) = 600 GeV, the LHC at √ s = 14 TeVrequires . . fb − to be sensitive at σ to a value of the coupling of . . , in theelectron channel. A measure of the Z (cid:48) B − L boson width is also possible over a range of masses.For a comparison to the case with muons in the final state, we refer to Ref. [404].Figure 6 shows a pictorial representation of the Z (cid:48) B − L properties (widths and cross sec-tions) for selected benchmark points on the σ lines for fb − of data at √ s = 14 TeV, plottingthe di-electron invariant mass to which just the cuts of Eq. (6) have been applied (without se-lecting any mass window).As before, if no evidence for a signal is found at this energy and luminosity configurationof the LHC,
C.L. exclusion limits can be derived. Due to the improved resolutions forboth electrons and muons, they have very similar exclusion powers for couplings g (cid:48) (cid:38) . ,therefore setting similar constraints (for details, see Ref. [404]). Depending on the amount ofdata that is collected, several maximum bounds can be set (see Fig. 7) (left): e.g. for fb − of data, the LHC at TeV can exclude masses, at the
C.L., up to roughly TeV for a125 ! , " s=14 TeV, L=10fb -1 Electrons M e + e - (TeV) d ! / d M e + e - ( pb / G e V ) -7 -6 -5 -4 -3 -2 Figure 6: dσdM ll ( pp → γ, Z, Z (cid:48) B − L → e + e − ) for several masses and couplings ( M Z (cid:48) / TeV, g (cid:48) , Γ Z (cid:48) / GeV):( . , . , . ), ( . , . , . ), ( . , . , . ), ( . , . , . ), ( . , . , . ) and ( . , . , ), ( √ s = 14 TeV), using 15 GeV binning. Notice that the asymmetry of the peaks is the result of ourchoice to consider here the full interference structure. -2 -1 g' =0.2, elg' =0.2, mug' =0.1, elg' =0.1, mug' =0.05, elg' =0.05, mug' =0.025, elg' =0.025, mu95% C.L. M Z' (TeV) I n t . Lu m . (f b - ) -2 -1 Figure 7: (left) Contour levels for
C.L. exclusion in the ( g (cid:48) , M Z (cid:48) ) plane at the LHC for selectedintegrated luminosities and (right) in the (integrated luminosity, M Z (cid:48) ) plane for selected values of g (cid:48) (inwhich only the allowed combination of masses and couplings are shown), for √ s = 14 TeV, for bothelectrons and muons. The shaded areas and the allowed ( M Z (cid:48) , g (cid:48) ) shown are in accordance with Eq. (15)(LEP bounds, in black) and table 1 (Tevatron bounds, in red for electrons and in green for muons). value of the coupling g (cid:48) = 0 . . For fb − and for the same value of the coupling, the LHC This is the largest allowed value for the consistency of the model up to a scale Q = 10 GeV, from a
C.L., up to roughly TeV. For fb − it will be possible toexclude a Z (cid:48) B − L boson of M Z (cid:48) = 600 GeV if the coupling is greater than . · − (9 · − ) formuons(electrons) and values of the coupling greater than . · − (7 · − ) for an integratedluminosity of fb − . Figure 7 (right) shows the integrated luminosity that is required toexcluded a certain Z (cid:48) B − L mass for fixed values of the coupling. For g (cid:48) ≥ . the same limits areobtained for electrons and muons. An integrated luminosity of fb − is required to excludea Z (cid:48) B − L boson mass up to . TeV for g (cid:48) = 0 . and fb − reduces this to g (cid:48) = 0 . . Foran integrated luminosity of fb − the LHC experiments will be able to exclude a Z (cid:48) B − L formasses up to . TeV for g (cid:48) = 0 . , . . TeV for g (cid:48) = 0 . and . . TeV for g (cid:48) = 0 . ,when considering the decay into electrons(muons). With and integrated luminosity of fb − of data more stringent bounds can be derived: for g (cid:48) = 0 . . the Z (cid:48) B − L boson can beexcluded for masses up to . . TeV in the electron channel, and up to . . TeV in themuons channel. Conclusions
We have presented the discovery potential for the Z (cid:48) gauge boson of the B − L minimal ex-tension of the SM at the LHC for CM energies of √ s = 7 and TeV, using the integratedluminosities expected at each stage. This has been done for both the Z (cid:48) B − L → e + e − and Z (cid:48) B − L → µ + µ − decay modes, and includes the most up-to-date constraints from LEP and theTevatron.A general feature is that greater sensitivity to the Z (cid:48) B − L resonance is provided by theelectron channel. At the LHC this has better energy resolution than the muon channel. A furtherconsequence of the better resolution of electrons is that an estimate of the gauge boson widthwould eventually be possible for smaller values of the Z (cid:48) B − L mass than in the muon channel.Limits from existing data imply that the first couplings that will start to be probed at the LHCare those around g (cid:48) = 0 . . Increased luminosity will enable both larger and smaller couplingsto be probed.Our comparison shows that, for an integrated luminosity of fb − , the Tevatron is stillcompetitive with the LHC in the small mass region, being able to probe the coupling at thelevel of σ down to a value of . . using electrons(muons). The LHC will start to becompetitive in such a region only for integrated luminosities close to fb − at √ s = 7 TeV. At √ s = 7 TeV the mass reach will be extended from the Tevatron value of M Z (cid:48) = 750 GeV up to . TeV for electrons(muons).When the data from the high energy runs at the LHC becomes available, the discoveryreach of Z (cid:48) B − L boson will be extended towards very high masses and small couplings in regionsof parameter space well beyond the reach of the Tevatron and comparable in scope with thoseaccessible at a future LC [411].If no evidence is found at any energies, C.L. limits can be derived, and, given theirbetter resolution, the bounds from electrons will be more stringent than those from muons,especially at smaller masses.While this work was in progress, other papers dealing with the discovery power at theLHC for the Z (cid:48) B − L boson appeared, for CM energies of TeV [413] and TeV [414], as wellas for other popular Z (cid:48) boson models. Our results broadly agree with those therein. Renormalisation Group (RG) analysis of the gauge sector of the model [390, 406]. cknowledgements
LB thanks Muge Karagoz Unel and Ian Tomalin for useful discussions. SM is financially sup-ported in part by the scheme ‘Visiting Professor - Azione D - Atto Integrativo tra la RegionePiemonte e gli Atenei Piemontesi’. 128 ontribution 11Single custodian production in warped extra dimensionalmodels
S. Gopalakrishna, G. Moreau and R.K. Singh
Abstract
We examine the single production of heavy fermions at the LHC andshow in particular the possible importance of some new types of pro-cesses, namely single production of the heavy fermion in associationwith a standard model gauge boson or a Higgs boson. The theoreticalframework is the Randall-Sundrum scenario, motivated by the gaugehierarchy problem, with a bulk custodial symmetry. In this context,the heavy fermion considered in the Kaluza-Klein (KK) excitation ofa fermion that does not have a zero mode. The location along thefifth dimension (affecting the single production) are selected such thatthey generate the masses of the third generation quarks. From both thetheoretical and phenomenological sides, the studied KK quarks can belighter than the KK excitations of gauge bosons which makes their po-tential discovery at the LHC easier, as illustrated by the cross sectionswe obtain numerically. Introduction
Recent alternatives to supersymmetric scenarios, like extra dimension theories, composite Higgsand little Higgs models (as well as twin Higgs and fourth generation models), predict the exis-tence of additional heavy fermions. Such fermions, e.g. exotic quarks, could be directly pro-duced at the LHC providing a clear discovery of new physics underlying the Standard Model(SM). In particular, the single production of such a heavy fermion is favored w.r.t. pair produc-tion from the point of view of the phase space .In the present work, we propose a systematic study of the various channels of single b (cid:48) (new quark with an electric charge of − / ) production at LHC. The exhaustive list of possibleelementary processes of type → body and → body is: qq → tb (cid:48) , qq → bb (cid:48) , qg → qtb (cid:48) , qg → qbb (cid:48) , bg → b (cid:48) Z , bg → b (cid:48) h , QQ/gg → bb (cid:48) Z , QQ/gg → bb (cid:48) h , qb → qb (cid:48) Z , qb → qb (cid:48) h , qq/gg → tb (cid:48) W and qb → qb (cid:48) W , where q stands for any SM quark except for the bottom andtop quarks which are denoted b and t respectively ( QQ denotes either the initial state qq or bb ) . The theoretical framework we consider is precisely defined: it is the Randall-Sundrum The pair production of exotic quarks has been studied in the warped extra dimension scenario with a custodialsymmetry [352, 415] (within the gauge-Higgs unification context [357]) or similarly within their dual compositeHiggs description [272]. See also Ref. [416] for the case of a strongly coupled fourth generation and Ref. [417,418]for more general approaches (using jet mass [331]). The possible b (cid:48) decay channels, whose branching ratios depend on the considered model, are b (cid:48) → bZ and b (cid:48) → tW . b (cid:48) , called the custodians, which appear in theextended gauge multiplets. These custodians are pure Kaluza-Klein (KK) excitations withoutzero-modes, due to specific boundary conditions.The well-known RS scenario is motivated by the gauge hierarchy problem and the cus-todial symmetry allows to satisfy the ElectroWeak Precision Test (EWPT) constraints for KKgauge boson masses in the vicinity of the TeV scale. The RS framework is also attractive as aflavor model and we will carefully consider quark locations, reproducing quite precisely the b and t masses, which are crucial for the single b (cid:48) production. In particular, in this flavor frame-work, the b (cid:48) quark, whose mass is controlled by the location of the right-handed t R (being in thesame multiplet under the custodial extension), tends to be particularly light due to the large topmass m t which also explains the motivation for analyzing its single production.The above type of single heavy fermion production processes with an EW gauge bosonor Higgs field in the final state were never studied before. In contrast, the other single processesalready considered have been within generic approaches [419] or within the different theoreticalcontexts of the composite Higgs models [353, 354], the little Higgs scenario [420] as well asthe twin Higgs mechanism [421]. The obtained production cross sections depend on the modelconsidered. There were even NLO estimations of this second class of single heavy fermionproduction (with only fermions in the final state) in the fourth generation context [422, 423].Nevertheless, to our knowledge, the contribution of KK excitations of gauge bosons, or eventheir mixing effect with SM bosons – both of which we will consider here – for this class ofsingle heavy fermion production have never been studied previously. Theoretical framework
We consider the RS scenario under the theoretical assumption of a bulk gauge custodial sym-metry
SU(2) L × SU(2) R × U(1) X which allows for a reduction of the final EWPT bound on themass of the first KK gauge boson excitation M KK (strictly speaking, the KK photon mass)from ∼ TeV down to a few TeV [321, 424], improving then the situation with regard tothe little hierarchy problem (fine tuning of the Higgs boson mass due to its loop level cor-rections sensitive to new physics). For simplicity, we take the minimal quark representationsunder the custodial symmetry [321]: the corresponding multiplets under the custodial symmetry
SU(2) L × SU(2) R × U(1) X for the bottom and top quarks are the three doublets Q L ≡ ( , ) / = ( t L , b L ) , Q b R ≡ ( , ) / = ( t (cid:48) R , b R ) , Q t R ≡ ( , ) / = ( t R , b (cid:48) R ) whereas the representation for the Higgs field, responsible for the EW symmetry breaking, is Σ ≡ ( , ) . Although for simplicity the quark representations just above are the only ones shown here,the effects we present in this paper are also qualitatively relevant for the model [322] wherethe Z ¯ b L b L coupling is protected by a custodial symmetry under which the quark doublet isenlarged to Q L ≡ ( , ) / and e.g. Q t R ≡ ( , ) / ⊕ ( , ) / . We will address these detailsin Ref. [425] (see also the discussion in next section).We will focus in this work on the production of b (cid:48) R at the LHC. For notational ease wewill denote the b (cid:48) R simply as b (cid:48) . 130 .2 EW Precision Tests The bulk custodial symmetry ensures that the global EW fit can reach a better goodness-of-fitthan for the SM case, in the EW gauge bosons and light fermions sector (not including thebottom and top quarks) as long as M KK > O (3) TeV, m h (cid:39) GeV and light fermions arelocalized towards the Planck-brane ( c light > . ) [426]. The c parameters are the dimensionlessquantities parameterizing the fermion five-dimensional (5 D ) masses which fix the profiles alongthe fifth dimension (see e.g. Ref. [321]).In the b , t sector, the R b observable is protected from excessively large deviations in the Zbb coupling (induced by b − b (cid:48) mixings) by taking m b (cid:48) > O (1 . TeV [298]. We will also con-sider some masses below O (1 . TeV but then the Z ¯ b L b L vertex can be protected by a subgroupof the custodial symmetry O(3) [322] . We thus assume such a symmetry for this low massspectrum, keeping in mind that this symmetry corresponds to quark representations differentfrom those given above but these new representations are not expected to modify significantlythe b (cid:48) couplings involved in its single production, and in turn our illustrative numerical results.Given the present simple theoretical context, we are forced to assume that the anomalies onthe forward-backward asymmetries A bF B and A tF B are due respectively to underestimated un-certainties and too preliminary data, so that we do not have to interpret them in terms of newphysics effects (see Ref. [427] and Ref. [428], respectively, for interpretations based on KKcontributions). b (cid:48) at the LHC We discuss here the production of the b (cid:48) in association with a (longitudinal) vector boson at theLHC. In particular, we are interested in the processes: gg → ¯ b L b (cid:48) R Z L /h and ¯ b L b L → ¯ b L b (cid:48) R Z L /h ,where Z L is the longitudinal polarization. Owing to the Goldstone boson equivalence theorem,the longitudinal polarization of the vector boson is nothing but the corresponding Goldstoneboson, and we have the correspondence V µL ↔ ∂ µ φ/M V where M V is the vector boson mass.We show in Fig. (1) example Feynman diagrams for these processes. The couplings φ ± tb (cid:48) , φ bb (cid:48) and hbb (cid:48) , involved in each single b (cid:48) production, are important due to a large wave functionoverlap near the TeV-brane owing to these fermion masses being sizable. b (cid:48) couplings For computing the b (cid:48) couplings, we will need the Yukawa couplings given, in terms of the 5 D Yukawa coupling constants, by L D ⊃ − λ t ¯ Q L Σ Q t R − λ b ¯ Q L Σ Q b R . (1)Electroweak symmetry is broken by (cid:104) Σ (cid:105) = diag( v, v ) (the Higgs boson vev is v ≈ GeV). The Goldstone bosons of electroweak symmetry breaking ( φ ) are contained in Σ = ve iφ a T a /v , written in the nonlinear realization, where T a are the generators of SU (2) L . Wework here in the unitary gauge (as for our numerical calculations) for which we absorb the The corrections to Z ¯ b R b R from the mixing with KK excitations of the Z boson are less dangerous than the Z ¯ b L b L deviations due to the vanishing SU(2) L isospin of b R leading to a smaller coupling to Z excitations.Moreover, we consider here only the domain c b R > . which tends to minimize the effective four-dimensional(4 D ) b R couplings to KK Z ( (cid:48) ) excitations. In a more precise analysis [425], we will give the numerical results forthe corrections to the Z ¯ b R b R vertex in the O(3) context. ’ R b L b L φ , h b’ R b L _ b’ R b L h φ , b’ R b’ R b L _b L b L _ γ ΚΚΚΚ g , ΚΚ Z , Z ’ ΚΚ g φ , h ggg _ Figure 1:
Examples of Feynman diagram contributing to the single custodian productions ¯ b L b L → ¯ b L b (cid:48) R φ , ¯ b L b (cid:48) R h and gg → ¯ b L b (cid:48) R φ , ¯ b L b (cid:48) R h at LHC. We indicate here the fermion fields in interactionbasis (those are not the mass eigenstates). g stands for the gluon, φ for the neutral Goldstone boson(equivalently the longitudinal Z ) and h for the Higgs field. Z KK represents the full KK tower for the Z boson (including the zero-mode), γ KK the full KK tower for the photon, Z (cid:48) KK is the KK tower for theadditional neutral gauge boson and similarly g KK stands for the (KK) gluons. Goldstone bosons as the longitudinal polarization of the gauge bosons. Nevertheless, for com-pleteness and to have a clear understanding of the involved couplings, a derivation of the cou-plings using Goldstone boson equivalence is presented in App. A.Electroweak symmetry breaking also gives rise to fermion masses. Taking the case of the b L , for example, the off-diagonal terms in the bottom mass matrix resulting from Eq. (1) lead tothe mixing of the fields ( b L , b (cid:48) ( n ) R L , b ( n ) L L , b ( n ) R L ) , where when two subscripts are present, the first L, R denote the gauge-group representation while the second subscript denotes Lorentz chiral-ity, and ( n ) denotes the n th KK state. Similarly, a mixing is induced amongst the correspondingLorentz R fields. Focusing on the dominant of these mixing terms for simplifying the presentdiscussion we have L D ⊃ − (cid:0) ¯ b L ¯ b (cid:48) L (cid:1) (cid:18) λ b vπR c f Q L ( πR c ) f b R ( πR c ) λ t vπR c f Q L ( πR c ) f b (cid:48) R ( πR c )0 m b (cid:48) (cid:19) (cid:18) b R b (cid:48) R (cid:19) + h . c . (2)where L, R above (and in the text that follows) denote Lorentz chirality, f i are the wavefunctionvalues of the corresponding fields at the IR-brane location. The above mass matrix is diago-nalized by biorthogonal rotations, and we denote the sine (cosine) of the mixing angles by s L,Rθ ( c L,Rθ ). The sine of mixing angle s Lθ is given in Appendix App. A. We denote the correspondingmass eigenstates as ( b b ) .Due to these mixings, the Z ¯ b L b L and W ¯ t L b L couplings are induced, where we simi-larly define t , as the top mass eigenstates and c α as the cosine of mixing angle between the Numerically, we take into account also the b mixing with the second KK excitations of the b (cid:48) custodian as wellas with the first KK excitations of the b L and b R fields. t (cid:48) field. The contribution to these couplings due e.g. to W ↔ W KK mixings arehigher order in v/M KK , which we have not shown here, but included in our numerical results.The couplings of the heavier mass eigenstate b in unitary gauge are given by L D ⊃ − e b L/R γ µ b L/R A µ + g s ¯ b L/R γ µ T α b L/R g αµ − (cid:18) gs Lθ c Lα √ t L γ µ b L W + µ + h . c . (cid:19) + g Z (cid:18) − s Lθ + 13 s W (cid:19) ¯ b L γ µ b L Z µ + (cid:20) g Z c Lθ s Lθ (cid:18) (cid:19) ¯ b L γ µ b L Z µ + h . c . (cid:21) + g Z (cid:18) s W (cid:19) ¯ b R γ µ b R Z µ , (3)where g Z ≡ (cid:112) g + g (cid:48) . The photon and gluon couplings are diagonal in the ( b b ) basis andare identical to the b-quark couplings.To derive the Higgs couplings to fermions, we start with Eq. (2) and apply the vev re-placement v → h/ √ . Writing in terms of the fermion mass eigenstates, we obtain L D ⊃ − h √ (cid:2) ¯ b L b R ( c Lθ c Rθ λ Db + c Lθ s Rθ λ Db (cid:48) ) + ¯ b L b R ( s Lθ s Rθ λ Db − s Lθ c Rθ λ Db (cid:48) )+¯ b L b R ( − c Lθ s Rθ λ Db + c Lθ c Rθ λ Db (cid:48) ) + ¯ b L b R ( − s Lθ c Rθ λ Db − s Lθ s Rθ λ Db (cid:48) ) (cid:3) + h . c . (4)where λ Db,b (cid:48) = λ b πR c f Q L ( πR c ) f b R ,b (cid:48) R ( πR c ) . For instance, f b (cid:48) R ( πR c ) is the value of the wave func-tion for b (cid:48) R (controlled by the c t R parameter of the Q t R doublet) taken at the TeV-brane — asinduced by the overlap with the peaked profile of the Higgs field. The motivation for the single b (cid:48) production resides in both its possibly low mass and its ratherlarge coupling to φ and h compared to lighter quark generations.Let us first consider the most severe experimental lower bound on a fourth generation b quark mass in order to get a rough idea ( a priori the b (cid:48) couplings differ from the b ones) ofthe realistic m b (cid:48) range: m b > GeV at
C.L. [2]. Now there might appear more severelower bounds on m b (cid:48) from FCNC considerations. Indeed, generally the new heavy fermionscan contribute to b → sγ at the one-loop level (see [429] for the two-site approach to 5 D AdSmodels with bulk Higgs, a framework that differs from the present one where e.g. the Higgsfield is confined on the TeV-brane). However such indirect constraints rely on the whole setof 5 D Yukawa couplings and light fermion locations along the extra dimension that we do notspecify here. A complete three-flavor study, beyond our scope, should also reproduce the CKMangles for the quark mixing [2].We consider a set of parameters, M KK = 2 . TeV, m h (cid:39) GeV and g Z (cid:48) = 1 . ,justified by considerations on the global EW fit [426]. Then we choose various values of c t R ( c t R is the main parameter determining the single b (cid:48) production cross section) fixing the right-handed top quark location and also mainly m b (cid:48) : c t R = − . ( m b (cid:48) = 225 GeV), c t R = − . ( m b (cid:48) = 756 GeV) and c t R = − . ( m b (cid:48) = 1574 GeV). Strictly speaking, the physical state is the b state, namely the second lightest bottom quark mass eigenstate composed in general mainlyof the first b (cid:48) custodian excitation but also partially of the pure SM b field (due to the b − b (cid:48) mixing effect). So the lightest b eigenstate is associated to the measured bottom quark mass m b , while the b eigenstate is associated to m b (cid:48) (that we should write m b but leave as m b (cid:48) forsimplification reasons). 133or each c t R value, examples of c Q L and the top quark 5 D Yukawa coupling constant λ t are chosen so that m t ∼ GeV. We note that the exact top mass value can only be fittedprecisely after a complete three-flavor treatment of the full quark mass matrix. Then order onecorrections can bring the top mass obtained here exactly to the measured value. Finally, weselect some values of c b R and λ b reproducing m b (cid:39) GeV.
Including the main contributions of the first KK gauge boson and fermion excitations, we obtainthe cross sections for the single b (cid:48) production at LHC given in Table 1. Strictly speaking, thecalculated amplitudes correspond to a final state with a unique b state, namely the secondlightest bottom quark mass eigenstate. A) c Q L = − . c t R = − . c b R = +0 . B) c Q L = − . c t R = − . c b R = +0 . C) c Q L = − . c t R = − . c b R = +0 . m b (cid:48) = 225 GeV
GeV
GeV m b (cid:39) [ m t (cid:39) ] 4 [197] GeV
GeV
GeV σ ( qq → tb (cid:48) ) (cid:39) fb . fb . fb σ ( qq → bb (cid:48) ) (cid:39) fb . fb . fb σ ( qg → qtb (cid:48) ) (cid:39) fb . fb . fb σ ( qg → qbb (cid:48) ) (cid:39) fb fb . fb σ ( bg → b (cid:48) Z ) (cid:39) fb . fb . fb σ ( bg → b (cid:48) h ) (cid:39) fb fb . fb σ ( QQ/gg → bb (cid:48) Z ) (cid:39) fb . fb . fb σ ( QQ/gg → bb (cid:48) h ) (cid:39) fb fb . fb σ ( qb → qb (cid:48) Z ) (cid:39) fb . fb . fb σ ( qb → qb (cid:48) h ) (cid:39) fb . fb . fb σ ( qq/gg → tb (cid:48) W ) (cid:39) fb . fb . fb σ ( qb → qb (cid:48) W ) (cid:39) fb . fb . fb σ ( QQ/gg → ¯ b (cid:48) b (cid:48) ) (cid:39) fb fb . fb Table 1:
Values (in fb) of the cross sections for the pair and the various single b (cid:48) production reactionsat LHC for three sets (A, B, C) of parameters with M KK = 2 . TeV, m h (cid:39) GeV, g Z (cid:48) = 1 . [seetext]. For each set, we also give the predicted values of m b (cid:48) , m b and m t (in GeV). QQ denotes either theinitial state qq or bb . We conclude from the examples of parameter sets considered in the table that within thepresent RS framework (reproducing m b,t ) both the single and pair b (cid:48) productions have promisingcross sections which might allow for their possible detection at LHC, if one assumes the typicalhigh luminosity regime expected ( L ∼ fb − ). The considerable number of events predictedhere should still lead to significant signals after including the decay analysis, hadronization134ffects, the Monte Carlo simulation and effects of detector response. Note also that for theextreme situation at m b (cid:48) = O (1 . TeV, if a b (cid:48) custodian is to be discovered at LHC it is mainlyvia the single production reaction QQ/gg → bb (cid:48) h only. Indeed, for this high custodian mass,the pair production rate is reduced too much by the phase space suppression.Note that in comparison, the resonant KK gluon production suffers from typically lowercross sections at LHC [341] (see also Ref. [350, 351, 430] for the production of KK EW gaugebosons) rendering its observation more tricky. This is essentially due to the high KK gluonmasses ( M KK > O (3) TeV) compared to the lighter b (cid:48) custodians considered here.While the dominant single production processes were thought to be only qg → qtb (cid:48) and qg → qbb (cid:48) (partially due to a polarization increase effect), we see that for instance the newsingle production reactions bg → b (cid:48) Z and QQ/gg → bb (cid:48) h – originally studied here (see alsoRef. [425]) – can be of comparable order or even larger than the previously thought dominantones. The reason is principally the possibly pure gluonic initial state for these new processes.The gg → tb (cid:48) W reaction, having only a possible pure gluonic initial state, is significant at a low m b (cid:48) value for which the parton density functions are significantly higher for the gg initial state.Therefore, from a more general point of view, novel reactions such as bg → b (cid:48) h , QQ/gg → bb (cid:48) Z or QQ/gg → bb (cid:48) h should be now included in the experimental investigations for any heavyquark production predicted by a scenario underlying the SM. Finally, note also that such newprocesses – like the ones drawn in Fig. (1) – constitute new channels for the Higgs boson pro-duction at LHC . Acknowledgements
The authors thank K. Agashe and A. Pukhov for interesting discussions and also
Les Houches conveners for organizing this nice Workshop where the present work was started. SG thanksBrookhaven National Laboratory for partial financial support to attend the workshop.
Appendices
App. A
Goldstone boson equivalence
Here we explicitly check the correspondence through the equivalence theorem between e.g. the φ ± tb (cid:48) and W ± L tb (cid:48) couplings in the R ξ and unitary gauges respectively. The correspondence inthe neutral sector can be shown analogously.In the linear realization, the Higgs is written as Σ = (cid:18) Φ ∗ φ + − φ − Φ (cid:19) , (5)with Electroweak symmetry broken by (cid:104) Σ (cid:105) = diag ( v, v ) (recall that v ≈ GeV in ournotation) and the three Goldstone bosons are φ ± and Im(Φ ) = φ / √ .After reducing the 5 D theory to an equivalent 4 D theory, we can write the mass matrix inthe b -sector as L D ⊃ − (cid:0) ¯ b L ¯ b (cid:48) L (cid:1) (cid:18) λ Db v λ Db (cid:48) v m b (cid:48) (cid:19) (cid:18) b R b (cid:48) R (cid:19) + h . c . , (6) In contrast, the corrections arising in the RS model to the usual Higgs production mechanisms have alreadybeen studied e.g. in Ref. [424, 431]. D Yukawa couplings given as λ Db = λ b πR c f Q L ( πR c ) f b R ( πR c ) and λ Db (cid:48) = λ t πR c f Q L ( πR c ) f b (cid:48) R ( πR c ) where the λ ’s are the corresponding 5 D Yukawa couplings.In the λ Db v → limit, the mass matrix is diagonalized by rotating only the L fields by s Lθ = − λ Db (cid:48) v/ (cid:112) m b (cid:48) + ( λ Db (cid:48) v ) with the heavier mass eigenvalue being m b = (cid:112) m b (cid:48) + ( λ Db (cid:48) v ) .Next we turn to showing the correspondence. For simplicity (and since it represents agood approximation) we do so in the λ Db v → limit. In order to extract the relevant interactionof the b (cid:48) quark, we focus on the following two couplings contained in the first term of Eq. (1): L ⊃ − λ Db (cid:48) φ + ¯ t L b (cid:48) R + λ Dt φ + ¯ t R b L + h . c . . (7)Here λ Dt = λ t πR c f Q L ( πR c ) f t R ( πR c ) . Writing in terms of the fermion mass eigenstates we obtain L ⊃ φ + (cid:0) − λ Db (cid:48) ¯ t L b R + c Lθ λ Dt ¯ t R b L − s Lθ λ Dt ¯ t R b L (cid:1) + h . c . . (8)We will show that we can recover Eq. (8) starting from the longitudinal W ± L unitary gaugecoupling in Eq. (3). Including the SM piece, this coupling in unitary gauge in the mass basis is L ⊃ g √ W L + µ ¯ t L γ µ (cid:0) c Lθ b L − s Lθ b L (cid:1) + h . c . , (9)where we have ignored the t ↔ t (cid:48) mixing (since m t (cid:48) (cid:29) m b (cid:48) for the parameter sets considered)and thus set c Lα = 1 . The Goldstone-boson equivalence theorem implies W L + µ ↔ ∂ µ φ + /m W = q µ φ + /m W where q µ is the momentum of the W boson. Using this and momentum conservation q µ = p t µ − p b µ (where p t is the momentum out of the vertex and p b into the vertex) and m W = gv/ √ , Eq. (9) becomes L ⊃ v φ + ¯ t L (cid:2) c Lθ (cid:0) /p t − /p b (cid:1) b L − s Lθ (cid:0) /p t − /p b (cid:1) b L (cid:3) + h . c . . (10)Under the assumption that all fermions are on mass-shell and using the equations of motion(Dirac equations: (cid:54) p ψ ( p ) = mψ ( p ) ) we recover Eq. (8), thus explicitly showing the corre-spondence. One has to recall that if the top quark mixing is neglected, the top mass reads as m t (cid:39) λ Dt v .The contributions to this coupling due to the mixing of the SM W boson with its KKexcitations constitute higher order corrections, in v/M KK , which we have already ignored inthe present theoretical demonstration (but not in the numerical results as mentioned previously).Finally, we note that the correspondence between the φ bb (cid:48) and Z L bb (cid:48) couplings can be similarlyshown. 136 ontribution 12Four top final states G. Servant, M. Vos, L. Gauthier and A.-I. Etienvre
Abstract
Of the many interesting final states that may be produced at the LHC,four top production is maybe one of the most spectacular. In this con-tribution, the sensitivity of this final state to several classes of physicsbeyond the Standard Model is discussed. The focus is on models wherethis topology is produced through a heavy resonance. The possibility toreconstruct the top and anti-top quarks in these events is explored. Introduction
Four top production occurs in the Standard Model through a large number of diagrams [432],two of which are indicated in figure 1(a-b). The total pp → tttt cross-section at 14 TeV is 7.5fb in the Standard Model. The production is dominated by gluon-initiated diagrams. ttttgg (a) ttttqq (b) ttgg tt KK* g (c) ttgg ttX (d) Figure 1: (a-b): Two Standard Model diagrams that give rise to the t ¯ tt ¯ t final state. (c-d): Two diagramsinvolving new physics, that yield to a non-zero event rate even if the new particle does not couple to lightquarks. (c) represents s-channel (resonant) t ¯ t production. The effective four-top interaction in (d) canresult from integrating out a heavy particle. The interest of this final states lies primarily in its sensitivity to beyond-the-standard-model physics, as recently discussed in [433–435]. These authors consider a composite topquark that would give rise to contact interactions like that of figure 1(d). The production cross-section through the contact interaction can be as large as several tens of fb.Another possibility is the production of the t ¯ tt ¯ t final state through an exotic heavy particle.In many extensions of the Standard Model the top quark plays a special role. New particles witha preference for the top quark could yield a sizeable cross-section through processes like thatdepicted in figure 1 (c), where the contact interaction is replaced by a resonance (in this case aKaluza–Klein gluon). It is interesting to note that this diagram involves only couplings of thenew particle to top quarks. In particular, the new particle does not have to couple to light quarksor gluons to be produced at the LHC. In section 2 two models giving rise to resonant four topproduction are discussed in some detail. 137 measurement of the four-top production rate would strongly constrain several models.While a complete, detector-level analysis is still missing, several authors [433–435] have in-vestigated the possibility to isolate this signal. A common aspect of these studies is that theisolation strategy consists in requiring two leptons with the same sign. Thus, processes like t ¯ t + jets production, with cross-sections that are several orders of magnitude larger than typi-cal signal cross-sections, are effectively reduced. The power of selecting same-sign dileptonevents to study ttW W final states from pair-production of heavy quarks was shown in detailin [353] and recently applied by CDF to put a strong bound on the mass of fourth generationdown-type quarks ( b (cid:48) ) [356]. The reduction of SM t ¯ t production using the same-sign criterionwas shown by ATLAS Monte Carlo studies of t ¯ tH production, with H → W W ∗ ( [436], pages1367-1368).Further experimental handles to distinguish the signal are particularly important in thelight of the large cross-section of several reducible background processes, like t ¯ tW + jets, t ¯ tW W + jets and t ¯ t + jets. Therefore, reconstruction of all or several of the top decays canstrengthen the robustness of the analysis considerably. This possibility is particularly interestingin searches for resonant production. Reconstruction of the top quarks allow the reconstructionof the mass of the resonance. The background level can then be normalized in the off-peakregion, thus considerably increasing the sensitivity of the search.A complete study into the reconstruction of this extremely challenging final state is clearlybeyond the scope of this contribution. The results from a first superficial exploration of someideas is presented in section 3. Resonant production
A prototype is based on the Randall-Sundrum (RS) setup where the hierarchy between thePlanck and electroweak scales is explained through warping of an extra dimension. The Stan-dard Model (SM) lives in the bulk [321] but the Higgs lives on the IR boundary where thenatural scale of physics is ∼ TeV. As a result of the localized Higgs, the zero mode of the right-handed top quark must also live close to the IR brane, in order to realize the large top mass . Asa consequence of the warping, all of the low level Kaluza–Klein (KK) modes have wave func-tions whose support is concentrated near the IR brane. Thus, they inevitably couple strongly to t R and the Higgs.The KK gluon [437] has a number of features that render it very interesting phenomeno-logically. First of all, it cannot be revealed by resonance searches in di-lepton final states. Asa coloured object it can be produced relatively abundantly (compared to partners of the electro-weak gauge bosons). With the couplings to light quarks of reference [437], g q = g Lb = − . g s , g Rb = g Lt = g s and g Rt = 4 g s , it is still sufficiently narrow ( Γ = 0 . M ) to yield a resonantsignature. Finally, it has a very sizeable branching ratio of 92.5 % into a top anti-top pair. Thecross-section for the four-top final state through the KK gluon is represented in figure 2. Itis noted here that the model of reference [437] is experimentally viable for KK gluon massesgreater than 2-3 TeV, in which case the 4 top production cross section is well below a fb andLHC prospects are not encouraging.However, one can easily envision some variation of the RS setup as follows. Consider a The left-handed top is usually chosen to be further from the IR brane, in order to mitigate constraints fromprecision electroweak tests [321]. g KK model "
14 TeV ! Z' model "
14 TeV... Z' model "
10 TeV ! effective interaction "
14 TeV ! SM "
14 TeV500 1000 1500 20000.0010.010.11 M Z ' , M KK , ! GeV " Σ ! pb " p p % t t & t t & Figure 2:
Four top production cross section at the LHC in the different theories discussed in this report.The orange curve refers to the effective 4-fermion interaction ( t R γ µ t R )( t R γ µ t R ) / Λ leading to fig. 1(d). top-philic Z (cid:48) described by the following lagrangian [438] L = L SM − F (cid:48) µν F (cid:48) µν + M Z (cid:48) Z (cid:48) µ Z (cid:48) µ + χ F (cid:48) µν F µνY + g Z (cid:48) t ¯ tγ µ P R Z (cid:48) µ t (1)where F (cid:48) µν ( F Yµν ) is the usual Abelian field strength for the Z (cid:48) (hypercharge boson), g Z (cid:48) t is the Z (cid:48) coupling to right-handed top quarks (we will take g Z (cid:48) t = 3 in our simulations). The parameter χ encapsulates the strength of kinetic mixing between the Z (cid:48) and SM hypercharge bosons (evenif absent in the UV, it is generated in the IR by loops of top quarks). These extra terms inthe lagrangian have a natural connection to Randall–Sundrum theories. The Z (cid:48) represents thelowest KK mode of the U (1) contained in SU (2) L × SU (2) R × U (1) B − L . It typically hasmixing with the electroweak bosons, resulting in strong constraints from precision data. Thiswill also be the case when the Z (cid:48) is a KK mode of the electroweak bosons. We circumvent theseconstraints by considering a Z (cid:48) whose mixing with the Z is kinetic. At large Z (cid:48) masses this isnot operationally different from the mass-mixing case, but it allows us to consider lower mass Z (cid:48) s which are not ruled out by precision data.Through the RS/CFT correspondence [304, 305], the extra-dimensional theory is thoughtto be dual to an approximately scale-invariant theory in which most of the Standard Model isfundamental, but with the Higgs and right-handed top largely composite. The Higgs couplesstrongly to composite fields, and the amount of admixture in a given SM fermion determines itsmass [212]. In this picture, the Z (cid:48) is one of the higher resonances, built out of the same preonsas t R . More generically, in models of partial fermion compositeness, it is natural to expect thatonly the top quark couples sizably to a new strongly interacting sector. As a simple exampleof a UV completion (see Appendix A of [438]) we can treat all SM fields (including t R ) as One can easily include a coupling to the left-handed top (and bottom). Our choice to ignore such a coupling fitswell with typical RS models, balancing the need for a large top Yukawa interaction with control over correctionsto precision electroweak observables. U (1) (cid:48) . We include a pair of fermions ψ L and ψ R , whose SM gauge quantumnumbers are identical to t R , but with equal charges under U (1) (cid:48) . To realize coupling of the Z (cid:48) to the top quark, we consider the gauge invariant masses and Yukawa couplings of the top- ψ sector, yH ¯ Q t R + µ ¯ ψ L ψ R + Y Φ ¯ ψ L t R (2)where Q is the 3rd family quark doublet, H is the SM Higgs doublet, Φ is the Higgs fieldresponsible for breaking U (1) (cid:48) , y and Y are dimensionless couplings, and µ is a gauge-invariantmass term for ψ . The lightest mass eigenstate is identified as the top quark. It can have a largecoupling to Z (cid:48) through its ψ component.Four-top production arises via the diagrams shown in Fig. 3. Given that the Z (cid:48) underconsideration has suppressed couplings to all SM fields (induced by the kinetic mixing χ ) butthe top quarks, constraints are weak and a mass of a few hundreds of GeV is allowed, whichcan lead to large four-top signals at the LHC. A detailed study is presented in [439] and com-pares with the non-resonant four-top events obtained from the effective four-fermion interaction ( t R γ µ t R )( t R γ µ t R ) / Λ leading to the diagram 1(d). The corresponding cross sections at LHC asa function of the Z (cid:48) mass and Λ are shown in Fig. 2. tttt −− Z’ tttt −− Z’ Figure 3:
Four-top production via Z (cid:48) An interesting way to probe the properties of the top interactions relies on measuringthe top polarization. The SM four top production being dominated by parity invariant QCDprocesses, we expect to generate an equal number of left and right-handed pairs. However, inthe new physics models discussed here, there is a strong bias towards RH tops. The angulardistribution of the leptons from the top decays enables to analyze the polarisation of the topquarks. The differential cross section can be written as σ dσd cos θ = A θ ) + 1 − A − cos θ ) (3)where θ is the angle between the direction of the lepton in the top rest frame and the directionof the top polarization. The corresponding distribution is illustrated in Fig. 4.In Fig. 5, we show the invariant mass M tt of the tt pair coming from the Z (cid:48) for different M Z (cid:48) masses as well as M tt from the SM four-top events. The latter peaks close to 600 GeV.We also display the maximum of the tt pair transverse energy distribution as a function of M Z (cid:48) . Fig. 6 compares the M tt distributions of the tt pair emitted by a Z (cid:48) with M Z (cid:48) = 1 . TeV,the spectator tt pair, which peaks around 500 GeV and the tt pair produced by the effective4-fermion contact interaction. 140 ! cos(-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 10.30.40.50.60.70.8 ) ! cos(-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 10.30.40.50.60.70.8 Polarisation of the topZ’ ModelSM
Z’ Model : A=0.78SM : A=0.50
Figure 4:
Distribution of cos ( θ ) for the Z (cid:48) Model with M Z (cid:48) =
800 GeV compared to the SM. ) [GeV]tM(t400 600 800 1000 1200 1400 1600 N u m b e r o f eve n t s / ( G e V ) -2 -1
10 ) [GeV]tM(t400 600 800 1000 1200 1400 1600 N u m b e r o f eve n t s / ( G e V ) -2 -1 =14TeVs and -1 L=200pbMz’=350 GeVMz’=500 GeVMz’=800 GeVMz’=1200 GeVSM
M(Z’) [GeV]400 600 800 1000 1200 1400 1600 )t(t T m ax i m u m o f E from Z’tt tspectator tRandom CombiSM ) versus M(Z’)t(t T maximum of E Figure 5: (a) Invariant mass of the tt pair coming from the Z (cid:48) compared with that from the SM four-topevents; (b) Position of the maximum of the E T distribution of the tt pair as a function of M Z (cid:48) . Reconstruction
Reconstruction of four top events is a challenge to the detector and event reconstruction. Thedecay of the top quarks gives rise to twelve fermions. To benefit from the same-sign leptonsignature two W bosons must decay to lepton-neutrino. The presence of two escaping neutri-nos then prevents a complete reconstruction of the twelve momenta. In the most abundantlyproduced final states, most of the remaining fermions will be quarks, giving rise to a large jetmultiplicity.The minimal approach to reconstruction merely registers the scalar sum of the transverse141 [GeV]tM(t400 600 800 1000 1200 1400 1600 1800 ] - [ G e V )t d M (t ! d ! ] - [ G e V )t d M (t ! d ! )tdistribution of M(t from the Z’ for M(Z’)=1.2TeVtt for M(Z’)=1.2TeVtspectator trandom combination for M(Z’)=1.2TeVSM = 500GeV " effective model for Figure 6:
Comparison of the different tt invariant mass distributions. energy of all final state objects. The H T distribution for a 500 GeV and 1 TeV Z’ resonance asdescribed in the previous section are shown in Figure 7. For sufficiently large resonance mass,i.e. for m Z (cid:48) = t ¯ tW ± + jets and t ¯ tW + W − .A further experimental signature of the four-top final states is the large b-jet multiplicitywhich can be used as a powerful tool to extract the signal even coming from a heavy resonanceas shown in Figure 7 and in [439]. Reconstruction of (some of) the top quarks in the event canprovide additional handles to reduce the background. [GeV] T H500 1000 1500 2000 2500 N u m b e r o f eve n t s / ( G e V ) T H500 1000 1500 2000 2500 N u m b e r o f eve n t s / ( G e V ) =14TeVs, -1 L = 10fb = 500GeV Z’ signal Mttpp -> tt - W + Wtpp -> t + jets ± Wtpp -> t + jets ± W - W + pp -> W + jets ± W ± pp -> W [GeV] T H500 1000 1500 2000 2500 3000 N u m b e r o f eve n t s / ( G e V ) T H500 1000 1500 2000 2500 3000 N u m b e r o f eve n t s / ( G e V ) -1 L = 100fb = 1TeV Z’ signal Mttpp -> tt - W + Wtpp -> t + jets ± Wtpp -> t + jets ± W - W + pp -> W + jets ± W ± pp -> W [GeV] T H500 1000 1500 2000 2500 3000 N u m b e r o f eve n t s / ( G e V ) T H500 1000 1500 2000 2500 3000 N u m b e r o f eve n t s / ( G e V ) =14TeVs, -1 L = 100fb = 1TeV Z’ signal Mttpp -> tt - W + Wtpp -> t + jets ± Wtpp -> t + jets ± W - W + pp -> W + jets ± W ± pp -> W Figure 7:
Total transverse energy after demanding n j ≥ , p T > GeV (first two plots) and in addition n b − jet ≥ (third plot). For a complete reconstruction of the t ¯ tt ¯ t final state, one must address the challenge of142ssigning 12 final state fermions to the four top candidates. Before exploring this very complexfinal state, we consider the reconstruction of t ¯ t events that is much better understood.For the reconstruction of t ¯ t pairs the W-mass constraint and a top mass constraint areused to find the correct pairing. This approach is quite successful for t ¯ t events and a bit less sofor the (simulated) t ¯ tH topology. An alternative approach presents itself when one considersthe reconstruction of t ¯ t events originating in the decay of a heavy resonance. Due to the boostof the top quark, its decay products are collimated in a narrow cone. This top mono-jet canbe identified as such by techniques revealing the jet substructure [70, 71, 339, 343, 437, 440].Importantly, for sufficiently large resonance mass the decay products of top and anti-top arecleanly separated. A simple assignment based on (geometrical) vicinity is sufficient to find thecorrect assignment of jets to top candidates. Thus, the ambiguities found in reconstruction of“tops at rest” disappear in regime of large top p T . Mtt (GeV)0 500 1000 1500 2000 2500 F r a c t i on o f e v en t s (a) mass (GeV)tt 0 500 1000 1500 2000 2500 - e v en t s / G e V / f b (b) reconstructed mass (GeV)0 500 1000 1500 2000 2500 - e v en t s / G e V / f b (c) Figure 8:
The probability as a function of resonance mass that final state fermions are correctly assignedto top and anti-top quarks in t ¯ t production (open circles) and t ¯ tt ¯ t production. The filled circles (triangles)indicate the probability to find two (four) correctly paired top quarks. The central panel shows theinvariant mass distribution of the two top quarks with highest p T in SM t ¯ tt ¯ t production (filled histogram)and for production through a 1.5 TeV KK gluon. The rightmost panel shows the invariant mass of thetwo reconstructed clusters with highest p T . To quantify this statement a parton level simulation of pp → X → t ¯ t has been analysed.Lepton+jets events are selected, where one of the W bosons decays to a lepton and a neutrinoand the second W boson decay to two jets. The neutrino is discarded and the momenta of theremaining five fermions is presented to the k T algorithm [441,442] for clustering . Clustering isconsidered correct whenever all decay products from the top (and anti-top) quark are clusteredtogether in a single jet. The result is represented with open circles in the leftmost plot offigure 8. For tops produced at rest the probability of correctly clustering the event is essentiallyequal to 0. For resonant t ¯ t production the probability to find the correct assignment increasesrapidly as the resonance mass is increased. The decay products of the top and anti-top quarkare collimated more and more in a narrow cone, while the top anti-top are emitted essentiallyback-to-back. Indeed, for a resonance mass of 1 TeV, the correct assignment is found in nearlyeighty percent of events. For a more exhaustive discussion, and results including a completedetector simulation the reader is referred to reference [444]. The implementation in FastJet [443] was used, with E-scheme recombination. The algorithms was used inexclusive mode, forcing it to return exactly two jets. The R-parameter was set to 2.5. For this, somewhat unusual,choice nearly all input objects are clustered into jets (rather than included in the “beam jets”). t ¯ tt ¯ t production, the simple clustering has to deal with amuch denser topology and is much less successful. As shown with blue triangles in figure 8the probability to find a perfectly clustered event is less than 10 % over the entire mass rangestudied here. Of course, the decay of a heavy resonance leads to only one pair of stronglyboosted top quarks, while the p T of the associated (spectator) top quarks remains relativelysmall. The third curve (filled circles) in 8 represents the probability that at least two tops out offour are clustered correctly. This probability is quite large even for relatively small resonancemass, reaching approximately 60 % for a 1 TeV resonance.The mass of the resonance is reconstructed as the invariant mass of the two objects withhighest p T in the event. At the parton level this yields good results: the combination of the twotop quarks with highest p T yields the distribution of the central panel of figure 8. The resonanceclearly stands out on top of the SM four top production (black). Applying the same criterionto the top quark candidates reconstructed by the clustering algorithm, the distribution in therightmost panel is obtained. Obviously, the resonant signature is washed out by false combina-tions and the energy carried away by the escaping neutrinos. Still, the signal and backgrounddistributions can clearly be distinguished.The additional handle of highly boosted top quarks is found to be quite useful to reducethe combinatoric problem of four top events. Reconstruction of a resonant signature may wellbe feasible, thus turning the counting experiment into a resonance search. Given the simple-minded nature of this attempt to reconstruct this complex final state this result must be consid-ered as encouragement to develop a more sophisticated approach. Conclusions
The four top final state is sensitive to new physics that is relatively unconstrained by precisionmeasurements at LEP or resonance searches at the Tevatron. Examples are models where thetop quark is composite, or where a new heavy particle couples strongly (or exclusively) to topquarks.Reduction of Standard Model processes is achieved primarily through the requirement oftwo same-sign leptons. The small signal cross-sections (typically 10s of fb) render a countingexperiment susceptible to large uncertainties due to large ( t ¯ t + jets) backgrounds.Partial or complete reconstruction of the event enhances the robustness of the measure-ment. In this contribution we have explored the reconstruction of this complex final state, atthe parton level, in the case where a t ¯ t pairs originates in the decay of a heavy resonance. Theboost of the top quarks is found to greatly reduce the combinatorics involved in assigning thefinal state fermions. Even with the limited performance of our simple-minded algorithm, a masspeak may be reconstructed for a resonance mass of (or greater than) 1.5 TeV. Reconstructionthus provides a way to distinguish signal and SM background. Acknowledgements
We thank Emmanuel Bussato and Javi Serra for discussions.144 ontribution 13LHC sensitivity to wide Randall–Sundrum gluon excitations
G. Brooijmans, G. Moreau and R.K. Singh
Abstract
We apply the results recently obtained by the ATLAS collaboration inthe reconstruction of high mass t ¯ t resonances to Kaluza–Klein excita-tions of gauge bosons as predicted in Randall–Sundrum models withfields propagating in the bulk. The resulting ATLAS sensitivity to suchsignals is determined. Introduction
Randall–Sundrum (RS) models [289] of extra dimensions are an attractive approach to dealingwith the hierarchy problem (for a short review, see Contribution 9 of these proceedings). If, fur-thermore, standard model fermions and gauge bosons are allowed to propagate in the bulk, themodel can offer solutions to other major open questions, such as for example the existence ofdark matter [297,298,445]. An interesting way to constrain such models is to hypothesize [427]that they are the source of the deviation between the standard model prediction [446] and ex-perimental measurement at LEP and elsewhere [447] of the forward-backward asymmetry A bF B in Z boson decays to bottom quarks. A study of such a scenario [341] and more generally thegeometrical mechanism generating a large top quark mass [272] implies that the principal LHCsignature is the production of broad, high mass resonances (mainly due to the Kaluza–Kleinexcitation of the gluon) decaying primarily to t ¯ t pairs. Indeed, electroweak precision tests in-directly force the first Kaluza–Klein mass of the gluon to be typically larger than ∼ TeV in acustodially protected framework [321, 424, 426].Due to the collimation of top quark decay products at large top quark momentum, thereconstruction of high mass t ¯ t resonances requires the development of new experimental ap-proaches [339, 346]. The ATLAS collaboration recently released a full simulation study [449]describing the effectiveness of such new techniques in the reconstruction of narrow t ¯ t res-onances. In this note, these results are applied to the four concrete scenarios described inRef. [341] to estimate the integrated luminosity required to exclude these models. Simulation
Both signal and t ¯ t continuum events are generated using the S HERPA [450] event generator forproton-proton collisions at √ s = 14 TeV. All four scenarios from Ref. [341] are considered.These scenarios correspond to different localizations of the bottom/top quarks along the extra-dimension, all typically reproducing the bottom/top quark masses. These different localizationslead to variations in the wave function overlaps between the bottom/top quarks and the Kaluza–Klein excitation of the gluon (whose profile is peaked on the so-called TeV-brane), and in turn The production of the Kaluza–Klein excitations of the gluon was also studied in [345, 448]. < m t ¯ t < < m t ¯ t < Table 1:
Cross-sections for production of t ¯ t pairs with 2 < m t ¯ t < Experimental reconstruction efficiencies
The ATLAS collaboration released a study [449] of the reconstruction efficiency of high mass t ¯ t resonances in the lepton ( e and µ ) plus jets channel using splitting scales obtained fromreclustering the jets using the k ⊥ jet clustering algorithm. In this analysis, fully hadronic topquark decays were identified using a combination of jet mass and the first three k ⊥ splittingscales into a likelihood variable y L . A combination of the fraction of visible top mass carriedby the lepton [69] and relative p T of the lepton w.r.t. the jet were used to tag semileptonic topquark decays. One of the main conclusions of the analysis is that after cuts, the by far dominantbackground is the standard model t ¯ t continuum. Backgrounds from QCD multijet and non- t ¯ tW boson plus jets are found to be substantially smaller. For the study described in this paper,the ATLAS operating point chosen is the one with a cut on the hadronic top likelihood variable y L > t ¯ t selection efficiency for e, µ + jets events is parametrized as a function ofthe hadronically decaying top quark’s transverse momentum ( p topT ) with a linear increase from0% to 35% for p topT from 500 to 900 GeV, and a constant value at 35% for values p topT > t ¯ t invariantmass is then smeared by 5% to reflect the resolution found in the ATLAS study. Results
The semi-frequentist
CLs method [451] based on a Poisson log-likelihood test statistic is usedto determine the sensitivity after taking into account branching ratios and reconstruction effi-ciencies. The invariant mass distribution between background-only and signal + background arecompared, taking into account flat systematic uncertainties on the integrated luminosity (6%),reconstruction efficiency (10%) and background cross-section (15%).The resulting luminosities (for √ s =
14 TeV) to exclude the model points at 95% C.L.are given in Table 2. In all cases these are well below 10 fb − , showing these models should beaccessible in the first few years of LHC running at √ s =
13 or 14 TeV. Conclusions
The LHC reach for broad, high mass excitations of the gauge bosons decaying to t ¯ t final statesin Randall–Sundrum models has been investigated using the results of a full simulation study146ignal Model Integrated Luminosity for 95% C.L. Exclusion (fb − )E1 + SM 2.5E2 + SM 5.4E3 + SM 1.8E4 + SM 6.7 Table 2:
Required integrated LHC luminosities for 95% C.L. exclusion of the model points for √ s = of high p T top quark reconstruction. Using the lepton plus jets channel, the LHC experimentsshould be sensitive to such models with integrated luminosities smaller than 10 fb − collectedat √ s = 14 TeV.
Acknowledgements
The authors would like to thank the Les Houches workshop organizers for a very stimulatingand enriching workshop. 147 ontribution 14Effects of nearby resonances at colliders
G. Cacciapaglia, A. Deandrea and S. De Curtis
Abstract
We describe propagators for particle resonances taking into accountthe quantum mechanical interference due to the width of two or morenearby states with common decay channels, incorporating the effectsarising from the imaginary parts of the one-loop self-energies. The in-terference effect, not usually taken into account in Montecarlo gener-ators, can modify the cross section or make the more long-lived reso-nance narrower. We give examples of New Physics models for whichthe effect is sizable for collider physics. Introduction
In the following we consider a generalisation of the Breit–Wigner description [452] whichmakes use of a matrix propagator including non-diagonal width terms in order to describe phys-ical examples in which these effects are relevant. Indeed for more than one meta-stable statecoupled to the same particles, loop effects generate mixings for the masses as well as mixedcontributions for the widths (imaginary parts). In general a diagonalisation procedure for themasses (mass eigenstates) will leave non-diagonal terms for the widths. Usually non-diagonalwidth terms are discarded. When two or more resonances are close-by and have common decaychannels such a description is not accurate. The usual Breit–Wigner approximation amounts tosum the modulus square of the various amplitudes neglecting the interference terms and this isthe usual procedure in Montecarlo generated events. When there are common decay channelsand the widths of the unstable particles are of the same order of the mass splitting, the interfer-ence terms may be non-negligible. In the following we shall consider models of physics Beyondthe Standard Model (BSM) in which new resonances play a crucial role. Based on these resultswe suggest that a proper treatment should be carefully implemented into Monte Carlo gener-ators as physical results may be dramatically different from a naive use of the Breit–Wignerapproximation. The formalism
We discuss here only the formalism for scalar fields which gives a simpler overview of theproblem without the extra complications of the gauge and Lorentz structure of the general case.A more detailed analysis can be done also including vector resonances [453].For a system involving many fields, which do couple to the same intermediate particles,loops will generate mixings in the masses, but also out-of-diagonal imaginary parts. In generalthe real and imaginary parts will not be diagonalisable at the same time. The kinetic function isin general a matrix : ( K s ) lk = ( p − m l ) δ lk + i Σ lk ( p ) . (1)148We are considering the imaginary part only, the real one is used to renormalise the masses.)The propagator of the fields can be defined as the inverse of the matrix: i (∆ s ) lk = i (cid:0) K − s (cid:1) lk . (2)For simplicity we give here the two-particle case : i ∆ s = iD s (cid:18) p − m + i Σ − i Σ − i Σ p − m + i Σ (cid:19) , (3)where D s = ( p − m + i Σ )( p − m + i Σ ) + Σ Σ . (4)For vanishing Σ and Σ , the propagator is diagonal and it reduces to two independent Breit–Wigner propagators with m i Γ i = Σ ii ( m i ) .However, the narrow width approximation is not valid if the off-diagonal terms are sizablecompared with the mass splitting. Defining M = m + m and δ = m − m , the poles ofthe propagator (zeros of D s ) are: ˜ m ± = M − i Σ + Σ ± i (cid:112) (Σ − Σ + 2 iδ ) + 4Σ Σ . (5)Note that the value of the masses is modified by the presence of the off-diagonal terms due to theimaginary part of the square root, at the same time the widths are affected. More importantly,the off-diagonal terms in the propagator will generate non-negligible interference, which can bein turn constructive or destructive. Numerical examples
We first study two heavy Higgses where both the scalars develop a vacuum expectation value(VEV) and therefore couple to the W and Z gauge bosons. This situation is common in super-symmetric models where two Higgses are required by writing supersymmetric Yukawa interac-tions for up and down type fermions, and generic two Higgs models. The interference betweennear degenerate Higgses has been studied in [454–456] focusing in CP violation effects.The couplings of the two CP-even Higgses to gauge bosons can be written as λ W W H = g m W cos α , λ W W H = g m W sin α ,λ ZZH = g m Z cos θ W cos α , λ ZZH = g m Z cos θ W sin α ; (6)where α is a mixing angle taking into account the mixing between the two mass eigenstates andthe difference between the two VEVs.Here we are interested in a generic production cross section of the two nearby Higgseson the resonances, with decay of the Higgses into gauge bosons (either W W or ZZ ). Theamplitude of this process is proportional to the resonant propagator weighted by the couplingsgiven in eq.(6). In the case we are considering, the common decay channels can give off-diagonal terms in eq.(3) which are sizable compared with the mass splitting. Here we assumethat the coupling to the initial particles are the same : (cid:12)(cid:12) (∆ s + ∆ s ) cos α + (∆ s + ∆ s ) sin α (cid:12)(cid:12) . (7)149
00 350 400 450 500 s s s s Figure 1:
Plots of the production cross section (in arbitrary units) of two nearby Higgses decaying into gaugeboson pairs for the naive Breit–Wigner (blue-dashed) and exact mixing (red-solid). The mass of the first resonanceis fixed to
GeV, the splitting respectively 50, 25, 10 and 5 GeV and α = π/ .
300 350 400 450 500 s s Figure 2:
Plots of the production cross section (in arbitrary units) for seven nearby Higgses equally coupled toSM gauge bosons: the naive Breit–Wigner (blue-dashed) bump reduces to a row of seven dwarfs when the exactmixing (red-solid) is taken into account. The mass of the first resonance is fixed to
GeV, the splitting betweenthe six Higgses respectively 10 and 5 GeV.
In Fig. 1, we plot this quantity in arbitrary units and compare it with the Breit–Wigner approx-imation: we fix m H = 400 GeV, and vary the splitting from 50 to 5 GeV. For simplicity, inthe following we will assume α = π/ , so that the two scalars have the same couplings. Theexact treatment of the resonances unveils a destructive interference that can drastically reducethe cross section.This effect can be even more important for scenarios with a large number of scalars aspredicted in some string models. Our analysis can be easily extended to an arbitrary numberof Higgses. Let’s take for example the couplings to the gauge bosons to be given by g SM / √ N ,150here g SM is the SM coupling of the gauge bosons and N is the number of Higgses. In Fig. 2we plot the cross section for seven nearby Higgses, with the first one at 400 GeV and the othersat a distance of 5 and 10 GeV, the width of each being 6.2 GeV. From the plot it is clear that thedestructive interference reduces the giant resonance (which is not distinguishable from a singleHiggs, once the experimental smearing is taken into account) to a bunch of gnometti (dwarfs),which will be very hard to detect. The cross section is in fact reduced by a significant factorwith respect to the naive expectation, and the smearing will wash out the peak structure. It isintriguing to compare this analysis with Un-Higgs models [457,458], where the Higgs in indeeda continuum: such behaviour may arise from the superposition of Kaluza-Klein resonances inextra dimensional realisations or deconstructed models.Another striking example involving vector resonances is given by Higgsless models [459,460], where the first two neutral resonances are nearly degenerate, and they correspond to thefirst KK excitation of the Z and of the photon. The masses can be approximated by m Z (cid:48) (cid:39) m KK + 4 m Z , m A (cid:48) (cid:39) m KK , (8)so that the mass difference is very small: m Z (cid:48) − m A (cid:48) (cid:39) m Z m KK ∼
16 GeV · (cid:18) m KK (cid:19) . (9)In terms of the parameters of the warped geometry ( R is the curvature, R (cid:48) the position of theInfra-Red brane in covariant coordinates): m KK ∼ . R (cid:48) , m W = 1 R (cid:48) log R (cid:48) R ; (10)therefore, given the value of the curvature R , the KK mass ( R (cid:48) ) is determined by the W mass.The determination of the couplings is more involved and we refer to [453] for details.We consider the following processes: Drell–Yan production and decay into gauge bosons W + W − (DY), Drell–Yan production and decay into a pair of leptons (Leptonic) and vectorboson fusion production followed by decay into gauge bosons (VBF). As in the scalar case, theamplitudes at the resonance are proportional to the propagators weighted by the couplings withthe incoming and outcoming particles.In Figure 3 we plot, for illustrative purposes, the squared matrix element of the three reso-nant production channels for A (cid:48) and Z (cid:48) as function of √ s for three different cases: m KK = 1000 GeV,
GeV and
GeV. For large masses, the effect of the interference is very importantand it can affect the value of the cross section significantly. We give in the table the ratio of thearea under the peaks in the figure obtained by the exact formula and the BW case. This roughlycorresponds to the ratio of the integrated cross sections. M KK =
60 980 1000 1020 1040 1060 1080 1100 s s s s s s s s s Figure 3:
Plots of production cross section (in arbitrary units) of the two low-lying neutral resonances of theHiggsless model for the naive Breit–Wigner (blue-dashed) and exact mixing (red-solid). The rows correspond(from top to bottom) to DY, Leptonic and VBF; the columns (from left to right) correspond to m KK = 1000 GeV,
GeV and
GeV. Conclusions
We have shown that for two or more unstable particles, when there are common decay channelsand the masses are nearby, the interference terms may be non-negligible. This kind of scenariois not uncommon in models of New Physics beyond the Standard Model, especially in models ofdynamical electroweak symmetry breaking or in extended Higgs sectors. In models with multi-Higgses and in Higgsless models with near degenerate neutral vector resonances, we showedthat interference induced by the off-diagonal propagators are very important and they can eithersuppress or enhance the total cross sections on resonance depending on the relative sign of thecouplings to the initial and final states. The interference effects can be crucial to study thephenomenology of such models at the LHC, and to determine its discovery potential. A propertreatment should be carefully and systematically implemented into Monte Carlo generators usedto study BSM models. 152 idden sectors ontribution 15An exotic photon cloud trigger for CMS
C. Henderson
Abstract
We propose a novel trigger to be sensitive to a new kind of beyond-Standard-Model physics: a ‘photon cloud’, consisting of ∼ GeVof transverse energy emitted through many soft photons ( (cid:46) GeVeach). Such an exotic event could potentially be overlooked by con-ventional trigger configurations. We demonstrate that by considering asimple new variable, Sum- E T in the electromagnetic calorimeter, whichis straight-forward to calculate in a high-level trigger, an experimentcould be sensitive to this photon-cloud scenario. We estimate rates forthis trigger for the expected LHC early luminosity scenario and showthat the proposal is feasible. This trigger is planned to be implementedin the CMS experiment for the 2010 LHC running period.At the Large Hadron Collider (LHC), the eventual goal is to collide proton bunches at rateup to 40 MHz. However, the rate of events which can be written to permanent storage is limitedto ∼
300 Hz. Therefore a sophisticated trigger system is required to make an online selectionof the most interesting collision events to be recorded.At the Compact Muon Solenoid (CMS) detector [162, 359], a two-level trigger systemis employed. At Level 1 [461], lower resolution information (with full eta-phi coverage) fromthe calorimeter and the muon chambers is used to create particle candidates, and specially-programmed firmware selects events at a rate up to 100 kHz. A Level 1 accept initiates thecomplete detector readout, and the full event is made available to the second trigger stage,the High-Level Trigger (HLT) [462]. This comprises essentially the full event reconstructionsoftware, running on a large PC farm. Standard particle objects such as jets, photons, muons,electrons, etc. . . are all reconstructed, and form the basis for further event selection. Generallyspeaking, events containing high- E T particles or combinations of particles are chosen, with thegoal of selecting the most interesting events for permanent storage, up to a maximum rate of ∼
300 Hz. The performance of the CMS trigger system in cosmic-ray operations during 2008 isdescribed in [463, 464].The trigger selection is therefore a crucial part of the experiment - events that do notpass the trigger can never be analysed. Thus it is of critical importance that we consider allpossible types of collision event, including those arising from exotic new physics beyond theStandard Model, and ensure that they are not being unwittingly rejected by the online triggerselection. Here we consider an unusual type of event topology that could potentially be missedby conventional trigger configurations, and propose a novel trigger selection to remedy this.The unusual event topology that we will consider is a ‘photon cloud’, which we take tobe a large amount of transverse energy ( (cid:38) − GeV) that is emitted as a large number of154oft photons ( (cid:46) GeV each). An exotic physics scenario which could potentially produce sucha photon cloud is described in [465]. The authors consider an extension of the Standard Modelthat contains a new asymptotically free SU(N’) gauge force (called QCD’) and new fermions q (cid:48) which are charged under this force (and also carry some SM gauge quantum numbers, sothey can be pair-produced in sufficiently high-energy collisions). For such a gauge field, a widerange of values for the confinement scale Λ can be considered ‘natural’, since the confinementscale is related to the fundamental gauge coupling g defined at a scale µ by: Λ = µ e − π /bg (1)where b is the one-loop coefficient of the SU(N’) beta function.The novelty in this scenario arises when, unlike QCD, one considers the confinementscale Λ << m q (cid:48) , where we take m q (cid:48) to be a few hundred GeV or greater (above current ob-servable limits). This scenario was first discussed in [466] and later revived in the context ofthe LHC in [467], where the new fermions were given the name “quirks”. The model of [465]specifically considers scalar quirks (“squirks”), which arise in the context of a folded super-symmetry scenario [468], where the superpartners which cancel the ultraviolet divergences inthe Higgs mass due to SM quarks, are charged under the new QCD’ force rather than havingnormal QCD color. Squirks could be pair-produced in LHC collisions via weak interactions (aDrell–Yan process, or gauge boson fusion), and their high mass means they will typically besemi-relativistic. The relationship Λ << m q (cid:48) can then result in striking new phenomenology.There will be a QCD’ string connecting the squirk-antisquirk pair, but unlike normal QCD,string fragmentation cannot occur because the energy density in the string, Λ , is much lessthan the typical energy density ( ∼ m q (cid:48) ) needed for quirk pair production in a standard ‘hadro-nisation’ mechanism. Instead, the heavy squirks will continue to separate, until eventually alltheir kinetic energy has been transferred into the stretched QCD’ string. At this point, the stringtension will cause the squirks to start oscillating, forming a “squirkonium” bound state. If thesquirks are electrically charged, this oscillation will then radiate photons, with characteristicfrequency given by: ω ∼ π Λ m q (cid:48) (2)It is possible that a large fraction of the squirkonium energy could be radiated this way, andassuming a squirk mass of 500 GeV, any value of the confinement scale Λ (cid:46) GeV will resultin photons with typical energy (cid:46) GeV.Such photon cloud events could represent a striking signature of physics beyond the Stan-dard Model. The problem is that in such an event, no individual detector region has very highactivity and therefore no typical high-energy object would be seen by the trigger. Triggering onthis kind of event therefore requires consideration of the global properties of the event, not justlocal regions of high-activity.We propose to introduce a new variable at the trigger level: a sum of the transverse energyobserved in all channels of the electromagnetic calorimeter (ECAL). This is a modification ofthe standard Sum- E T trigger, which typically sums contributions from both the electromag-netic and hadronic calorimeters. Not only is this new ECAL-only Sum- E T measure moredirectly sensitive to the photon cloud signature, it also takes advantage of the reduced back-ground from the electromagnetic component of minimum bias proton-proton collisions relativeto the hadronic component. In addition, by considering only one sub-detector, it is less affected155y detector noise, and it is particularly well suited for the CMS detector, which has a high-performance electromagnetic calorimeter [469]. These combined benefits allow for a lowertrigger threshold, increasing our sensitivity to potential exotic physics.At the High-Level Trigger stage in CMS, the full detector readout is available and essen-tially complete ‘offline-like’ event reconstruction can be performed. Thus it is straight-forwardin the software to construct the ECAL-only Sum- E T variable for event selection at the HLT.Ideally one would do a similar thing at the Level 1 stage also. However, unfortunately in thiscontext, the design of the CMS Level 1 trigger system is such that only the complete calorime-ter Sum- E T can be computed - the electromagnetic and hadronic Sum- E T components are notavailable separately. Fortunately though, we can tolerate a higher trigger rate at Level 1, sincefurther refinement will be done at the HLT. A cut at Level 1 on the complete calorimeter Sum- E T of ∼ GeV is expected to be fully efficient for the signal we are considering, whilestill producing an acceptable rate. For the remainder of this paper, we will discuss only thefinal high-level trigger selection. For simplicity, we choose to focus only on the ECAL Barrelsub-detector, which spans the central pseudorapidity region | η | < . .Since the HLT is required to be able to accept incoming events at a rate up to 100 kHzwithout incurring deadtime, this imposes a limit on the average total time which can be spentprocessing each event. The design goal is that the average HLT event processing time not exceed50 ms [470]. Unlike other region-of-interest based triggers, our proposed new global triggerrequires the raw data from all ∼ , ECAL barrel channels to be unpacked; it is thereforeimportant to verify that the time spent doing this will not be prohibitive for HLT operations. Wehave studied the CPU-time required for the new selection on minimum bias Monte Carlo events,and verified that it is well within the acceptable limits. As well as checking the average time,we must also ensure that high-occupancy events do not result in unacceptably long processingtimes. An excellent opportunity to test this was provided during the LHC commissioning phasewith so-called ‘beam splash’ events where essentially every calorimeter channel is illuminatedby high-energy particles. We studied the time taken for this trigger path on the data collectedduring ‘beam splashes’ in November 2009. The time taken in these extreme high-occupancyconditions forms an upper bound for the trigger path, and the result is found to be well withinthe acceptable range for operations.The background to this photon cloud signature will come from rare proton-proton colli-sion events which contain a large number of final-state photons, either from prompt productionor from π decays. We estimate this background using minimum bias events generated by the PYTHIA
Monte Carlo event generator [271].Based on the Monte Carlo generator-level information, the distribution of the Sum- E T of all final-state photons in the simulated event is shown in Figure 1. In combination with theanticipated accelerator luminosity profile, this will allow us to estimate the expected triggerrate as a function of selected threshold value. Our goal is to choose a trigger threshold thatcorresponds to a trigger rate < Hz (in order to have minimal impact on the limited triggerbandwidth, and hence the rest of the CMS physics program).Sum- E T triggers are especially sensitive to the effect of multiple minimum bias collisionswithin a single bunch crossing, a phenomenon known as pileup. Because the trigger merelysums the total energy deposited in the calorimeter, it cannot distinguish the separate contribu- As part of the accelerator commissioning, the beam is fired into the collimators, generating a large spray ofsecondary particles that can be seen in the detector. These events are called ‘beam splashes’. ntries 3.137568e+07Mean 3.593RMS 5.3Overflow 0 (GeV) T E ! MC Photon 0 50 100 150 F r ac t i on o f C r o ss - sec t i on -8 -6 -4 -2
10 1
Entries 3.137568e+07Mean 3.593RMS 5.3Overflow 0
Figure 1:
Distribution of the Sum- E T of final-state photons in PYTHIA minimum bias collisions, basedon the generator-level information.
Entries 3e+08Mean 3.85RMS 5.227Overflow 52 (GeV) T E ! MC Photon 0 50 100 1501 Entries 3e+08Mean 3.85RMS 5.227Overflow 52
N Collisions = 1
Entries 3e+08Mean 7.702RMS 7.393Overflow 101 (GeV) T E ! MC Photon 0 50 100 150110 Entries 3e+08Mean 7.702RMS 7.393Overflow 101
N Collisions = 2
Entries 2e+08Mean 11.55RMS 9.055Overflow 117 (GeV) T E ! MC Photon 0 50 100 150110 Entries 2e+08Mean 11.55RMS 9.055Overflow 117
N Collisions = 3
Entries 1e+08Mean 15.4RMS 10.45Overflow 93 (GeV) T E ! MC Photon 0 50 100 150110 Entries 1e+08Mean 15.4RMS 10.45Overflow 93
N Collisions = 4
Entries 1e+08Mean 19.25RMS 11.69Overflow 150 (GeV) T E ! MC Photon 0 50 100 150110 Entries 1e+08Mean 19.25RMS 11.69Overflow 150
N Collisions = 5
Entries 1e+08Mean 23.11RMS 12.8Overflow 244 (GeV) T E ! MC Photon 0 50 100 15010 Entries 1e+08Mean 23.11RMS 12.8Overflow 244
N Collisions = 6
Figure 2:
Effect of additional collisions per bunch crossing on the total photon Sum- E T distribution. tions from the individual collisions. Figure 2 shows how the total photon Sum- E T distributionevolves as a function of the number of separate minimum bias collisions per bunch crossing. Toillustrate the point, consider that a total of, say, 100 GeV in the detector can be obtained by two157imultaneous collisions if both generate 50 GeV each, or one generates 60 GeV and the other40 GeV, or 70 GeV and 30 GeV, and so on. The probability to exceed a given Sum- E T thresholdtherefore grows nonlinearly with the number of independent collisions per bunch crossing, asdisplayed in Figure 3. N collisions0 1 2 3 4 5 6 7 P r ob a b ili t y t o excee d S u m - E t t h r es ho l d -12 -10 -8 -6 -4 -2
10 1 > 50 GeV> 80 GeV> 100 GeV> 120 GeV> 140 GeV
Figure 3:
Probability to exceed a given threshold value of Sum- E T as a function of the number ofindependent collisions per bunch crossing. The likelihood of there being N collisions per bunch crossing follows a Poisson distri-bution characterised by < n > , the mean number of interactions per crossing. < n > can bedetermined from the expected LHC luminosity parameters, in particular the instantaneous lumi-nosity and the number of colliding bunches. A benchmark for the early part of the 2010 runningperiod is for the LHC to reach an instantaneous luminosity of cm − s − with × colliding bunches [471]. Taking a proton-proton cross-section of mb at TeV, this corre-sponds to < n > ≈ . . Figure 4 shows the Poisson-weighted probability to exceed a giventrigger threshold, assuming < n > = 0 . . Fortunately, we find that the reduced likelihood ofhaving an extra collision more than compensates for the increase due to combinatorics, and theobserved rate hence remains under control. Estimated trigger rates for the cm − s − lumi-nosity scenario as a function of the chosen threshold value are shown in Table 1. Note that thevalues shown here are just from the generator-level photon Sum- E T – the actual trigger thresh-old value that should be used in the experiment must also account for detector effects such asnoise and resolution, and will therefore be somewhat higher. However, it is clear from this tablethat we can expect to be sensitive to potential photon clouds with Sum- E T ∼ GeV whilemaintaining a trigger rate < Hz, which was our goal.In summary, we have considered an unusual potential event topology: a ‘photon cloud’,consisting of ∼ GeV of transverse energy emitted through many soft photons ( (cid:46) GeVeach). Such an event could arise from the production and decay of ‘squirks’ in a folded super-158 collisions0 1 2 3 4 5 6 7 P o i ss on - w e i gh t e d P r ob a b ili t y -15 -12 -9 -6 -3 > 50 GeV> 80 GeV> 100 GeV> 120 GeV> 140 GeV Figure 4:
Poisson-weighted probability to exceed a given threshold value of Sum- E T as a function ofthe number of independent collisions per bunch crossing, assuming < n > = 0 . . Photon Sum- E T (GeV) Trigger Rate (Hz)50 28580 7100 1.5120 0.4140 0.15 Table 1:
Estimated rates as a function of chosen trigger threshold, for LHC cm − s − startupluminosity scenario. symmetry scenario. A photon cloud event could be missed by conventional trigger selections,because no individual detector region has particularly high activity. We have proposed a newvariable, Sum- E T in the electromagnetic calorimeter, which can be easily computed in an exper-iment’s high-level trigger and which would be sensitive to this unusual new physics signature.We have estimated the expected rates for this new trigger for the LHC early luminosity scenario,and shown that the proposal is feasible. This trigger is being implemented for the CMS detectorand it is planned to be operational for high-energy LHC collisions in 2010.159 ontribution 16Models and benchmarks for long-lived exotica production atthe LHC/Tevatron M.J. Strassler and I. Tomalin
Abstract
Numerous theories predict that long-lived exotic particles may be pro-duced at the LHC/Tevatron. These yield highly unusual signatures,which can prove a major challenge for triggering and event reconstruc-tion. This note highlights some of these theories, and uses them to de-fine simple analysis benchmarks, as might be appropriate for early LHCsearches. The note also defines reconstruction/trigger benchmarks, whichare deliberately more difficult to find, designed to stress the detector andso identify weaknesses to be remedied. Motivation
New long-lived particles, with lifetimes such that their decays commonly occur at distancesoutside the beam-pipe but inside the detector volume, do not generally arise in the most popularmodels of electroweak symmetry breaking. Perhaps for this reason, they were little studied byLHC/Tevatron theorists and experimentalists until recently. Only those cases arising in partic-ularly simple models of gauge-mediated [472] supersymmetry, with neutral LSP’s decaying tophoton plus gravitino, were covered by early LHC studies.However, looking across the literature, one finds plenty of models in which a long-livedparticle appears; see for example [473–475]. Also, it has been emphasized recently in the lit-erature that long-lived particles arise very commonly in models with hidden sectors and a massgap (hidden valleys [476]) in which a number of new particles may naturally arise with a varietyof long lifetimes. (The example of the QCD spectrum, which has many long-lived hadrons withwidely varying decay lengths and final states, is instructive.) Moreover, it has also been shownthat finding long-lived particles with current hadron colliders can be exceptionally difficult,because of challenges in triggering, reconstruction and detector backgrounds (e.g. secondaryinteractions.) This motivates a serious effort to ensure that the collaborations have a plan toperform searches for long-lived particles and ensure that the detector hardware and software isused in a way which helps, rather than hinders, this effort.Toward this end, it was decided that benchmark models were needed as targets for theexperimental collaborations. We describe the current status of that enterprise here. Our effortsare organized along two different lines, with differing goals.First, we aim to provide benchmarks for long-lived particle searches appropriate for earlydays at the LHC. The goal in this case is to provide simple models, with moderately large cross-sections and with signatures that are relatively uncomplicated. These models have a smallnumber of variable parameters on which limits could be placed if no discovery is made. Wewill propose specific possibilities below. 160econd, we aim to provide benchmark models that would serve as a stress-test of thetrigger system and reconstruction software. The goals in this case are to check whether a chal-lenging signature might cause problems either for the trigger pathways or the reconstructionsoftware, or even the methods of data storage. The underlying concerns are that a long-livedparticle signature might (a) be rejected by the trigger, (b) cause the event to fail quality controlcuts, or (c) be inefficiently reconstructed. Even where events are successfully kept, poor recon-struction can confuse subsequent event skims, so making it difficult to select a sample of eventson which detailed analysis should be performed. The variety of possible signatures and thecomplexity of the software involved make it difficult to guess whether the currently designedsystem is robust without a test. We have considered models with signatures that are in somecases simple, in others exceptionally complex, though always realistic. For each model we aremaking available a data set that will serve in such a test, as well as providing, where possible,the information as to how to simulate the model without special-purpose software. Classes of models used in the benchmarks
In all the models chosen, long-lived particles either arise when a visible-sector particle decaysslowly into a hidden sector, or when a hidden-sector particle decays slowly to standard modelparticles. We will first discuss the production mechanisms for the long-lived particle(s). Thenwe will address the final states emerging in the decay of the long-lived particle(s).
The models chosen for the benchmarks produce long-lived particles via three mechanism: (1)decays of a singly-produced light resonance (such as a Higgs boson or Z (cid:48) ) into a hidden sectorwith long-lived particles [476,477]; (2) R-parity-conserving decays of the lightest super-partnerof any standard-model particle (LSP), in the case that the LSP is heavier than the lightest super-partner in the hidden sector (vLSP) [472, 478, 479]: and (3) annihilation of quirks [466, 467,476, 480], confined particles charged under standard model and hidden sector gauge groups,to hidden valley gluons (also called “infracolor”.) In each case, the number of hidden sectorparticles produced, and the lifetimes of the long-lived particles, are strong functions of theessential dynamics in the hidden sector and of the parameters of the model. However, theproduction rate for a Higgs boson, Z (cid:48) , LSP or other particles occurs as usual in the standardmodel or in its appropriate extensions by an extra U (1) and/or supersymmetry. For the usualreasons, these rates may be of order 1 – 10 pb for light resonances or supersymmetry, and oforder 100 fb – 1 pb for a Z (cid:48) or moderately heavy scalar. (These numbers are only guidelinesand can be exceeded in particular models.) Quirks are produced through usual standard modelpair production, and to estimate their cross-sections one need merely compare standard modelproduction rates for standard model fermions of the same mass and quantum numbers, possiblymultiplied by a hidden color factor (typically a small integer.)Let us note here that the Higgs scalar that dominantly gives mass to the W and Z bosonsthrough its expectation value will dominantly decay to W W and ZZ when these channels be-comes kinematically allowed. For this reason, only a light Higgs is likely to have exotic decaysto long-lived particles with a large branching fraction (above a percent). However, any scalarsor pseudoscalars which are not responsible for electroweak symmetry breaking (including forexample the CP-odd A boson in two-higgs doublet models such as supersymmetry) will not(dominantly) decay to gauge bosons and typically will have small widths to decay to standard161odel particles. This makes them susceptible,whatever their masses might be, to new couplingsto a hidden sector, and thus to new exotic decays.We should also note that when designing an analysis, care should be taken to assure thatquantum numbers are properly accounted for in any process considered. For example, just as ρ → π π is forbidden by Bose statistics, a process Z (cid:48) → XX where X is a real scalar isforbidden. However Z (cid:48) → X ¯ X where X is complex is allowed, as is ρ → π + π − . For simplicity we limit ourselves to scenarios that are both “natural” and popular. It must be em-phasized that other scenarios cannot be excluded on either theoretical or experimental grounds.All long-lived particles discussed below are charge- and color-neutral unless explicitly statedotherwise. Such particles, if sufficiently weakly coupled, are poorly constrained by experimentand may be extremely light, as emphasized for example in [476, 481, 482].New hidden-sector scalar or pseudoscalar particles tend to decay to the heaviest fermionpair available, due to helicity suppression and/or coupling proportional to standard model Yukawacouplings. For moderate masses these particles will decay dominantly to b ¯ b , and perhaps τ + τ − .An interesting special case occurs when the mass becomes of order 1 GeV, as for example inthe case of a “dark axion” motivated by current anomalies in dark matter experiments. In thiscase, among experimentally detectable decays, µ + µ − may be dominant.New hidden-sector vector or axial-vector particles usually decay to leptons and quarks ina generation-independent fashion, subject of course to kinematic constraints. In this case µ + µ − and e + e − may be the best channels, though the large rate for dijet decays may outweigh theirrelative difficulty. Again a special case occurs for masses of order or below 1 GeV. A vectorboson (such as a “dark photon” or “dark gluon”) at this mass, mixing with the photon, have (likedark axions) been proposed to explain astrophysics signals from the PAMELA/ATIC and nowFermi-LAT experiments [483, 484]. It would decay equally to µ + µ − and e + e − (unless below m µ .) The relative branching fractions for decays to leptons versus π + π − is a complicated butknown function of the particle’s mass, determined by electromagnetic couplings and by mixingsof the photon with hadronic resonances [485, 486]. Additional results on such particles appearin [487–489].New hidden-sector scalar or tensor particles may decay to gauge boson pairs, includinggluon pairs, photon pairs, and (when kinematically allowed) W + W − and ZZ , and perhaps Zγ .In some cases a decay to two Higgs bosons may be permitted. Branching fractions of thesevarious final states may vary widely.The case of a long-lived neutral LSP (or any analogous particle in models with KK-parity,T-parity, or other new global symmetries) offers two different possibilities. (a) Even without R-parity violation, the LSP may be an long-lived particle and decay in flight to a partly visible finalstate. Well-known examples include decay to a gravitino plus a photon, Z or Higgs boson [478].It is also possible for the decay to go to new hidden sector particles (as in many supersymmetrichidden valley models [479,485,488], of which many of the recent dark matter models [484,490]are examples) in which case their decay products might appear all at the point of the LSP decay.(b) Conversely, it is also possible that the LSP decays promptly into the hidden sector, producingamong other things a long-lived hidden-sector particle that may decay as described above.In all cases (except the LSP → γ + gravitino decay mode) quoted so far, the decay prod-ucts observed form a resonance — generally a new resonance, except in the case of decays of162he LSP to a Z boson (plus a gravitino). Like the LSP, new hidden sector particles also mighthave decays emitting a photon, Z or Higgs and a second hidden-sector particle [479, 480].However, more complicated decays, where the final state might produce invariant-mass edgesor endpoints rather than resonances, are possible. For example, a hidden sector particle, of anyspin, might decay to a second while emitting two standard model fermions ( X → f ¯ f X (cid:48) ), wherethe f ¯ f invariant mass is a continuum, with an edge or endpoint. Another interesting exampleis the model [34], which predicts Z (cid:48) → ν H ¯ ν H → ( (cid:96) − W + )( (cid:96) + W − ) , where ν H is a long-lived,heavy neutrino. Even more complex decays are possible, if for example, the decay X → f ¯ f X (cid:48) is followed by a prompt decay [476] such as X (cid:48) → f ¯ f or X → W + W − . In this case many stan-dard model particles will be emitted from a single vertex. The number of possibilities rapidlybecomes very large. Once there is experience with searching in the simpler scenarios, and itis clearer how robust the initial analyses are and where the gaps lie between them, these morecomplex signatures should be considered, and attempts made where necessary to check for theirpresence.One might ask if there are strong constraints on the masses and lifetimes of the newparticles. Unfortunately there are not. In general, in any fixed model, the lifetime of a particularparticle is often a strongly decreasing function of its mass (for roughly the same reason that themuon lifetime varies inversely with the fifth power of its mass) or of other parameters. However,across models there is no correlation between mass and lifetime. Even within a model theremay be very long-lived particles with large masses (just as B mesons live a bit longer than D mesons.) Thus one ought if possible to treat these parameters as independent, since otherwisemodel-based assumptions will limit the applicability of the results obtained. The decay of a visible sector particle into the hidden sector may lead to a final state with anynumber of long-lived decays, subject only to kinematical constraints. This is partly because ofthe wide variety of dynamics that can be present in hidden sectors, affecting the intrinsic mul-tiplicities of hidden particles produced, and partly because hidden sectors may contain severalnew particles with different lifetimes, some of which may decay promptly, others of which maybe stable or may decay far outside the detector.For example, models exist in which a Z (cid:48) may decay into two identical particles with iden-tical lifetimes (analogous to ρ → π + π − ). In this case one has two back-to-back particles withthe same average lifetime. But it may also decay to a single long-lived particle and a promptly-decaying particle (analogous to a → πρ ) or to a long-lived particle and an invisibly-decaying(or stable) particle. At the other extreme, it may decay to particles which have showeringor cascade decays that may lead to a very large number of hidden-sector particles being pro-duced, possibly producing many displaced vertices [476, 491]. The same is true if the LSPin a SUSY model decays promptly to long-lived particles in the hidden sector; the number ofvisibly-decaying long-lived hidden-sector particles produced may vary, and consequently thenumber in each event may range from one to a very large number [479, 488]. By contrast, inSUSY models where the LSP itself is long-lived [472, 475] the number of displaced vertices isgenerally equal to two (except when one LSP happens to live too long or too short a time for itsdisplaced decay to be detected.) 163 Analysis benchmarks
A key requirement for early-data analysis benchmarks is that they should be sharply defined,theoretically well-motivated, relatively simple experimentally, and contain a small number oftunable parameters. It should be possible to imagine a search strategy (or small number ofstrategies) that would make these models appropriate targets for analyses within the comingfew months.In all the cases listed below, the signatures are simple enough that it is very easy to imple-ment these models in MadGraph or other similar event generators. In particular, all decays inthese models are a sequence of two-body decays. To obtain precise limits, spin effects shouldbe properly included, since angular distributions will affect efficiencies. Z (cid:48) or H → B ¯ B → ( f ¯ f )( f (cid:48) ¯ f (cid:48) ) In this benchmark, B is a long-lived, neutral boson and f and f (cid:48) are Standard Model fermions.Experimentally, these events can be identified by reconstructing the displaced difermion ver-tices in the Tracker or using time-of-flight measurements. One can also require that very fewpromptly produced hadrons should be flying in a similar direction to the displaced fermions.One should trigger on the displaced fermions, to avoid assumptions about the rest of the event.Modifications to the trigger are often needed to accomplish this. Such analyses have been per-formed at the Tevatron [492–494].One should publish the measured the cross-section σ [ Z (cid:48) /H → B ¯ B → ( f ¯ f )( f (cid:48) ¯ f (cid:48) )] forindividual f ¯ f and f (cid:48) ¯ f (cid:48) species. As explained in Sect. 2.2, e + e − , µ + µ − and q ¯ q final states shouldgive better sensitivity to spin 1 bosons B , whilst τ + τ − or b ¯ b should be better for the spin 0 case.Searching for mixed states where one boson decays to leptons and the other to jets may also beuseful, since the leptons will facilitate triggering and suppress backgrounds.The cross-section measurements will depend on 3 parameters, which can be taken to bethe masses M Z (cid:48) and M B and the mean B lifetime (or decay length). Quoting results as asimultaneous function of all 3 variables is not practical, so one may instead show them as afunction of each in turn. (N.B. For analyses that rely on reconstructing the displaced verticeswith a Tracker, the reconstruction efficiency is a strong function of the mean decay length, so itthen makes sense to parametrize the results in terms of it, rather than the lifetime).The published measurements should be sufficiently complete, to allow one to subse-quently combine the different channels into a measurement of σ ( Z (cid:48) /H → B ¯ B ) . This steprequires theoretical assumptions about the branching ratios B.R. ( B → f ¯ f ). e.g. If B is apseudo-scalar, a simple model might assume these to be equal to those of a pseudoscalar Higgsat the corresponding mass (i.e. no decays to W W , ZZ ). Z (cid:48) or H → B ¯ B → ( f ¯ f )( f (cid:48) ¯ f (cid:48) ) It is worth considering two variations on the previous benchmark, both of which can be studiedwith only minor changes in analysis software. They are:1. As explained in Sect. 2.1, the B ¯ B system may originate from something other than aresonance decay (e.g. a SUSY cascade decay chain). One could therefore repeat theanalysis of the previous section, but without requiring the presence of a Z (cid:48) /H peak inthe reconstructed B ¯ B mass. This leads to a measurement of σ [ B ¯ B → ( f ¯ f )( f (cid:48) ¯ f (cid:48) )] , as a164unction of 3 principal parameters, which can be taken to be M B ¯ B , M B and the mean B decay length.2. As explained in Sect. 2.3, the Z (cid:48) or H may decay to only one boson that yields a visible,displaced f ¯ f vertex. One should therefore publish inclusive measurements of σ [ B → f ¯ f ] , as a function of 3 parameters, which can be taken to be the transverse momentum P t B , mass M B and mean decay length of the boson B .In these two variants (especially the second one), the backgrounds will be larger and harder tocontrol. They may therefore only be practical for leptonic final states.A more complex analysis (but perhaps possible in a well understood detector) is to searchfor a pair of long-lived fermions, each decaying to a 3-fermion final state. However, to reducethe number of free parameters, the early benchmarks described here consider only the specialcase, where this occurs via an intermediate 2-body state (fermion plus boson, where the bosonthen decays promptly to f ¯ f (cid:48) ). One suitable benchmark is the model [34], which predicts Z (cid:48) → ν H ¯ ν H , where ν H is a long-lived, heavy neutrino, followed by ν H → (cid:96) − W + or νZ .The case where both ν H decay to νZ is simplest. One can begin by using this channelto make a measurement of the cross-section σ [ ν H ¯ ν H → ( νZ )(¯ νZ )] , as a function of M ν H ¯ ν H , M ν H and the mean decay length of ν H . In this initial measurement, one should explicitly re-construct the displaced Z , but not attempt to reconstruct the ν H or Z (cid:48) . This ensures that theresults obtained will also constrain other models predicting displaced Z . (e.g. Long-lived 4 th generation quarks b (cid:48) → bZ [495]).The case where both ν H decay to (cid:96) − W + is expected to have a larger cross-section, sois also worth pursuing. The mixed channel, where one W decays hadronically and the otherleptonically is perhaps most promising. Measurements of σ [ Z (cid:48) → ν H ¯ ν H → ( (cid:96) − W + )( (cid:96) + W − )] could be quoted as function of M Z (cid:48) , M ν H and the mean decay length of ν H . (Here one wouldfully reconstruct the ν H and Z (cid:48) ). Ultimately, this benchmark could be further generalized toconsider long-lived fermions decaying to a quark/lepton plus an unknown boson (charged orneutral) which in turn decays promptly to f ¯ f (cid:48) . ˜ χ → ˜ DB In this model, ˜ D is a stable, invisible fermion, assumed to be of negligible mass, (e.g. a grav-itino). The neutral boson B can be a photon (already covered in standard GMSB searches andignored here), a Z , a Higgs (with a new parameter, the Higgs mass), or a new exotic boson (withan unknown mass). This boson is assumed to decay to difermions f ¯ f .The underlying SUSY event can be based on a standard SUSY benchmark [4], in whichthe ˜ χ is the LSP. However, it must be understood that the search strategy, and the parameterslimited by the analysis, will depend strongly upon the particular SUSY benchmark point chosen.A key feature of these events is that, in addition to the fermions from the long-livedexotic, the rest of the event will contain other hard particles from the SUSY decay chain. Onecan use these additional particles for triggering or background rejection. Doing so allows oneto explore regions of parameter space which would otherwise be inaccessible. (e.g. Wherethe fermions from the long-lived exotic are too soft to be triggered upon). However, relyingon these additional particles does make the results very dependent on the particular choice ofSUSY benchmark. It is therefore strongly advisable to also quote results for the case wherethese extra particles have not been used. 165here are two simple benchmarks based on this model:1. The χ decays promptly and the boson B is long-lived. This leads to an inclusive cross-section measurement of σ [ BB → ( f ¯ f )( f (cid:48) ¯ f (cid:48) )] , as a function of the B boson’s mass andmean decay length. (N.B. The results obtained will also be influenced by the P t of the bo-son B , which depends on the particular SUSY benchmark chosen). One would probablynot need to reconstruct the ˜ χ , except to establish the exact nature of a discovery.Technically, this benchmark is very similar to the first benchmark described in Sect. 3.1.1.Indeed, unless one is using the rest of the SUSY event for triggering/background rejec-tion, they are almost identical, and one can query if it is worth studying both. Doing sodoes, however, allow one to check if the more crowded environment of the SUSY eventaffects the signal selection efficiency.2. The ˜ χ is long-lived and the boson B decays promptly. In this case, one can measurethe inclusive cross-section σ [ ˜ χ ˜ χ → ( ˜ DB )( ˜ DB ) → ( ˜ Df ¯ f )( ˜ Df (cid:48) ¯ f (cid:48) )] as a function ofthe B boson’s mass and the ˜ χ ’s mean decay length. (The results will also depend onthe mass and P t of the ˜ χ , which depend on the choice of SUSY benchmark). Thisis the more experimentally challenging of the two variants, because one can no longersuppress background by assuming that the momentum vector of the displaced f ¯ f systemis collinear with direction from the beam spot to the displaced vertex. (However, onecould limit oneself to the special case when B = Z , which allows one to use the Z mass constraint to suppress background, and makes the cross-section dependent on oneless free parameter).Technically, this benchmark is very similar to that based on ν H → νZ . described inSect. 3.1.1. So again, it may not be necessary to study both. As was the case for thatbenchmark, more generally applicable results could be obtained by only explicitely re-constructing the boson B , and not attempting to find the χ . In all the above benchmarks, one can consider the possibility of the long-lived exotic beingcharged. The analysis remains similar, although the final states will be subtly different (e.g. l − ¯ ν instead of l − l + ). It may be worth specifically searching not only for the tracks produced bythe daughters of the exotic, but also for the highly ionizing track produced by the exotic itself.These should all form a common vertex. (In the case a signal is seen, this will help clarify itsexact nature, whereas if no signal is seen, it may suppress background). This technique may beparticularly helpful for signatures such as the GMSB SUSY ˜ τ − → τ − ˜ G , where ˜ τ − is long-livedand the gravitino ˜ G is invisible. The decay point of the ˜ τ − can be found by searching for thevertex of the ˜ τ − and τ − tracks.For colored exotics decaying in flight, there are subtle issues. (An example of such anexotic is the long-lived gluino, predicted in ‘Split SUSY’ theories, where the gluino is muchlighter than the squarks [475]). These exotics may form neutral or charged exotic hadrons,which may or may not have a track pointing to the decay vertex. Moreover, there is someprobability of a nearby pion being formed in the hadronization process (even though gluonradiation is suppressed for a massive particle) and this could give a nearby soft track that couldimpact vertex-isolation requirements in an analysis. On the other hand, colored particles aregenerally most often pair-produced (rather than produced in a decay of a heavier particle) and166herefore they travel at a variety of speeds. Some fraction of them might therefore be detectablethrough the late arrival of their decay products at the ECAL. A special case occurs if the long-lived exotic is very light (less than or of order 1 GeV/ c ).In this low mass case, kinematics mean that it can only decay to pairs of leptons (or pairs ofhadrons). These will often have a very small opening angle, which can make them difficultto resolve experimentally (and prevent them being selected by isolated lepton triggers) [486].Also, they can appear in clusters, further complicating isolation requirements.As explained in Sect. 2.3, there is theoretical motivation for this scenario, in which thelong-lived particles are known as “dark photons”. This makes it well worth pursuing. Sincethese decay to normal matter only as a result of mixing with the photon, their decay branchingratios are determined by electromagnetic couplings. After measuring the cross-section of theseparticles to decay to individual fermion species, one can therefore subsequently use this theo-retical knowledge to combine the separate channels and/or obtain a measurement of the darkphoton production cross-section. Trigger/reconstruction benchmarks
A serious concern that has arisen in the study of models with new long-lived particles involvesthe behavior of the triggering and reconstruction software of the LHC (and Tevatron) experi-ments for such events. It is now rather well-appreciated that triggering can be a serious chal-lenge in certain events with long-lived particles. Studies of various examples are on-going andthere are some public results [491]. However, only preliminary studies have been performed ofthe risks at the step following triggering: reconstruction and data storage. The primary recon-struction software, designed for obvious reasons to look for jets, leptons and photons emerg-ing from or near the interaction point, may behave unpredictably when faced with long-liveddecays, especially in cases that have not already been actively studied in the context of gauge-mediated supersymmetry breaking with photons [472]. While it is certainly necessary, whenlooking for long-lived particles, to run special reconstruction algorithms, it is often not feasibleto run these algorithms on the full data set. It is therefore important that data set be reducedthrough some initial event selection to a manageable size. (This is particularly true for finalstates involving only jets, because of the large QCD background). If the primary reconstructionprocess in some way fails to identify events with long-lived particles as candidates for special-ized reconstruction, it might turn out to be impossible to collect a large fraction of the signalinto an analysis sample. Conversely, if the primary reconstruction software can recognize andflag events with unusual features that merit inclusion in a long-lived particle analysis, this maysignificantly increase the fraction of signal that can be collected.After some discussion of this issue, it was generally agreed that a stress-test of the triggersystem and reconstruction software of the experiments is warranted. Toward this end, a numberof simulated data sets, from a variety of models with long-lived particles, has been assembled.Some of the models produce simple signatures, while others produce relatively extreme (thoughrealistic) signals that, though not especially probable, are appropriate for testing the behavior167f the reconstruction software. It must be emphasized that for this reason these more complexmodels should not be viewed as proper benchmark models for early LHC analyses, and theymay turn out not to be good benchmark models for later analyses either. Moreover, for somemodels, the simulation techniques employed in the event generator are crude. Any seriousexperimental analysis would deserve more carefully constructed event generators (which formost of these models are under construction or consideration.) However, in all models theevents themselves are consistent with the underlying physical process — only the statisticaldistribution of the events over phase space is not entirely correct. For this reason, the limitationsjust described should not much affect the realism of individual events, and so any problemsobserved in the reconstruction of these events should still allow important lessons regarding thebehavior of the software to be drawn.Though relatively simple in their signatures, the simple models discussed in Sect. 3 asanalysis benchmarks are also appropriate for the trigger/reconstruction stress-test. It is alreadyknown that certain models with mainly low-energy jets produced by long-lived particles ( e.g. H → B ¯ B followed by displaced B → b ¯ b ) can cause trouble for the ATLAS trigger [491]. Evenwhen triggered, B ¯ B final states with decay lengths of order or greater than ∼
20 cm may lackhits in the pixel detectors, which can lead to a failure of track reconstruction. For B → dijets,a partially successful reconstruction may be insufficient, because of large QCD backgroundsfrom secondary interactions.Special problems arise when B is very light (and potentially also for larger masses if B issufficiently boosted.) Even when prompt, very light ( ∼ e.g. darkphotons) with a boost factor (cid:29) can cause various problems for triggering and reconstruction,including but not limited to a failure of isolation requirements (with each lepton ruining theisolation of the other) or because only one of the two close leptons is identified. A long lifetimefor the new particle compounds these problems.While there have been some public trigger studies at ATLAS [491], the question ofwhether these relatively simple signatures cause problems for reconstruction at the LHC ex-periments has so far been only subject to preliminary studies. Any issues that arise in thesesimpler settings will need to be addressed before there is any hope of understanding the situa-tion for more complex signatures, which we now discuss.Complex signatures easily arise once the multiplicity of long-lived particles exceeds oneor two. The distribution of decay vertices and their daughter particles around the detector canbe enormously variable and complicated. Many classes of Hidden Valley models can producehigh-multiplicity final states and/or long-lived particles in some regions of parameter space.Consequently, Hidden Valleys serve as a useful set from which to select examples of physicallyrealistic phenomena that might be especially challenging for reconstruction software.As emphasized in [476], high-multiplicity states may result through a number of mecha-nisms, including cascade decays within the hidden sector, parton showering within the hiddensector, and/or hidden sector hadronization. For those hidden sector particles which are for-bidden by kinematics and/or quantum numbers from decaying to final states of purely hiddensector particles, their decays to standard model particles are relatively slow, and their lifetimesrelatively long. This arises because of the weak couplings (through small mixings or irrelevantinteractions) of the hidden sector to the standard model sector. As for hadrons in QCD, theirlifetimes may be further enhanced by approximately conserved quantum numbers (analogousto strangeness or CP). Furthermore, a given hidden sector often has multiple metastable hidden-sector particles with relatively long lifetimes. There is typically, therefore, a wide range of168arameters over which a hidden valley model will produce at least one type of particle that willgenerally decay with an observably displaced vertex, often well within the detector volume.The models chosen for the reconstruction stress-test draw upon the same three produc-tion mechanisms for v-particles described earlier: resonance decay, LSP (or similar) decay,and quirk annihilation. (Other mechanisms certainly may arise but these three suffice to givethe wide variety of kinematic distributions needed for the stress-test.) Holding the produc-tion mechanism fixed, and within a given class of models, experience has shown that one mayusually vary the v-particle decay chains or showering rates, masses, and lifetimes as almostindependent parameters, , subject to relatively weak constraints from existing data. These phe-nomenological parameters may be adjusted so as to create unusual but nevertheless realistic andplausible final states which are qualitatively unlike those for which reconstruction software wasdesigned.For any hidden valley model, key aspects of its signatures are determined by the quantumnumbers of any metastable v-particles which produce standard model particles in its decays.Since v-particles are always neutral their masses are not directly constrained by experiment,so they may be very light. On the other hand, their neutrality limits their possible final statesto manageable sets. As discussed earlier, there are three types of two-body resonances thatcommonly arise, assuming the hidden sector does not strongly violate standard model flavorsymmetries: (1) scalar or pseudoscalar v-bosons that decay to the heaviest fermions available(or to three pions, etc., if sufficiently light); (2) vector v-bosons that decay in a generation-democratic way to fermion pairs (or to light lepton and meson pairs, if sufficiently light) witha “dark photon” (a particle coupling to the standard model only through kinetic mixing withthe photon) as a special case; (3) spin-0 and spin-2 v-bosons that couple to photon pairs, gluonpairs, and (if kinematically allowed) weak boson pairs. Also commonly arising are v-particles(of any spin) that decay to other v-particles via single emission of a photon, Z or Higgs boson,or via emission of a fermion pair. Each one of these types of particles arises in at least onemodel used in the stress test, and in some models several such particles may be present.Obviously the lifetimes of the v-particles play a key role in determining the signature. Inmodels with more than one stable type of v-particle, the lifetimes of these particles may varywidely, potentially leading to prompt decays, highly-displaced decays, and/or missing energyin the same event. At least one model of this class arises below. For many of the models givenbelow, two variants are presented with different lifetimes for the long-lived particle(s).A final key determinant of the final states is the multiplicity and clustering of the v-particles. This is highly model-dependent and depends crucially on the details of how particleproduction in the hidden sector proceeds. A general example of a signature with high mul-tiplicity and complex clustering was discussed in [476], where the visibly decaying particlesmight decay to dijets [496] or to a mixture of jets and leptons [497]. Another is the dark-matter-motivated example of a “lepton jet”, where we mean in this context a jet made from morethan one very light particles which decay (with a branching fraction that is substantial thoughpossibly not unity) to lepton pairs. This is not true of the most minimal hidden valley models, where certain relations between these quantitiesand overall cross-sections often hold. For early-LHC analyses , these more minimal models are more suitable asbenchmarks. Here, for reconstruction stress-tests , a more complex signature, even if it only arises from a non-minimal hidden valley model, is sometimes more appropriate. This is simply by definition; charged or colored particles that couple to the hidden sector are not v-particles butrather “communicators” or “mediators” between the two sectors, and are constrained by experiment to be ratherheavy. ∼ strassler/LesHouchesModels is presentedthe list of models appropriate for the trigger and reconstruction stress tests. For each model areprovided1. A Les Houches Accord event file modelname.lhe with at least one thousand events. Theevent samples themselves are in the form of LHE files at parton level; they must still bepiped through a showering Monte Carlo to account for standard-model showering, decaysand hadronization.2. A description of the model (and the simulation technique used) in the file modelname.mdl3. If the model was generated using a standard Monte Carlo, the appropriate run-card com-mands will be given in modelname.run4. If appropriate, an SLHA file that was used in the Monte Carlo generation will be given inmodelname.spc.For some models the full set of particles in intermediate steps is provided in the LHEfiles, but in other cases they are not; this depends on the simulation method used. However inall cases the mother pointers in the LHE file are internally consistent. Some of the particlesin intermediate steps, and certain stable particles, are new and have non-standard PDG codes,though since they are charge- and color-neutral no conflicts or challenges should arise withsimulation. Conclusions ∼ strassler/LesHouchesModels.There are a number of other topics that were discussed at the Les Houches workshop thatwe have not covered here. These include interesting but exotic prompt signatures, includingfour-lepton (non- Z ) decays of the Higgs boson, prompt light dilepton resonances, producedisolated or in clusters; Monte Carlo implementation of non-abelian hidden sectors in HERWIGand in SHERPA; and the simulation of the various phenomena associated with (microscopic)quirks. Acknowledgements
The authors would like to thank the University of Washington LHC group for hosting the work-shop “Signatures of Long-Lived Exotic Particles at the LHC” at which some of this work wasinitiated, and the DOE which supported that workshop under Task TeV of contract DE-FGO2-96-ER40956. The work of MJS also was supported under grants DOE–DE-FG02-96ER40959and NSF–PHY0904069. 170 ontribution 17A benchmark SUSY Abelian hidden sector
D.E. Morrissey, D. Poland and K.M. Zurek
Abstract
New and unusual collider signatures can arise if the MSSM couplesto a light hidden sector through gauge kinetic mixing. In this note wedescribe the minimal supersymmetric realization of this scenario. Thismodel provides a simple benchmark for future LHC collider studies oflight hidden sectors. Introduction
Supersymmetry is a well-motivated candidate for new physics beyond the Standard Model (SM),and the collider and cosmology signals of the minimal supersymmetric Standard Model (MSSM)have been studied very extensively [162, 163]. However, many new possibilities can arise if thefield content of the MSSM is expanded. One interesting extension consists of the MSSM cou-pled to a new gauged hidden sector with characteristic mass scale near a GeV [483]. Models ofthis type have received attention recently in relation to potential hints of dark matter, but theyare also worthy of study in their own right since they can produce new and unusual signals atparticle colliders [483].The largest effect of such a light hidden sector on the phenomenology at the LHC comesfrom the fact that, even with exact R -parity, the lightest MSSM superpartner will no longerbe stable. Instead, the LSP of the full theory will lie in the hidden sector. Supersymmetriccascades initiated by QCD-charged MSSM states will therefore terminate at the lightest MSSMsuperpartner, which will subsequently decay into the hidden sector. These decays, or subsequentcascades within the hidden sector itself, can potentially generate highly boosted leptons or jets.In these proceedings we describe the minimal MSSM extension containing a gauged lighthidden sector, consisting of a Higgsed Abelian U (1) x gauge group that couples to the MSSMthrough gauge kinetic mixing. This simple extension can display a wide variety of collidersignals, and could potentially serve of as benchmark example for collider studies of light hiddensectors. Model, parameters, and spectrum
The extension of the MSSM that we consider consists of a new supersymmetric U (1) x gaugemultiplet ( X µ , ˜ X ) , together with a pair of hidden Higgs fields H and H (cid:48) with charges x = ± under U (1) x , but singlets under the Standard Model gauge groups. We take their superpotentialto be [498] W x = − µ (cid:48) H H (cid:48) . (1)171e also assume soft supersymmetry-breaking couplings of the form −L ⊃ m H | H | + m H (cid:48) | H (cid:48) | + (cid:20) − ( Bµ ) (cid:48) HH (cid:48) + 12 M x ˜ X ˜ X + h.c. (cid:21) . (2)By redefining these fields, we can take ( Bµ ) (cid:48) and M x to be real and positive with no loss ofgenerality. For reasons we will discuss below, all dimensionful terms in the hidden sector areassumed to be on the order of a GeV.To connect this hidden sector to the MSSM, we introduce a kinetic mixing coupling be-tween U (1) x and hypercharge. At the supersymmetric level, this is given by L ⊃ (cid:90) d θ (cid:18) B α B α + 14 X α X α + (cid:15) B α X α (cid:19) + h.c. (3) ⊃ − B µν B µν − X µν X µν − (cid:15) X µν B µν + i(cid:15) ˜ X † ¯ σ µ ∂ µ ˜ B + . . . Such a term will be generated radiatively when there are fields charged under both U (1) x and U (1) Y [499–501], ∆ (cid:15) ( µ ) (cid:39) g x ( µ ) g Y ( µ )16 π (cid:88) i x i Y i ln Λ µ , (4)where x i and Y i denote the charges of the i -th field, Λ is the UV cutoff scale, and the log is cutoff below µ (cid:39) m i , where m i is the (supersymmetric) mass of the i -th field. This leads to valuesof the kinetic mixing in the typical range (cid:15) (cid:39) − − − . Conversely, the kinetic mixingparameter (cid:15) can be highly suppressed or absent if there exist no such bi-fundamentals, or if theunderlying gauge structure consists of a simple group.In the present work we concentrate on the case where all the dimensionful couplings in thehidden U (1) x sector are on the order of a GeV, but with the soft supersymmetry breaking (and µ ) terms in the MSSM larger than a few hundred GeV. This can arise naturally if the mediatorof supersymmetry breaking couples more strongly to the MSSM than to the hidden sector.The canonical example we have in mind is gauge mediation in which the gauge messengersare charged under the Standard Model gauge groups but not U (1) x . Supersymmetry breakingin the MSSM sector will then be communicated to the hidden sector through gauge kineticmixing with characteristic size m ( x ) soft ≤ (cid:15) m ( MSSM ) soft [502]. We additionally assume that thereare supergravity contributions to supersymmetry-breaking parameters in all sectors of order m / ∼ GeV. These contributions will be subleading in the MSSM sector (and sufficientlysmall so as to avoid flavor mixing problems), but can be very important in the hidden sector.Supergravity contributions of this size also provide a natural origin for µ (cid:48) ∼ GeV through theGiudice-Masiero mechanism [503].Supersymmetry breaking in the MSSM will also induce an effectively supersymmetriccontribution to the hidden-sector potential. A non-vanishing hypercharge D -term arises in theMSSM when the visible-sector Higgs fields acquire VEVs with tan β (cid:54) = 1 (induced by SUSYbreaking) ξ Y = − g Y c β v . (5)The kinetic mixing operator of Eq. (3) then leads to an effective Fayet-Iliopoulos [504] term inthe hidden-sector D -term potential [487, 501, 505, 506] V D = g x c (cid:15) (cid:18) | H | − | H (cid:48) | − (cid:15)g x ξ Y (cid:19) , (6)172ith c (cid:15) = 1 / √ − (cid:15) . The hypercharge D -term potential retains its usual form.Putting together all the contributions, the hidden sector scalar potential can be written as V = ( | µ (cid:48) | + ˜ m H ) | H | + ( | µ | + ˜ m H (cid:48) ) | H (cid:48) | − [( Bµ ) (cid:48) HH (cid:48) + ( h.c. )] + g x (cid:0) | H | − | H (cid:48) | (cid:1) , (7)where we have defined ˜ m H = m H − (cid:15) g x ξ Y , ˜ m H (cid:48) = m H (cid:48) + (cid:15) g x ξ Y , (8)and we have dropped terms of O ( (cid:15) ) . This potential is structurally identical to the Higgs po-tential in the MSSM. Since we can take ( Bµ ) (cid:48) to be real and positive (after a suitable fieldredefinition), this potential is minimized with real and non-negative vacuum expectation val-ues (VEVs). It is convenient to write them as (cid:104) H (cid:105) = η sin α, (cid:104) H (cid:48) (cid:105) = η cos α, (9)with sin α, cos α ≥ .Extremizing the potential, we find sin 2 α = 2( Bµ ) (cid:48) | µ (cid:48) | + ˜ m H + ˜ m H (cid:48) (10) η = − | µ (cid:48) | g x + − ˜ m H tan α + ˜ m H (cid:48) g x (tan α − . (11)This solution defines a consistent local minimum provided sin 2 α ≤ and η ≥ .For non-zero η , U (1) x is broken and we obtain a massive gauge boson Z x with mass m Z x = √ g x η. (12)This state will mix with the photon and the Z due to the kinetic coupling, but the effect on themass eigenvalues are minuscule for (cid:15) (cid:28) . Symmetry breaking leads to two physical CP-evenscalars and one physical CP-odd scalar. The tree-level mass of the CP-odd state A x is m A x = 2( Bµ ) (cid:48) sin 2 α = 2 | µ (cid:48) | + ˜ m H + ˜ m H (cid:48) . (13)For the two CP-even states, h x and H x , the mass matrix is given by M h x = (cid:18) m Z x s α + m A x c α − ( m Z x + m A x ) s α c α − ( m Z x + m A x ) s α c α m Z x c α + m A x s α (cid:19) . (14)The remaining states in the hidden sector consist of fermions from the U (1) x neutralino (xino)and the hidden higgsinos. Their mass matrix in the basis ( ˜ X, ˜ H, ˜ H (cid:48) ) is M = M x m Z x s α − m Z x c α m Z x s α − µ (cid:48) − m Z x c α − µ (cid:48) . (15)Note that this mass matrix acquires a zero eigenvalue in the limit that M x → and sin α → .When the only source of supersymmetry breaking in the hidden sector is residual gauge me-diation through gauge kinetic mixing, one naturally obtains M x , ( Bµ ) (cid:48) (cid:28) m Z x and is pushed173o this limit. The resulting light neutralino state can be highly problematic cosmologically andcould potentially overclose the universe. It is for this reason we include additional supergravitycontributions to the soft parameters from the beginning.The masses in this sector share a strong structural similarity with the MSSM. In particular,we see that the lightest CP-even hidden Higgs mass is bounded from above at tree level by m h x ≤ m Z x cos α. (16)This has important implications for the decay properties of the lighter Higgs state since thereneed not be any light hidden fermions in the spectrum. In particular, the decay of h x can be veryslow if the only channel available to it involves a loop or a pair of highly off-shell Z x gaugebosons.For the fermions, the mass matrix has a similar form to the MSSM, but without a winostate. Only the hidden higgsinos couple to gauge bosons. In the absence of large mixing be-tween the xino and the higgsinos, the relative mass gap between two mostly higgsino states willbe less than the Z x mass. Thus, for M x (cid:28) µ (cid:48) or M x (cid:29) µ (cid:48) , the heaviest neutralino will decayto the lighter state(s) primarily by emitting a hidden Higgs. In the latter case, the decay fromthe slightly heavier mostly higgsino state to the lighter one will have to involve an off-shell Z x .Only when there is a great deal of mixing between the xino and higgsinos will neutralino decaysthrough on-shell gauge bosons be relevant [507].So far we have not considered the effects of gauge kinetic mixing on the spectrum. Toleading order in (cid:15) , the gauge boson kinetic mixing can be removed by shifting the photon fieldaccording to [487] A µ → A µ − (cid:15)X µ . (17)This shift induces a coupling between MSSM fields carrying electromagnetic charge and the X µ gauge boson through the operator − L ⊃ − (cid:15) X µ J µem . (18)This coupling allows the decay Z x → f ¯ f , where f is a light Standard Model fermion withwidth on the order of Γ x ∼ (cid:15) g x m x / π [508]. Despite its small width, these decay channelswill dominate if there are no channels open kinematically in the hidden sector. A small mixingwith the Standard Model Z is also induced [487].Among the neutralinos, the kinetic mixing between the bino and the xino can be removedmost conveniently by shifting the xino according to ˜ X → ˜ X − (cid:15) ˜ B. (19)This shift induces a very small mass mixing between the hidden and MSSM neutralinos on theorder of (cid:15) m Z x /M . More importantly, it leads to a coupling of the Bino to the hidden sector, − L ⊃ − (cid:15) √ g x (cid:16) H ˜ H ˜ B − H (cid:48) ˜ H ˜ B + h.c. (cid:17) . (20)On account of these couplings, a would-be MSSM neutralino LSP will decay to the hiddensector according to ˜ χ → ˜ χ xi h x , ˜ χ xi H x , ˜ χ xi A x . These decays will typically be prompt for (cid:15) > − . In the case of a non-neutralino LSP, this coupling will also permit its decay, albeit at alower rate. 174or the purposes of defining a benchmark model, let us point out that the phenomenologyof the model (in the hidden sector) can be specified by the seven parameters: { g x , (cid:15), m Z x , m A x , tan α, µ (cid:48) , M x } . (21)All these parameters are defined at the low (GeV) scale. Note that we have implicitly used theminimization conditions to eliminate ˜ m H , ˜ m H (cid:48) , and ( Bµ ) (cid:48) in favor of m Z x , m A x , and tan α which have a more intuitive interpretation. We could also have included a soft mass mixingbetween the bino and xino, but it is consistent to neglect such a term provided the origin of (cid:15) issupersymmetric. Constraints and signatures
A light hidden sector of this form can produce striking signals at the LHC [483]. These originateprimarily from supersymmetric MSSM cascade decays. Instead of terminating at the lightestMSSM superpartner, these cascades will continue into the hidden sector. The cascade willcontinue in the hidden sector until the lightest superpartner is reached. If these produce oneor more Z x gauge bosons or h x hidden Higgses, a distinctive signature will arise from decaysof the Z x or h x to Standard Model leptons. These leptons will typically have a small invariantmass on account of their origin, but they will also be highly collimated since their energy scaleis set by the mass of the much heavier MSSM states [487, 488]. This construction is a simpleand perturbative example of a hidden valley scenario [476, 477].Before discussing the collider signatures of this scenario, let us first outline the boundson a light hidden sector with a kinetic mixing to hypercharge. The most stringent bounds comefrom the induced Z x coupling to SM fermions. For m Z x ∼ GeV, and with Z x → f ¯ f thedominant decay mode, limits from ( g − µ and B -physics force (cid:15) < − or so [482, 509].Additional constraints can arise if h x is long-lived [508, 510–512]. This arises naturally if m h x < m Z x and there are no lighter x -sector fermions, in which case it can only decay througha loop with Z x ’s or through an off-shell pair of Z x ’s to the SM [508]. Constraints on both the Z x and h x states are weakened considerably if they decay primarily to the invisible hidden-sectorLSP. A wide range of collider signals can arise from the minimal hidden sector model dis-cussed here. The dominant mode of production at the LHC is expected to be through MSSMcascade decays. These cascades will proceed to the MSSM LSP, which we assume here to bethe lightest MSSM neutralino. This LSP will subsequently decay to the hidden sector throughthe interaction of Eq. (20), and the decay cascade will continue in the hidden sector down to theLSP. Along the way, h x , H x , A x and Z x states will be emitted, which can potentially generatenew signals.Consider first the situation for min { µ (cid:48) , M x } < m h x / . In this case the Z x and h x bosonsemitted in the decay cascades will almost always decay to pairs of the lightest neutralino LSP.The LHC collider signatures are then nearly identical to those of the MSSM in the absence ofthe hidden sector: even though the decay cascades continue, they still only produce missingenergy.A more interesting case is m A x (cid:29) µ (cid:48) ∼ M x > m Z x such that all the hidden neutralinosare well-mixed with m χ x > m Z x and m χ x − m χ x > m Z x . This spectrum allows the decays H x → Z x Z x along with χ x → χ x Z x , while the only decay channel open to the gauge bosonis back to the Standard Model, Z x → f ¯ f . In this situation, some of the decay cascades will175roduce new visible signatures from the Z x decays to highly boosted Standard Model states.For m Z x < MeV, these decay products will frequently form lepton jets consisting of highlycollimated lepton pairs [483, 487, 488]. Note that in this case, the h x state is typically verylong-lived and escapes the detector [508]. Thus, some of these can also include missing energy.We are currently working on implementing this simple benchmark model into togetherwith a spectrum generator to generate the necessary SLHA format input data [507]. Conclusions
Light hidden sectors can lead to exciting new signals at the LHC. We have presented here asimple benchmark Abelian hidden in the context of supersymmetry that could be useful forfuture collider studies.
Acknowledgements
We thank the organizers and participants of the “Physics at Tev Colliders” program for anexcellent workshop, as well as the École de Physique des Houches for the use of their facilitiesand the delicious cheese. 176 ibliography http://lepsusy.web.cern.ch/lepsusy/
Quarks and Leptons , eds. M.Lèvy et al. (Plenum, New York ),p.707