866 resultados para LHC Ê


Relevância:

20.00% 20.00%

Publicador:

Resumo:

A search has been performed, using the full 20.3  fb −1 data sample of 8 TeV proton-proton collisions collected in 2012 with the ATLAS detector at the LHC, for photons originating from a displaced vertex due to the decay of a neutral long-lived particle into a photon and an invisible particle. The analysis investigates the diphoton plus missing transverse momentum final state, and is therefore most sensitive to pair production of long-lived particles. The analysis technique exploits the capabilities of the ATLAS electromagnetic calorimeter to make precise measurements of the flight direction, as well as the time of flight, of photons. No excess is observed over the Standard Model predictions for background. Exclusion limits are set within the context of gauge mediated supersymmetry breaking models, with the lightest neutralino being the next-to-lightest supersymmetric particle and decaying into a photon and gravitino with a lifetime in the range from 250 ps to about 100 ns.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents the electron and photon energy calibration achieved with the ATLAS detector using about 25 fb−1 of LHC proton–proton collision data taken at centre-of-mass energies of √s = 7 and 8 TeV. The reconstruction of electron and photon energies is optimised using multivariate algorithms. The response of the calorimeter layers is equalised in data and simulation, and the longitudinal profile of the electromagnetic showers is exploited to estimate the passive material in front of the calorimeter and reoptimise the detector simulation. After all corrections, the Z resonance is used to set the absolute energy scale. For electrons from Z decays, the achieved calibration is typically accurate to 0.05% in most of the detector acceptance, rising to 0.2% in regions with large amounts of passive material. The remaining inaccuracy is less than 0.2–1% for electrons with a transverse energy of 10 GeV, and is on average 0.3% for photons. The detector resolution is determined with a relative inaccuracy of less than 10% for electrons and photons up to 60 GeV transverse energy, rising to 40% for transverse energies above 500 GeV.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents the performance of the ATLAS muon reconstruction during the LHC run with pp collisions at √s = 7–8 TeV in 2011–2012, focusing mainly on data collected in 2012. Measurements of the reconstruction efficiency and of the momentum scale and resolution, based on large reference samples of J/ψ → μμ, Z → μμ and ϒ → μμ decays, are presented and compared to Monte Carlo simulations. Corrections to the simulation, to be used in physics analysis, are provided. Over most of the covered phase space (muon |η| < 2.7 and 5 ≲ pT ≲ 100 GeV) the efficiency is above 99% and is measured with per-mille precision. The momentum resolution ranges from 1.7% at central rapidity and for transverse momentum pT ≅ 10 GeV, to 4% at large rapidity and pT ≅ 100 GeV. The momentum scale is known with an uncertainty of 0.05% to 0.2% depending on rapidity. A method for the recovery of final state radiation from the muons is also presented.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Many of the interesting physics processes to be measured at the LHC have a signature involving one or more isolated electrons. The electron reconstruction and identification efficiencies of the ATLAS detector at the LHC have been evaluated using proton–proton collision data collected in 2011 at √s = 7 TeV and corresponding to an integrated luminosity of 4.7 fb−1. Tag-and-probe methods using events with leptonic decays of W and Z bosons and J/ψ mesons are employed to benchmark these performance parameters. The combination of all measurements results in identification efficiencies determined with an accuracy at the few per mil level for electron transverse energy greater than 30 GeV.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Very recently, the ATLAS and CMS Collaborations reported diboson and dijet excesses above standard model expectations in the invariant mass region of 1.8–2.0 TeV. Interpreting the diboson excess of events in a model independent fashion suggests that the vector boson pair production searches are best described by WZ or ZZ topologies, because states decaying into W+W− pairs are strongly constrained by semileptonic searches. Under the assumption of a low string scale, we show that both the diboson and dijet excesses can be steered by an anomalous U(1) field with very small coupling to leptons. The Drell–Yan bounds are then readily avoided because of the leptophobic nature of the massive Z′ gauge boson. The non-negligible decay into ZZ required to accommodate the data is a characteristic footprint of intersecting D-brane models, wherein the Landau–Yang theorem can be evaded by anomaly-induced operators involving a longitudinal Z. The model presented herein can be viewed purely field-theoretically, although it is particularly well motivated from string theory. Should the excesses become statistically significant at the LHC13, the associated Zγ topology would become a signature consistent only with a stringy origin.

Relevância:

20.00% 20.00%

Publicador:

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Heavy-ion collisions are a powerful tool to study hot and dense QCD matter, the so-called Quark Gluon Plasma (QGP). Since heavy quarks (charm and beauty) are dominantly produced in the early stages of the collision, they experience the complete evolution of the system. Measurements of electrons from heavy-flavour hadron decay is one possible way to study the interaction of these particles with the QGP. With ALICE at LHC, electrons can be identified with high efficiency and purity. A strong suppression of heavy-flavour decay electrons has been observed at high $p_{m T}$ in Pb-Pb collisions at 2.76 TeV. Measurements in p-Pb collisions are crucial to understand cold nuclear matter effects on heavy-flavour production in heavy-ion collisions. The spectrum of electrons from the decays of hadrons containing charm and beauty was measured in p-Pb collisions at $\\sqrt = 5.02$ TeV. The heavy flavour decay electrons were measured by using the Time Projection Chamber (TPC) and the Electromagnetic Calorimeter (EMCal) detectors from ALICE in the transverse-momentum range $2 < p_ < 20$ GeV/c. The measurements were done in two different data set: minimum bias collisions and data using the EMCal trigger. The non-heavy flavour electron background was removed using an invariant mass method. The results are compatible with one ($R_ \\approx$ 1) and the cold nuclear matter effects in p-Pb collisions are small for the electrons from heavy-flavour hadron decays.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

L’obiettivo di tutto il mio lavoro è stato quello di misurare le sezioni d’urto di produzione dei bosoni deboli W ± e Z nei loro decadimenti leptonici (e, μ) coi dati raccolti dal rivelatore ATLAS a LHC con un’energia del centro di massa di √s = 13 TeV relativi all’estate 2015. Gli eventi selezionati sono gli stessi di quelli del recente articolo della Collaborazione ATLAS sullo stesso argomento, in modo anche da poter operare un confronto tra i risultati ottenuti. Confronto peraltro necessario, poichè i risultati sono stati ottenuti con due metodologie differenti: tradizionale (classica) per l’articolo, bayesiana in questa tesi. L’approccio bayesiano permette di combinare i vari canali e di trattare gli effetti sistematici in modo del tutto naturale. I risultati ottenuti sono in ottimo accordo con le predizioni dello Standard Model e con quelli pubblicati da ATLAS.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Neste trabalho de disserta¸c˜ao, investigamos os efeitos nucleares em processos de produ¸c˜ao de quarkonium no Relativistic Heavy Ion Collider (RHIC) e no Large Hadron Collider (LHC). Para tanto, consideramos o Modelo de Evapora¸c˜ao de Cor (CEM), baseado em processos partˆonicos calculados mediante a QCD perturbativa e em intera¸c˜oes n˜ao perturbativas via troca de gl´uons suaves para a forma¸c˜ao do quarkonium. Supress˜ao de quarkonium ´e um dos sinais de forma¸c˜ao do assim chamado Plasma de Quarks e Gl´uons (QGP) em colis˜oes ultrarelativ´ısticas de ´ıons pesados. No entanto, a supress˜ao n˜ao ´e somente causada em colis˜oes n´ucleo-n´ucleo (AA) devido `a forma¸c˜ao do QGP. De fato, a supress˜ao de quarkonium tamb´em foi observada em colis˜oes pr´oton-n´ucleo (pA). A fim de separar os efeitos da mat´eria quente (devidos ao QGP) e fria (efeitos n˜ao devidos ao QGP), pode-se olhar primeiro para colis˜oes pA, onde somente efeitos de mat´eria fria desempenham um papel fundamental, e depois aplicar esses efeitos em colis˜oes AA, uma vez que parte da supress˜ao ´e devido a efeitos de mat´eria fria. No regime de altas energias, a produ¸c˜ao do quarkonium ´e fortemente dependente da distribui¸c˜ao de gl´uons nuclear, o que viabiliza uma oportunidade ´unica de estudar o comportamento de pequeno x dos gl´uons dentro do n´ucleo e, consequentemente, restringir os efeitos nucleares. Estudamos os processos nucleares utilizando distintas parametriza¸c˜oes para as distribui¸c˜oes partˆonicas nucleares. Calculamos a raz˜ao nuclear para processos pA e AA em fun¸c˜ao da vari´avel rapidez para a produ¸c˜ao de quarkonium, o que permite estimar os efeitos nucleares. Al´em disso, apresentamos uma compara¸c˜ao com os dados do RHIC para a produ¸c˜ao do m´eson J/Ψ em colis˜oes pA, demonstrando que a an´alise deste observ´avel ´e uma quest˜ao em aberto na literatura. Adicionalmente, estimamos a produ¸c˜ao de quarks pesados e quarkonium na etapa inicial e durante a fase termal de uma colis˜ao ultrarelativ´ıstica de ´ıons pesados. O objetivo deste estudo ´e estimar as distintas contribui¸c˜oes para a produ¸c˜ao e de alguns efeitos do meio nuclear.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Since it has been found that the MadGraph Monte Carlo generator offers superior flavour-matching capability as compared to Alpgen, the suitability of MadGraph for the generation of ttb¯ ¯b events is explored, with a view to simulating this background in searches for the Standard Model Higgs production and decay process ttH, H ¯ → b ¯b. Comparisons are performed between the output of MadGraph and that of Alpgen, showing that satisfactory agreement in their predictions can be obtained with the appropriate generator settings. A search for the Standard Model Higgs boson, produced in association with the top quark and decaying into a b ¯b pair, using 20.3 fb−1 of 8 TeV collision data collected in 2012 by the ATLAS experiment at CERN’s Large Hadron Collider, is presented. The GlaNtp analysis framework, together with the RooFit package and associated software, are used to obtain an expected 95% confidence-level limit of 4.2 +4.1 −2.0 times the Standard Model expectation, and the corresponding observed limit is found to be 5.9; this is within experimental uncertainty of the published result of the analysis performed by the ATLAS collaboration. A search for a heavy charged Higgs boson of mass mH± in the range 200 ≤ mH± /GeV ≤ 600, where the Higgs mediates the five-flavour beyond-theStandard-Model physics process gb → tH± → ttb, with one top quark decaying leptonically and the other decaying hadronically, is presented, using the 20.3 fb−1 8 TeV ATLAS data set. Upper limits on the product of the production cross-section and the branching ratio of the H± boson are computed for six mass points, and these are found to be compatible within experimental uncertainty with those obtained by the corresponding published ATLAS analysis.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Crossing the Franco-Swiss border, the Large Hadron Collider (LHC), designed to collide 7 TeV proton beams, is the world's largest and most powerful particle accelerator the operation of which was originally intended to commence in 2008. Unfortunately, due to an interconnect discontinuity in one of the main dipole circuit's 13 kA superconducting busbars, a catastrophic quench event occurred during initial magnet training, causing significant physical system damage. Furthermore, investigation into the cause found that such discontinuities were not only present in the circuit in question, but throughout the entire LHC. This prevented further magnet training and ultimately resulted in the maximum sustainable beam energy being limited to approximately half that of the design nominal, 3.5-4 TeV, for the first three years of operation (Run 1, 2009-2012) and a major consolidation campaign being scheduled for the first long shutdown (LS 1, 2012-2014). Throughout Run 1, a series of studies attempted to predict the amount of post-installation training quenches still required to qualify each circuit to nominal-energy current levels. With predictions in excess of 80 quenches (each having a recovery time of 8-12+ hours) just to achieve 6.5 TeV and close to 1000 quenches for 7 TeV, it was decided that for Run 2, all systems be at least qualified for 6.5 TeV operation. However, even with all interconnect discontinuities scheduled to be repaired during LS 1, numerous other concerns regarding circuit stability arose. In particular, observations of an erratic behaviour of magnet bypass diodes and the degradation of other potentially weak busbar sections, as well as observations of seemingly random millisecond spikes in beam losses, known as unidentified falling object (UFO) events, which, if persist at 6.5 TeV, may eventually deposit sufficient energy to quench adjacent magnets. In light of the above, the thesis hypothesis states that, even with the observed issues, the LHC main dipole circuits can safely support and sustain near-nominal proton beam energies of at least 6.5 TeV. Research into minimising the risk of magnet training led to the development and implementation of a new qualification method, capable of providing conclusive evidence that all aspects of all circuits, other than the magnets and their internal joints, can safely withstand a quench event at near-nominal current levels, allowing for magnet training to be carried out both systematically and without risk. This method has become known as the Copper Stabiliser Continuity Measurement (CSCM). Results were a success, with all circuits eventually being subject to a full current decay from 6.5 TeV equivalent current levels, with no measurable damage occurring. Research into UFO events led to the development of a numerical model capable of simulating typical UFO events, reproducing entire Run 1 measured event data sets and extrapolating to 6.5 TeV, predicting the likelihood of UFO-induced magnet quenches. Results provided interesting insights into the involved phenomena as well as confirming the possibility of UFO-induced magnet quenches. The model was also capable of predicting that such events, if left unaccounted for, are likely to be commonplace or not, resulting in significant long-term issues for 6.5+ TeV operation. Addressing the thesis hypothesis, the following written works detail the development and results of all CSCM qualification tests and subsequent magnet training as well as the development and simulation results of both 4 TeV and 6.5 TeV UFO event modelling. The thesis concludes, post-LS 1, with the LHC successfully sustaining 6.5 TeV proton beams, but with UFO events, as predicted, resulting in otherwise uninitiated magnet quenches and being at the forefront of system availability issues.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The scalar sector of the simplest version of the 3-3-1 electroweak model is constructed with three Higgs triplets only. We show that a relation involving two of the constants of the model, two vacuum expectation values of the neutral scalars, and the mass of the doubly charged Higgs boson leads to important information concerning the signals of this scalar particle.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Using a peculiar version of the SU(3)(L) circle times U(1)(N) electroweak model, we investigate the production of doubly charged Higgs boson at the Large Hadron Collider. Our results include branching ratio calculations for the doubly charged Higgs and for one of the neutral scalar bosons of the model. (c) 2006 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Complex singlet extension of the Standard Model (CxSM) is the simplest extension that provides scenarios for Higgs pair production with different masses. The model has two interesting phases: the dark matter phase, with a Standard Model-like Higgs boson, a new scalar and a dark matter candidate; and the broken phase, with all three neutral scalars mixing. In the latter phase Higgs decays into a pair of two different Higgs bosons are possible. In this study we analyse Higgs-to-Higgs decays in the framework of singlet extensions of the Standard Model (SM), with focus on the CxSM. After demonstrating that scenarios with large rates for such chain decays are possible we perform a comparison between the NMSSM and the CxSM. We find that, based on Higgs-to-Higgs decays, the only possibility to distinguish the two models at the LHC run 2 is through final states with two different scalars. This conclusion builds a strong case for searches for final states with two different scalars at the LHC run 2. Finally, we propose a set of benchmark points for the real and complex singlet extensions to be tested at the LHC run 2. They have been chosen such that the discovery prospects of the involved scalars are maximised and they fulfil the dark matter constraints. Furthermore, for some of the points the theory is stable up to high energy scales. For the computation of the decay widths and branching ratios we developed the Fortran code sHDECAY, which is based on the implementation of the real and complex singlet extensions of the SM in HDECAY.