605 resultados para Interacoes hadron-hadron


Relevância:

20.00% 20.00%

Publicador:

Resumo:

ALICE is one of four major experiments of particle accelerator LHC installed in the European laboratory CERN. The management committee of the LHC accelerator has just approved a program update for this experiment. Among the upgrades planned for the coming years of the ALICE experiment is to improve the resolution and tracking efficiency maintaining the excellent particles identification ability, and to increase the read-out event rate to 100 KHz. In order to achieve this, it is necessary to update the Time Projection Chamber detector (TPC) and Muon tracking (MCH) detector modifying the read-out electronics, which is not suitable for this migration. To overcome this limitation the design, fabrication and experimental test of new ASIC named SAMPA has been proposed . This ASIC will support both positive and negative polarities, with 32 channels per chip and continuous data readout with smaller power consumption than the previous versions. This work aims to design, fabrication and experimental test of a readout front-end in 130nm CMOS technology with configurable polarity (positive/negative), peaking time and sensitivity. The new SAMPA ASIC can be used in both chambers (TPC and MCH). The proposed front-end is composed of a Charge Sensitive Amplifier (CSA) and a Semi-Gaussian shaper. In order to obtain an ASIC integrating 32 channels per chip, the design of the proposed front-end requires small area and low power consumption, but at the same time requires low noise. In this sense, a new Noise and PSRR (Power Supply Rejection Ratio) improvement technique for the CSA design without power and area impact is proposed in this work. The analysis and equations of the proposed circuit are presented which were verified by electrical simulations and experimental test of a produced chip with 5 channels of the designed front-end. The measured equivalent noise charge was <550e for 30mV/fC of sensitivity at a input capacitance of 18.5pF. The total core area of the front-end was 2300?m × 150?m, and the measured total power consumption was 9.1mW per channel.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We have studied the radial dependence of the energy deposition of the secondary electron generated by swift proton beams incident with energies T = 50 keV–5 MeV on poly(methylmethacrylate) (PMMA). Two different approaches have been used to model the electronic excitation spectrum of PMMA through its energy loss function (ELF), namely the extended-Drude ELF and the Mermin ELF. The singly differential cross section and the total cross section for ionization, as well as the average energy of the generated secondary electrons, show sizeable differences at T ⩽ 0.1 MeV when evaluated with these two ELF models. In order to know the radial distribution around the proton track of the energy deposited by the cascade of secondary electrons, a simulation has been performed that follows the motion of the electrons through the target taking into account both the inelastic interactions (via electronic ionizations and excitations as well as electron-phonon and electron trapping by polaron creation) and the elastic interactions. The radial distribution of the energy deposited by the secondary electrons around the proton track shows notable differences between the simulations performed with the extended-Drude ELF or the Mermin ELF, being the former more spread out (and, therefore, less peaked) than the latter. The highest intensity and sharpness of the deposited energy distributions takes place for proton beams incident with T ~ 0.1–1 MeV. We have also studied the influence in the radial distribution of deposited energy of using a full energy distribution of secondary electrons generated by proton impact or using a single value (namely, the average value of the distribution); our results show that differences between both simulations become important for proton energies larger than ~0.1 MeV. The results presented in this work have potential applications in materials science, as well as hadron therapy (due to the use of PMMA as a tissue phantom) in order to properly consider the generation of electrons by proton beams and their subsequent transport and energy deposition through the target in nanometric scales.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Measurements of hadron production in p+C interactions at 31 GeV/c are performed using the NA61/SHINE spectrometer at the CERN SPS. The analysis is based on the full set of data collected in 2009 using a graphite target with a thickness of 4% of a nuclear interaction length. Inelastic and production cross sections as well as spectra of π±, K±, p, K0s and Λ are measured with high precision. These measurements are essential for improved calculations of the initial neutrino fluxes in the T2K long-baseline neutrino oscillation experiment in Japan. A comparison of the NA61/SHINE measurements with predictions of several hadroproduction models is presented.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Thesis (Master's)--University of Washington, 2016-06

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Thesis (Ph.D.)--University of Washington, 2016-06

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Most experiments in particle physics are scattering experiments, the analysis of which leads to masses, scattering phases, decay widths and other properties of one or multi-particle systems. Until the advent of Lattice Quantum Chromodynamics (LQCD) it was difficult to compare experimental results on low energy hadron-hadron scattering processes to the predictions of QCD, the current theory of strong interactions. The reason being, at low energies the QCD coupling constant becomes large and the perturbation expansion for scattering; amplitudes does not converge. To overcome this, one puts the theory onto a lattice, imposes a momentum cutoff, and computes the integral numerically. For particle masses, predictions of LQCD agree with experiment, but the area of decay widths is largely unexplored. ^ LQCD provides ab initio access to unusual hadrons like exotic mesons that are predicted to contain real gluonic structure. To study decays of these type resonances the energy spectra of a two-particle decay state in a finite volume of dimension L can be related to the associated scattering phase shift δ(k) at momentum k through exact formulae derived by Lüscher. Because the spectra can be computed using numerical Monte Carlo techniques, the scattering phases can thus be determined using Lüscher's formulae, and the corresponding decay widths can be found by fitting Breit-Wigner functions. ^ Results of such a decay width calculation for an exotic hybrid( h) meson (JPC = 1-+) are presented for the decay channel h → πa 1. This calculation employed Lüscher's formulae and an approximation of LQCD called the quenched approximation. Energy spectra for the h and πa1 systems were extracted using eigenvalues of a correlation matrix, and the corresponding scattering phase shifts were determined for a discrete set of πa1 momenta. Although the number of phase shift data points was sparse, fits to a Breit-Wigner model were made, resulting in a decay width of about 60 MeV. ^

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A Partial Waves Analysis (PWA) of γp → Δ ++X → pπ+ π - (η) data taken with the CLAS detector at Jefferson Lab is presented in this work. This reaction is of interest because the Δ++ restricts the isospin of the possible X states, leaving the PWA with a smaller combination of partial waves, making it ideal to look for exotic mesons. It was proposed by Isgur and Paton that photoproduction is a plausible source for the Jpc=1–+ state through flux tube excitation. The π1(1400) is such a state that has been produced with the use of hadron production but it has yet to be seen in photoproduction. A mass independent amplitude analysis of this channel was performed, followed by a mass dependent fit to extract the resonance parameters. The procedure used an event-based maximum likelihood method to maintain all correlations in the kinematics. The intensity and phase motion is mapped out for the contributing signals without requiring assumptions about the underlying processes. The strength of the PWA is in the analysis of the phase motion, which for resonance behavior is well defined. In the data presented, the ηπ– invariant mass spectrum shows contributions from the a0(980) and a2(1320) partial waves. No π1 was observed under a clear a2 signal after the angular distributions of the decay products were analyzed using an amplitude analysis. In addition, this dissertation discusses trends in the data, along with the implemented techniques.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Al Large Hadron Collider (LHC) ogni anno di acquisizione dati vengono raccolti più di 30 petabyte di dati dalle collisioni. Per processare questi dati è necessario produrre un grande volume di eventi simulati attraverso tecniche Monte Carlo. Inoltre l'analisi fisica richiede accesso giornaliero a formati di dati derivati per centinaia di utenti. La Worldwide LHC Computing GRID (WLCG) è una collaborazione interazionale di scienziati e centri di calcolo che ha affrontato le sfide tecnologiche di LHC, rendendone possibile il programma scientifico. Con il prosieguo dell'acquisizione dati e la recente approvazione di progetti ambiziosi come l'High-Luminosity LHC, si raggiungerà presto il limite delle attuali capacità di calcolo. Una delle chiavi per superare queste sfide nel prossimo decennio, anche alla luce delle ristrettezze economiche dalle varie funding agency nazionali, consiste nell'ottimizzare efficientemente l'uso delle risorse di calcolo a disposizione. Il lavoro mira a sviluppare e valutare strumenti per migliorare la comprensione di come vengono monitorati i dati sia di produzione che di analisi in CMS. Per questa ragione il lavoro è comprensivo di due parti. La prima, per quanto riguarda l'analisi distribuita, consiste nello sviluppo di uno strumento che consenta di analizzare velocemente i log file derivanti dalle sottomissioni di job terminati per consentire all'utente, alla sottomissione successiva, di sfruttare meglio le risorse di calcolo. La seconda parte, che riguarda il monitoring di jobs sia di produzione che di analisi, sfrutta tecnologie nel campo dei Big Data per un servizio di monitoring più efficiente e flessibile. Un aspetto degno di nota di tali miglioramenti è la possibilità di evitare un'elevato livello di aggregazione dei dati già in uno stadio iniziale, nonché di raccogliere dati di monitoring con una granularità elevata che tuttavia consenta riprocessamento successivo e aggregazione “on-demand”.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A measurement of the production cross sections of top quark pairs in association with a W or Z boson is presented. The measurement uses 20.3 fb−1 of data from proton-proton collisions at √s = 8 TeV collected by the ATLAS detector at the Large Hadron Collider. Four different final states are considered: two opposite-sign leptons, two same-sign leptons, three leptons, and four leptons. The t t̅ W and t t̅ Z cross sections are simultaneously extracted using a maximum likelihood fit over all the final states. The t t̅ Z cross section is measured to be 176+58−52 fb, corresponding to a signal significance of 4.2σ. The t t̅ W cross section is measured to be 369+100−91 fb, corresponding to a signal significance of 5.0σ. The results are consistent with next-to-leading-order calculations for the tt̅W and tt̅Z processes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A search for new heavy resonances decaying to boson pairs (WZ, WW or ZZ) using 20.3 inverse femtobarns of proton-proton collision data at a center of mass energy of 8 TeV is presented. The data were recorded by the ATLAS detector at the Large Hadron Collider (LHC) in 2012. The analysis combines several search channels with the leptonic, semi-leptonic and fully hadronic final states. The diboson invariant mass spectrum is studied for local excesses above the Standard Model background prediction, and no significant excess is observed for the combined analysis. 95$\%$ confidence limits are set on the cross section times branching ratios for three signal models: an extended gauge model with a heavy W boson, a bulk Randall-Sundrum model with a spin-2 graviton, and a simplified model with a heavy vector triplet. Among the individual search channels, the fully-hadronic channel is predominantly presented where boson tagging technique and jet substructure cuts are used. Local excesses are found in the dijet mass distribution around 2 TeV, leading to a global significance of 2.5 standard deviations. This deviation from the Standard Model prediction results in many theory explanations, and the possibilities could be further explored using the LHC Run 2 data.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Searches for the supersymmetric partner of the top quark (stop) are motivated by natural supersymmetry, where the stop has to be light to cancel the large radiative corrections to the Higgs boson mass. This thesis presents three different searches for the stop at √s = 8 TeV and √s = 13 TeV using data from the ATLAS experiment at CERN’s Large Hadron Collider. The thesis also includes a study of the primary vertex reconstruction performance in data and simulation at √s = 7 TeV using tt and Z events. All stop searches presented are carried out in final states with a single lepton, four or more jets and large missing transverse energy. A search for direct stop pair production is conducted with 20.3 fb−1 of data at a center-of-mass energy of √s = 8 TeV. Several stop decay scenarios are considered, including those to a top quark and the lightest neutralino and to a bottom quark and the lightest chargino. The sensitivity of the analysis is also studied in the context of various phenomenological MSSM models in which more complex decay scenarios can be present. Two different analyses are carried out at √s = 13 TeV. The first one is a search for both gluino-mediated and direct stop pair production with 3.2 fb−1 of data while the second one is a search for direct stop pair production with 13.2 fb−1 of data in the decay scenario to a bottom quark and the lightest chargino. The results of the analyses show no significant excess over the Standard Model predictions in the observed data. Consequently, exclusion limits are set at 95% CL on the masses of the stop and the lightest neutralino.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

If magnetism is universal in nature, magnetic materials are ubiquitous. A life without magnetism is unthinkable and a day without the influence of a magnetic material is unimaginable. They find innumerable applications in the form of many passive and active devices namely, compass, electric motor, generator, microphone, loud speaker, maglev train, magnetic resonance imaging, data recording and reading, hadron collider etc. The list is endless. Such is the influence of magnetism and magnetic materials in ones day to day life. With the advent of nanoscience and nanotechnology, along with the emergence of new areas/fields such as spintronics, multiferroics and magnetic refrigeration, the importance of magnetism is ever increasing and attracting the attention of researchers worldwide. The search for a fluid which exhibits magnetism has been on for quite some time. However nature has not bestowed us with a magnetic fluid and hence it has been the dream of many researchers to synthesize a magnetic fluid which is thought to revolutionize many applications based on magnetism. The discovery of a magnetic fluid by Jacob Rabinow in the year 1952 paved the way for a new branch of Physics/Engineering which later became magnetic fluids. This gave birth to a new class of material called magnetorheological materials. Magnetorheological materials are considered superior to electrorheological materials in that magnetorheology is a contactless operation and often inexpensive.Most of the studies in the past on magnetorheological materials were based on magnetic fluids. Recently the focus has been on the solid state analogue of magnetic fluids which are called Magnetorheological Elastomers (MREs). The very word magnetorheological elastomer implies that the rheological properties of these materials can be altered by the influence of an external applied magnetic field and this process is reversible. If the application of an external magnetic field modifies the viscosity of a magnetic fluid, the effect of external magnetic stimuli on a magnetorheological elastomer is in the modification of its stiffness. They are reversible too. Magnetorheological materials exhibit variable stiffness and find applications in adaptive structures of aerospace, automotive civil and electrical engineering applications. The major advantage of MRE is that the particles are not able to settle with time and hence there is no need of a vessel to hold it. The possibility of hazardous waste leakage is no more with a solid MRE. Moreover, the particles in a solid MRE will not affect the performance and durability of the equipment. Usually MR solids work only in the pre yield region while MR fluids, typically work in the post yield state. The application of an external magnetic field modifies the stiffness constant, shear modulus and loss modulus which are complex quantities. In viscoelastic materials a part of the input energy is stored and released during each cycle and a part is dissipated as heat. The storage modulus G′ represents the capacity of the material to store energy of deformation, which contribute to material stiffness. The loss modulusG′′ represents the ability of the material to dissipate the energy of deformation. Such materials can find applications in the form of adaptive vibration absorbers (ATVAs), stiffness tunable mounts and variable impedance surfaces. MREs are an important material for automobile giants and became the focus of this research for eventual automatic vibration control, sound isolation, brakes, clutches and suspension systems

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Abstract Heading into the 2020s, Physics and Astronomy are undergoing experimental revolutions that will reshape our picture of the fabric of the Universe. The Large Hadron Collider (LHC), the largest particle physics project in the world, produces 30 petabytes of data annually that need to be sifted through, analysed, and modelled. In astrophysics, the Large Synoptic Survey Telescope (LSST) will be taking a high-resolution image of the full sky every 3 days, leading to data rates of 30 terabytes per night over ten years. These experiments endeavour to answer the question why 96% of the content of the universe currently elude our physical understanding. Both the LHC and LSST share the 5-dimensional nature of their data, with position, energy and time being the fundamental axes. This talk will present an overview of the experiments and data that is gathered, and outlines the challenges in extracting information. Common strategies employed are very similar to industrial data! Science problems (e.g., data filtering, machine learning, statistical interpretation) and provide a seed for exchange of knowledge between academia and industry. Speaker Biography Professor Mark Sullivan Mark Sullivan is a Professor of Astrophysics in the Department of Physics and Astronomy. Mark completed his PhD at Cambridge, and following postdoctoral study in Durham, Toronto and Oxford, now leads a research group at Southampton studying dark energy using exploding stars called "type Ia supernovae". Mark has many years' experience of research that involves repeatedly imaging the night sky to track the arrival of transient objects, involving significant challenges in data handling, processing, classification and analysis.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

COMPASS is an experiment at CERN’s SPS whose goal is to study hadron structure and spectroscopy. The experiment includes a wide acceptance RICH detector, operating since 2001 and subject to a major upgrade of the central region of its photodetectors in 2006. The remaining 75% of the photodetection area are still using MWPCs from the original design, who suffer from limitations in gain due to aging of the photocathodes from ion bombardment and due to ion-induced instabilities. Besides the mentioned limitations, the increased luminosity conditions expected for the upcoming years of the experiment make an upgrade to the remaining detectors pertinent. This upgrade should be accomplished in 2016, using hybrid detectors composed of ThGEMs and MICROMEGAS. This work presents the study, development and characterization of gaseous photon detectors envisaging the foreseen upgrade, and the progress in production and evaluation techniques necessary to reach increasingly larger area detectors with the performances required. It includes reports on the studies performed under particle beam environment of such detectors. MPGD structures can also be used in a variety of other applications, of which nuclear medical imaging is a notorious example. This work includes, additionally, the initial steps in simulating, assembling and characterizing a prototype of a gaseous detector for application as a Compton Camera.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Since it has been found that the MadGraph Monte Carlo generator offers superior flavour-matching capability as compared to Alpgen, the suitability of MadGraph for the generation of ttb¯ ¯b events is explored, with a view to simulating this background in searches for the Standard Model Higgs production and decay process ttH, H ¯ → b ¯b. Comparisons are performed between the output of MadGraph and that of Alpgen, showing that satisfactory agreement in their predictions can be obtained with the appropriate generator settings. A search for the Standard Model Higgs boson, produced in association with the top quark and decaying into a b ¯b pair, using 20.3 fb−1 of 8 TeV collision data collected in 2012 by the ATLAS experiment at CERN’s Large Hadron Collider, is presented. The GlaNtp analysis framework, together with the RooFit package and associated software, are used to obtain an expected 95% confidence-level limit of 4.2 +4.1 −2.0 times the Standard Model expectation, and the corresponding observed limit is found to be 5.9; this is within experimental uncertainty of the published result of the analysis performed by the ATLAS collaboration. A search for a heavy charged Higgs boson of mass mH± in the range 200 ≤ mH± /GeV ≤ 600, where the Higgs mediates the five-flavour beyond-theStandard-Model physics process gb → tH± → ttb, with one top quark decaying leptonically and the other decaying hadronically, is presented, using the 20.3 fb−1 8 TeV ATLAS data set. Upper limits on the product of the production cross-section and the branching ratio of the H± boson are computed for six mass points, and these are found to be compatible within experimental uncertainty with those obtained by the corresponding published ATLAS analysis.