972 resultados para Read Out Driver, Data Acquisition, Electronics, FPGA, ATLAS, IBL, Pixel Detector, LHC, VME
Resumo:
Stimulated echoes are widely used for imaging functional tissue parameters such as diffusion coefficient, perfusion, and flow rates. They are potentially interesting for the assessment of various cardiac functions. However, severe limitations of the stimulated echo acquisition mode occur, which are related to the special dynamic properties of the beating heart and flowing blood. To the well-known signal decay due to longitudinal relaxation and through-plane motion between the preparation and the read-out period of the stimulated echoes, additional signal loss is often observed. As the prepared magnetization is fixed with respect to the tissue, this signal loss is caused by the tissue deformation during the cardiac cycle, which leads to a modification of the modulation frequency of the magnetization. These effects are theoretically derived and corroborated by phantom and in vivo experiments.
Resumo:
L'expérience LHCb sera installée sur le futur accélérateur LHC du CERN. LHCb est un spectromètre à un bras consacré aux mesures de précision de la violation CP et à l'étude des désintégrations rares des particules qui contiennent un quark b. Actuellement LHCb se trouve dans la phase finale de recherche et développement et de conception. La construction a déjà commencé pour l'aimant et les calorimètres. Dans le Modèle Standard, la violation CP est causée par une phase complexe dans la matrice 3x3 CKM (Cabibbo-Kobayashi-Maskawa) de mélange des quarks. L'expérience LHCb compte utiliser les mesons B pour tester l'unitarité de cette matrice, en mesurant de diverses manières indépendantes tous les angles et côtés du "triangle d'unitarité". Cela permettra de surdéterminer le modèle et, peut-être, de mettre en évidence des incohérences qui seraient le signal de l'existence d'une physique au-delà du Modèle Standard. La reconstruction du vertex de désintégration des particules est une condition fondamentale pour l'expérience LHCb. La présence d'un vertex secondaire déplacé est une signature de la désintégration de particules avec un quark b. Cette signature est utilisée dans le trigger topologique du LHCb. Le Vertex Locator (VeLo) doit fournir des mesures précises de coordonnées de passage des traces près de la région d'interaction. Ces points sont ensuite utilisés pour reconstruire les trajectoires des particules et l'identification des vertices secondaires et la mesure des temps de vie des hadrons avec quark b. L'électronique du VeLo est une partie essentielle du système d'acquisition de données et doit se conformer aux spécifications de l'électronique de LHCb. La conception des circuits doit maximiser le rapport signal/bruit pour obtenir la meilleure performance de reconstruction des traces dans le détecteur. L'électronique, conçue en parallèle avec le développement du détecteur de silicium, a parcouru plusieurs phases de "prototyping" décrites dans cette thèse.<br/><br/>The LHCb experiment is being built at the future LHC accelerator at CERN. It is a forward single-arm spectrometer dedicated to precision measurements of CP violation and rare decays in the b quark sector. Presently it is finishing its R&D and final design stage. The construction already started for the magnet and calorimeters. In the Standard Model, CP violation arises via the complex phase of the 3 x 3 CKM (Cabibbo-Kobayashi-Maskawa) quark mixing matrix. The LHCb experiment will test the unitarity of this matrix by measuring in several theoretically unrelated ways all angles and sides of the so-called "unitary triangle". This will allow to over-constrain the model and - hopefully - to exhibit inconsistencies which will be a signal of physics beyond the Standard Model. The Vertex reconstruction is a fundamental requirement for the LHCb experiment. Displaced secondary vertices are a distinctive feature of b-hadron decays. This signature is used in the LHCb topology trigger. The Vertex Locator (VeLo) has to provide precise measurements of track coordinates close to the interaction region. These are used to reconstruct production and decay vertices of beauty-hadrons and to provide accurate measurements of their decay lifetimes. The Vertex Locator electronics is an essential part of the data acquisition system and must conform to the overall LHCb electronics specification. The design of the electronics must maximise the signal to noise ratio in order to achieve the best tracking reconstruction performance in the detector. The electronics is being designed in parallel with the silicon detector development and went trough several prototyping phases, which are described in this thesis.
Resumo:
Työn tarkoituksena oli kerätä käyttövarmuustietoa savukaasulinjasta kahdelta suomalaiselta sellutehtaalta niiden käyttöönotosta aina tähän päivään asti. Käyttövarmuustieto koostuu luotettavuustiedoista sekä kunnossapitotiedoista. Kerätyn tiedon avulla on mahdollista kuvata tarkasti laitoksen käyttövarmuutta seuraavilla tunnusluvuilla: suunnittelemattomien häiriöiden lukumäärä ja korjausajat, laitteiden seisokkiaika, vikojen todennäköisyys ja korjaavan kunnossapidon kustannukset suhteessa savukaasulinjan korjaavan kunnossapidon kokonaiskustannuksiin. Käyttövarmuustiedon keräysmetodi on esitelty. Savukaasulinjan kriittisten laitteiden määrittelyyn käytetty metodi on yhdistelmä kyselytutkimuksesta ja muunnellusta vian vaikutus- ja kriittisyysanalyysistä. Laitteiden valitsemiskriteerit lopulliseen kriittisyysanalyysiin päätettiin käyttövarmuustietojen sekä kyselytutkimuksen perusteella. Kriittisten laitteiden määrittämisen tarkoitus on löytää savukaasulinjasta ne laitteet, joiden odottamaton vikaantuminen aiheuttaa vakavimmat seuraukset savukaasulinjan luotettavuuteen, tuotantoon, turvallisuuteen, päästöihin ja kustannuksiin. Tiedon avulla rajoitetut kunnossapidon resurssit voidaan suunnata oikein. Kriittisten laitteiden määrittämisen tuloksena todetaan, että kolme kriittisintä laitetta savukaasulinjassa ovat molemmille sellutehtaille yhteisesti: savukaasupuhaltimet, laahakuljettimet sekä ketjukuljettimet. Käyttövarmuustieto osoittaa, että laitteiden luotettavuus on tehdaskohtaista, mutta periaatteessa samat päälinjat voidaan nähdä suunnittelemattomien vikojen todennäköisyyttä esittävissä kuvissa. Kustannukset, jotka esitetään laitteen suunnittelemattomien kunnossapitokustannusten suhteena savukaasulinjan kokonaiskustannuksiin, noudattelevat hyvin pitkälle luotettavuuskäyrää, joka on laskettu laitteen seisokkiajan suhteena käyttötunteihin. Käyttövarmuustiedon keräys yhdistettynä kriittisten laitteiden määrittämiseen mahdollistavat ennakoivan kunnossapidon oikean kohdistamisen ja ajoittamisen laitteiston elinaikana siten, että luotettavuus- ja kustannustehokkuusvaatimukset saavutetaan.
Resumo:
Tämä työ tehtiin globaaliin elektroniikka-alan yritykseen. Diplomityö liittyy haasteeseen, jonka lisääntynyt globalisaatio ja kiristyvä kilpailu ovat luoneet: case yrityksen on selvitettävä kuinka se voi saavuttaa kasvutavoitteet myös tulevaisuudessa hankkimalla uusia asiakkaita ja olemalla yhä enenevissä määrin maailmanlaajuisesti läsnä. Tutkimuksen tavoite oli löytää sopiva malli potentiaalisten avainasiakkaiden identifiointiin ja valintaan, sekä testata ja modifioida valittua mallia case yrityksen tarpeiden mukaisesti. Erityisesti raakadatan kerääminen, asiakkaiden houkuttelevuuskriteerit ja kohdemarkkinarako olivat asioita, jotka tarvitsivat tutkimuksessa huomiota. Kirjallisuuskatsauksessa keskityttiin yritysmarkkinoihin, eri asiakassuhteenhallinnan lähestymistapoihin ja avainasiakkaiden määrittämiseen. CRM:n, KAM:n ja Customer Insight-ajattelun perusteet esiteltiin yhdessä eri avainasiakkaiden identifiointimallien kanssa. Valittua Chevertonin mallia testattiin ja muokattiin työn empiirisessä osassa. Tutkimuksen empiirinen kontribuutio on modifioitu malli potentiaalisten avainasiakkaiden identifiointiin. Se auttaa päätöksentekijöitä etenemään systemaattisesti ja organisoidusti askel askeleelta kohti potentiaalisten asiakkaiden listaa tietyltä markkina-alueelta. Työ tarjoaa työkalun tähän prosessiin sekä luo pohjaa tulevaisuuden tutkimukselle ja toimenpiteille.
Resumo:
Most of the applications of airborne laser scanner data to forestry require that the point cloud be normalized, i.e., each point represents height from the ground instead of elevation. To normalize the point cloud, a digital terrain model (DTM), which is derived from the ground returns in the point cloud, is employed. Unfortunately, extracting accurate DTMs from airborne laser scanner data is a challenging task, especially in tropical forests where the canopy is normally very thick (partially closed), leading to a situation in which only a limited number of laser pulses reach the ground. Therefore, robust algorithms for extracting accurate DTMs in low-ground-point-densitysituations are needed in order to realize the full potential of airborne laser scanner data to forestry. The objective of this thesis is to develop algorithms for processing airborne laser scanner data in order to: (1) extract DTMs in demanding forest conditions (complex terrain and low number of ground points) for applications in forestry; (2) estimate canopy base height (CBH) for forest fire behavior modeling; and (3) assess the robustness of LiDAR-based high-resolution biomass estimation models against different field plot designs. Here, the aim is to find out if field plot data gathered by professional foresters can be combined with field plot data gathered by professionally trained community foresters and used in LiDAR-based high-resolution biomass estimation modeling without affecting prediction performance. The question of interest in this case is whether or not the local forest communities can achieve the level technical proficiency required for accurate forest monitoring. The algorithms for extracting DTMs from LiDAR point clouds presented in this thesis address the challenges of extracting DTMs in low-ground-point situations and in complex terrain while the algorithm for CBH estimation addresses the challenge of variations in the distribution of points in the LiDAR point cloud caused by things like variations in tree species and season of data acquisition. These algorithms are adaptive (with respect to point cloud characteristics) and exhibit a high degree of tolerance to variations in the density and distribution of points in the LiDAR point cloud. Results of comparison with existing DTM extraction algorithms showed that DTM extraction algorithms proposed in this thesis performed better with respect to accuracy of estimating tree heights from airborne laser scanner data. On the other hand, the proposed DTM extraction algorithms, being mostly based on trend surface interpolation, can not retain small artifacts in the terrain (e.g., bumps, small hills and depressions). Therefore, the DTMs generated by these algorithms are only suitable for forestry applications where the primary objective is to estimate tree heights from normalized airborne laser scanner data. On the other hand, the algorithm for estimating CBH proposed in this thesis is based on the idea of moving voxel in which gaps (openings in the canopy) which act as fuel breaks are located and their height is estimated. Test results showed a slight improvement in CBH estimation accuracy over existing CBH estimation methods which are based on height percentiles in the airborne laser scanner data. However, being based on the idea of moving voxel, this algorithm has one main advantage over existing CBH estimation methods in the context of forest fire modeling: it has great potential in providing information about vertical fuel continuity. This information can be used to create vertical fuel continuity maps which can provide more realistic information on the risk of crown fires compared to CBH.
Resumo:
Les seize détecteurs MPX constituant le réseau ATLAS-MPX ont été placés à différentes positions dans le détecteur ATLAS et sa averne au CERN dans le but de mesurer en emps réel les champs de radiation produits ar des particules primaires (protons des faisceaux) et des particules secondaires (kaons, pions, g, protons) issues des collisions proton-proton. Des films de polyéthylène (PE) et de fluorure de lithium (6LiF) recouvrent les détecteurs afin d’augmenter leur sensibilité aux neutrons produits par les particules primaires et secondaires interagissant avec les matériaux présents dans l’environnement d’ATLAS. La reconnaissance des traces laissées par les particules dans un détecteur ATLAS-MPX se fait à partir des algorithmes du logiciel MAFalda (“Medipix Analysis Framework”) basé sur les librairies et le logiciel d’analyse de données ROOT. Une étude sur le taux d’identifications erronées et le chevauchement d’amas a été faite en reconstruisant les activités des sources 106Ru et 137Cs. L’efficacité de détection des neutrons rapides a été mesurée à l’aide des sources 252Cf et 241AmBe (neutrons d’énergie moyenne de 2.13 et 4.08 MeV respectivement). La moyenne des efficacités de détection mesurées pour les neutrons produits par les sources 252C f et 241AmBe a été calculée pour les convertisseurs 6LiF et PE et donnent (0.8580 ± 0.1490)% et (0.0254 ± 0.0031)% pour LiF et (0.0510 ± 0.0061)% et (0.0591 ± 0.0063)% pour PE à bas et à haut seuil d’énergie respectivement. Une simulation du calcul de l’efficacité de détection des neutrons dans le détecteur MPX a été réalisée avec le logiciel GEANT4. Des données MPX correspondant aux collisions proton-proton à 2.4 TeV et à 7 TeV dans le centre de masse ont été analysées. Les flux détectés d’électrons et de photons sont particulièrement élevés dans les détecteurs MPX01 et MPX14 car ils sont plus près du point de collision. Des flux de neutrons ont été estimés en utilisant les efficacités de détection mesurées. Une corrélation avec la luminosité du LHC a été établie et on prédit que pour les collisions à 14 TeV dans le centre de masse et avec une luminosité de 10^34 cm-1*s-1 il y aura environ 5.1x10^8 ± 1.5x10^7 et 1.6x10^9 ± 6.3x10^7 particules détectées par les détecteurs MPX01 et MPX14 respectivement.
Resumo:
The TCABR data analysis and acquisition system has been upgraded to support a joint research programme using remote participation technologies. The architecture of the new system uses Java language as programming environment. Since application parameters and hardware in a joint experiment are complex with a large variability of components, requirements and specification solutions need to be flexible and modular, independent from operating system and computer architecture. To describe and organize the information on all the components and the connections among them, systems are developed using the extensible Markup Language (XML) technology. The communication between clients and servers uses remote procedure call (RPC) based on the XML (RPC-XML technology). The integration among Java language, XML and RPC-XML technologies allows to develop easily a standard data and communication access layer between users and laboratories using common software libraries and Web application. The libraries allow data retrieval using the same methods for all user laboratories in the joint collaboration, and the Web application allows a simple graphical user interface (GUI) access. The TCABR tokamak team in collaboration with the IPFN (Instituto de Plasmas e Fusao Nuclear, Instituto Superior Tecnico, Universidade Tecnica de Lisboa) is implementing this remote participation technologies. The first version was tested at the Joint Experiment on TCABR (TCABRJE), a Host Laboratory Experiment, organized in cooperation with the IAEA (International Atomic Energy Agency) in the framework of the IAEA Coordinated Research Project (CRP) on ""Joint Research Using Small Tokamaks"". (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
Each plasma physics laboratory has a proprietary scheme to control and data acquisition system. Usually, it is different from one laboratory to another. It means that each laboratory has its own way to control the experiment and retrieving data from the database. Fusion research relies to a great extent on international collaboration and this private system makes it difficult to follow the work remotely. The TCABR data analysis and acquisition system has been upgraded to support a joint research programme using remote participation technologies. The choice of MDSplus (Model Driven System plus) is proved by the fact that it is widely utilized, and the scientists from different institutions may use the same system in different experiments in different tokamaks without the need to know how each system treats its acquisition system and data analysis. Another important point is the fact that the MDSplus has a library system that allows communication between different types of language (JAVA, Fortran, C, C++, Python) and programs such as MATLAB, IDL, OCTAVE. In the case of tokamak TCABR interfaces (object of this paper) between the system already in use and MDSplus were developed, instead of using the MDSplus at all stages, from the control, and data acquisition to the data analysis. This was done in the way to preserve a complex system already in operation and otherwise it would take a long time to migrate. This implementation also allows add new components using the MDSplus fully at all stages. (c) 2012 Elsevier B.V. All rights reserved.
Resumo:
This work proposes the development of an Adaptive Neuro-fuzzy Inference System (ANFIS) estimator applied to speed control in a three-phase induction motor sensorless drive. Usually, ANFIS is used to replace the traditional PI controller in induction motor drives. The evaluation of the estimation capability of the ANFIS in a sensorless drive is one of the contributions of this work. The ANFIS speed estimator is validated in a magnetizing flux oriented control scheme, consisting in one more contribution. As an open-loop estimator, it is applied to moderate performance drives and it is not the proposal of this work to solve the low and zero speed estimation problems. Simulations to evaluate the performance of the estimator considering the vector drive system were done from the Matlab/Simulink(R) software. To determine the benefits of the proposed model, a practical system was implemented using a voltage source inverter (VSI) to drive the motor and the vector control including the ANFIS estimator, which is carried out by the Real Time Toolbox from Matlab/Simulink(R) software and a data acquisition card from National Instruments.
Resumo:
Several activities were conducted during my PhD activity. For the NEMO experiment a collaboration between the INFN/University groups of Catania and Bologna led to the development and production of a mixed signal acquisition board for the Nemo Km3 telescope. The research concerned the feasibility study for a different acquisition technique quite far from that adopted in the NEMO Phase 1 telescope. The DAQ board that we realized exploits the LIRA06 front-end chip for the analog acquisition of anodic an dynodic sources of a PMT (Photo-Multiplier Tube). The low-power analog acquisition allows to sample contemporaneously multiple channels of the PMT at different gain factors in order to increase the signal response linearity over a wider dynamic range. Also the auto triggering and self-event-classification features help to improve the acquisition performance and the knowledge on the neutrino event. A fully functional interface towards the first level data concentrator, the Floor Control Module, has been integrated as well on the board, and a specific firmware has been realized to comply with the present communication protocols. This stage of the project foresees the use of an FPGA, a high speed configurable device, to provide the board with a flexible digital logic control core. After the validation of the whole front-end architecture this feature would be probably integrated in a common mixed-signal ASIC (Application Specific Integrated Circuit). The volatile nature of the configuration memory of the FPGA implied the integration of a flash ISP (In System Programming) memory and a smart architecture for a safe remote reconfiguration of it. All the integrated features of the board have been tested. At the Catania laboratory the behavior of the LIRA chip has been investigated in the digital environment of the DAQ board and we succeeded in driving the acquisition with the FPGA. The PMT pulses generated with an arbitrary waveform generator were correctly triggered and acquired by the analog chip, and successively they were digitized by the on board ADC under the supervision of the FPGA. For the communication towards the data concentrator a test bench has been realized in Bologna where, thanks to a lending of the Roma University and INFN, a full readout chain equivalent to that present in the NEMO phase-1 was installed. These tests showed a good behavior of the digital electronic that was able to receive and to execute command imparted by the PC console and to answer back with a reply. The remotely configurable logic behaved well too and demonstrated, at least in principle, the validity of this technique. A new prototype board is now under development at the Catania laboratory as an evolution of the one described above. This board is going to be deployed within the NEMO Phase-2 tower in one of its floors dedicated to new front-end proposals. This board will integrate a new analog acquisition chip called SAS (Smart Auto-triggering Sampler) introducing thus a new analog front-end but inheriting most of the digital logic present in the current DAQ board discussed in this thesis. For what concern the activity on high-resolution vertex detectors, I worked within the SLIM5 collaboration for the characterization of a MAPS (Monolithic Active Pixel Sensor) device called APSEL-4D. The mentioned chip is a matrix of 4096 active pixel sensors with deep N-well implantations meant for charge collection and to shield the analog electronics from digital noise. The chip integrates the full-custom sensors matrix and the sparsifification/readout logic realized with standard-cells in STM CMOS technology 130 nm. For the chip characterization a test-beam has been set up on the 12 GeV PS (Proton Synchrotron) line facility at CERN of Geneva (CH). The collaboration prepared a silicon strip telescope and a DAQ system (hardware and software) for data acquisition and control of the telescope that allowed to store about 90 million events in 7 equivalent days of live-time of the beam. My activities concerned basically the realization of a firmware interface towards and from the MAPS chip in order to integrate it on the general DAQ system. Thereafter I worked on the DAQ software to implement on it a proper Slow Control interface of the APSEL4D. Several APSEL4D chips with different thinning have been tested during the test beam. Those with 100 and 300 um presented an overall efficiency of about 90% imparting a threshold of 450 electrons. The test-beam allowed to estimate also the resolution of the pixel sensor providing good results consistent with the pitch/sqrt(12) formula. The MAPS intrinsic resolution has been extracted from the width of the residual plot taking into account the multiple scattering effect.
Resumo:
Mit der Erweiterung des Elektronenbeschleunigers MAMI um eine dritte Stufe ist es möglich geworden, am Institut für Kernphysik Teilchen mit offener Strangeness zu produzieren. Für deren Nachweis ist die Drei-Spektrometeranlage der Kollaboration A1 um das von der GSI in Darmstadt übernommene KAOS-Spektrometer erweitert worden. Untersucht wird damit die elementare Reaktion p(e,e' K+)Lambda/Sigma0 wobei das auslaufende Elektron und das Kaon nachgewiesen werden müssen. Wird als Target nicht Wasserstoff verwendet, besteht die Möglichkeit dass sich ein Hyperkern bildet. Spektroskopische Untersuchungen an diesen bieten die Möglichkeit das Potential von Hyperonen in Atomkernen und die Hyperon-Nukleon-Wechselwirkung zu untersuchen. Aufgrund der hervorragenden Strahlqualität bei der Elektroproduktion können hier Massenauflösungen von einigen hundert keV/c² erreicht werden. Mit Hilfe von GEANT4 wurden die Detektoren und die Abbildungseigenschaften des Spektrometers simuliert. Geeignete Ereignisgeneratoren wurden implementiert. Es wurde untersucht, wie mögliche Treffermuster in den Detektoren aussehen, die von einem Trigger auf FPGA-Basis selektiert werden müssen. Ebenso konnte hieraus eine erste Abbildung der Spurkoordinaten auf die Targetkoordinaten und den Teilchenimpuls gewonnen werden. Für das Hyperkernprogramm muss KAOS unter 0° Vorwärtsrichung betrieben werden und der Primärstrahl mit Hilfe einer Schikane durch den Dipol gelenkt werden. Die Simulation zeigt hier eine nur moderate Erhöhung der Strahlenbelastung, vor allem im Bereich des Strahlfängers. Somit ist es möglich, KAOS als doppelseitiges Spektrometer in der Spektrometerhalle zu betreiben. Im Rahmen dieser Arbeit wurden die für sämtliche Detektoren nötige Auslese- und Steuerungselektronik in das vorhandene Datenerfassungssystem und das Steuerungssystem eingebunden. In zwei Strahlzeiten im Herbst 2008 wurden Kaonen im Winkelbereich von 20°-40° mit Impulsen zwischen 400MeV/c und 600MeV/c nachgewiesen. Die aus der Simulation gewonnenen Daten zum Trigger und zur Abbildung kamen zum Einsatz. Es konnte die für eine gute Teilchenidentifikation nötige Zeitauflösung von ca. 1ns FWHM erreicht werden. Die erreichte Winkel- und Impulsauflösung war ausreichend um Lambda und Sigma0-Hyperonen im Spektrum der fehlenden Masse leicht trennen zu können.
Resumo:
Diese Arbeit befasst sich mit der photoinduzierten Erzeugung neutraler Pionen, sehr nahe an der Schwellenenergie. Dabei werden zwei Ziele verfolgt: Zum einen die Überprüfung von Vorhersagen dieser effektiven Theorien und Modelle, zum anderen werden hier erstmals alle relevanten Partialwellenamplituden modellunabhängig aus gemessenen Observablen bestimmt. Diese Methode soll in Zukunft auch bei höheren Energien im Bereich der Nukleonresonanzen Anwendung finden. rnrnKonkret wird die Durchführung und Analyse eines Experiments vorgestellt, welches am Mainzer Mikrotron (MAMI) in den Jahren 2010 bis 2013 mit zirkular polarisiertem Photonenstrahl stattfand. Der Photonenstrahl wurde an einer Anlage zur Erzeugung energiemarkierter Bremsstrahlung aus dem Elektronenstrahl von MAMI gewonnen. Zum Nachweis der Reaktionsprodukte diente das hermetische 4pi CB/TAPS-Detektorsystem. Erstmalig standen bei derartigen Messungen auch transversal polarisierte Protonen zur Verfügung. Dazu wird Butanol in einer speziellen Apparatur dynamisch polarisiert. Molekularer Wasserstoff lässt sich aufgrund der para-Konfiguration nicht polarisieren. Wegen der Verwendung von Butanol als Targetmaterial, bei dem weniger als 5% aller erzeugten Pionen an polarisierten Protonen produziert wurden, ist die Behandlung des Untergrunds eine zentrale Aufgabe. rnrnEs werden zwei Methoden der Untergrundseparation vorgestellt, wovon die bessere in der Analyse angewendet wurde. Abschließend findet eine ausführliche Bewertung systematischer Fehler statt.rnrnDie erstmalige Verwendung transversal polarisierter Protonen ermöglicht den Zugang zu bisher nicht gemessenen Spin"=Freiheitsgraden. In Kombination mit einem komplementären Vorläufer-Experiment aus dem Jahr 2008 mit linear polarisiertem Photonenstrahl konnten aus den gewonnenen Daten erstmals alle komplexen s- und p-Partialwellenamplituden modellunabhängig bestimmt werden. rnrnDarüber hinaus wurden im Rahmen dieser Arbeit wesentliche Verbesserungen am apparativen Aufbau erzielt. Beispiele sind ein Elektronenstrahl-Polarimeter, ein zellularer CB-Multiplizitätstrigger, sowie signifikante Verbesserungen der Datennahmeelektronik und des Triggersystems, die teilweise in dieser Arbeit vorgestellt werden.
Resumo:
The ArgoNeuT liquid argon time projection chamber has collected thousands of neutrino and anti-neutrino events during an extended run period in the NuMI beam-line at Fermilab. This paper focuses on the main aspects of the detector layout and related technical features, including the cryogenic equipment, time projection chamber, read-out electronics, and off-line data treatment. The detector commissioning phase, physics run, and first neutrino event displays are also reported. The characterization of the main working parameters of the detector during data-taking, the ionization electron drift velocity and lifetime in liquid argon, as obtained from through-going muon data complete the present report.
Resumo:
A search for scalar particles decaying via narrow resonances into two photons in the mass range 65–600 GeV is performed using 20.3 fb −1 of s √ =8 TeV pp collision data collected with the ATLAS detector at the Large Hadron Collider. The recently discovered Higgs boson is treated as a background. No significant evidence for an additional signal is observed. The results are presented as limits at the 95% confidence level on the production cross section of a scalar boson times branching ratio into two photons, in a fiducial volume where the reconstruction efficiency is approximately independent of the event topology. The upper limits set extend over a considerably wider mass range than previous searches.
Resumo:
Results of a search for supersymmetry via direct production of third-generation squarks are reported, using 20.3 fb −1 of proton-proton collision data at √s =8 TeV recorded by the ATLAS experiment at the LHC in 2012. Two different analysis strategies based on monojetlike and c -tagged event selections are carried out to optimize the sensitivity for direct top squark-pair production in the decay channel to a charm quark and the lightest neutralino (t 1 →c+χ ˜ 0 1 ) across the top squark–neutralino mass parameter space. No excess above the Standard Model background expectation is observed. The results are interpreted in the context of direct pair production of top squarks and presented in terms of exclusion limits in the m ˜t 1, m ˜ X0 1 ) parameter space. A top squark of mass up to about 240 GeV is excluded at 95% confidence level for arbitrary neutralino masses, within the kinematic boundaries. Top squark masses up to 270 GeV are excluded for a neutralino mass of 200 GeV. In a scenario where the top squark and the lightest neutralino are nearly degenerate in mass, top squark masses up to 260 GeV are excluded. The results from the monojetlike analysis are also interpreted in terms of compressed scenarios for top squark-pair production in the decay channel t ˜ 1 →b+ff ′ +χ ˜ 0 1 and sbottom pair production with b ˜ 1 →b+χ ˜ 0 1 , leading to a similar exclusion for nearly mass-degenerate third-generation squarks and the lightest neutralino. The results in this paper significantly extend previous results at colliders.