949 resultados para missing data imputation
Resumo:
A search is presented for new particles decaying to large numbers (7 to greater or equal to 10) of jets, missing transverse momentum and no isolated electrons or muons. This analysis uses 20.3/fb of pp collision data at sqrt(s)=8 TeV collected by the ATLAS experiment at the Large Hadron Collider. The sensitivity of the search is enhanced by considering the number of b-tagged jets and the scalar sum of masses of large-radius jets in an event. No evidence is found for physics beyond the Standard Model. The results are interpreted in the context of various simplified supersymmetry-inspired models where gluinos are pair produced, as well as a mSUGRA/CMSSM model.
Resumo:
A search is presented for dark matter pair production in association with a W or Z boson in pp collisions representing 20.3 fb(-1) of integrated luminosity at root s = 8 TeV using data recorded with the ATLAS detector at the Large Hadron Collider. Events with a hadronic jet with the jet mass consistent with a W or Z boson, and with large missing transverse momentum are analyzed. The data are consistent with the standard model expectations. Limits are set on the mass scale in effective field theories that describe the interaction of dark matter and standard model particles, and on the cross section of Higgs production and decay to invisible particles. In addition, cross section limits on the anomalous production of W or Z bosons with large missing transverse momentum are set in two fiducial regions.
Resumo:
Background Aerosolized vaccine can be used as a needle-free method of immunization against measles, a disease that remains a major cause of illness and death. Data on the immunogenicity of aerosolized vaccine against measles in children are inconsistent. Methods We conducted an open-label noninferiority trial involving children 9.0 to 11.9 months of age in India who were eligible to receive a first dose of measles vaccine. Children were randomly assigned to receive a single dose of vaccine by means of either aerosol inhalation or a subcutaneous injection. The primary end points were seropositivity for antibodies against measles and adverse events 91 days after vaccination. The noninferiority margin was 5 percentage points. Results A total of 1001 children were assigned to receive aerosolized vaccine, and 1003 children were assigned to receive subcutaneous vaccine; 1956 of all the children (97.6%) were followed to day 91, but outcome data were missing for 331 children because of thawed specimens. In the per-protocol population, data on 1560 of 2004 children (77.8%) could be evaluated. At day 91, a total of 662 of 775 children (85.4%; 95% confidence interval [CI], 82.5 to 88.0) in the aerosol group, as compared with 743 of 785 children (94.6%; 95% CI, 92.7 to 96.1) in the subcutaneous group, were seropositive, a difference of -9.2 percentage points (95% CI, -12.2 to -6.3). Findings were similar in the full-analysis set (673 of 788 children in the aerosol group [85.4%] and 754 of 796 children in the subcutaneous group [94.7%] were seropositive at day 91, a difference of -9.3 percentage points [95% CI, -12.3 to -6.4]) and after multiple imputation of missing results. No serious adverse events were attributable to measles vaccination. Adverse-event profiles were similar in the two groups. Conclusions Aerosolized vaccine against measles was immunogenic, but, at the prespecified margin, the aerosolized vaccine was inferior to the subcutaneous vaccine with respect to the rate of seropositivity. (Funded by the Bill and Melinda Gates Foundation; Measles Aerosol Vaccine Project Clinical Trials Registry-India number, CTRI/2009/091/000673 .).
Resumo:
A search has been performed, using the full 20.3 fb −1 data sample of 8 TeV proton-proton collisions collected in 2012 with the ATLAS detector at the LHC, for photons originating from a displaced vertex due to the decay of a neutral long-lived particle into a photon and an invisible particle. The analysis investigates the diphoton plus missing transverse momentum final state, and is therefore most sensitive to pair production of long-lived particles. The analysis technique exploits the capabilities of the ATLAS electromagnetic calorimeter to make precise measurements of the flight direction, as well as the time of flight, of photons. No excess is observed over the Standard Model predictions for background. Exclusion limits are set within the context of gauge mediated supersymmetry breaking models, with the lightest neutralino being the next-to-lightest supersymmetric particle and decaying into a photon and gravitino with a lifetime in the range from 250 ps to about 100 ns.
Resumo:
This paper presents a search for new particles in events with one lepton (electron or muon) and missing transverse momentum using 20.3 fb−1 of proton-proton collision data at TeX = 8 TeV recorded by the ATLAS experiment at the Large Hadron Collider. No significant excess beyond Standard Model expectations is observed. A W ′ with Sequential Standard Model couplings is excluded at the 95% confidence level for masses up to 3.24 TeV. Excited chiral bosons (W *) with equivalent coupling strengths are excluded for masses up to 3.21 TeV. In the framework of an effective field theory limits are also set on the dark matter-nucleon scattering cross-section as well as the mass scale M * of the unknown mediating interaction for dark matter pair production in association with a leptonically decaying W.
Resumo:
The liquid argon calorimeter is a key component of the ATLAS detector installed at the CERN Large Hadron Collider. The primary purpose of this calorimeter is the measurement of electron and photon kinematic properties. It also provides a crucial input for measuring jets and missing transverse momentum. An advanced data monitoring procedure was designed to quickly identify issues that would affect detector performance and ensure that only the best quality data are used for physics analysis. This article presents the validation procedure developed during the 2011 and 2012 LHC data-taking periods, in which more than 98% of the proton-proton luminosity recorded by ATLAS at a centre-of-mass energy of 7–8 TeV had calorimeter data quality suitable for physics analysis.
Resumo:
Searches for the electroweak production of charginos, neutralinos and sleptons in final states characterized by the presence of two leptons (electrons and muons) and missing transverse momentum are performed using 20.3 fb−1 of proton-proton collision data at ps = 8TeV recorded with the ATLAS experiment at the Large Hadron Collider. No significant excess beyond Standard Model expectations is observed. Limits are set on the masses of the lightest chargino, next-to-lightest neutralino and sleptons for different lightest-neutralino mass hypotheses in simplified models. Results are also interpreted in various scenarios of the phenomenological Minimal Supersymmetric Standard Model.
Resumo:
A search is presented for direct top squark pair production using events with at least two leptons including a same-flavour opposite-sign pair with invariant mass consistent with the Z boson mass, jets tagged as originating from b-quarks and missing transverse momentum. The analysis is performed with proton–proton collision data at √ s = 8 TeV collected with the ATLAS detector at the LHC in 2012 corresponding to an integrated luminosity of 20.3 fb−1. No excess beyond the Standard Model expectation is observed. Interpretations of the results are provided in models based on the direct pair production of the heavier top squark state (˜t2) followed by the decay to the lighter top squark state (˜t1) via ˜t2 → Z ˜t1, and for ˜t1 pair production in natural gaugemediated supersymmetry breaking scenarios where the neutralino (˜χ 01 ) is the next-to-lightest supersymmetric particle and decays producing a Z boson and a gravitino ( ˜G ) via the ˜χ 01→ Z ˜G process.
Resumo:
A search for the direct production of charginos and neutralinos in final states with three leptons and missing transverse momentum is presented. The analysis is based on 20.3 fb−1 of √s = 8TeV proton-proton collision data delivered by the Large Hadron Collider and recorded with the ATLAS detector. Observations are consistent with the Standard Model expectations and limits are set in R-parity-conserving phenomenological Minimal Supersymmetric Standard Models and in simplified supersymmetric models, significantly extending previous results. For simplified supersymmetric models of direct chargino (˜χ±1 ) and next-to-lightest neutralino (˜χ02) production with decays to lightest neutralino(˜χ01) via either all three generations of sleptons, staus only, gauge bosons, or Higgs bosons, ˜χ±1 and ˜χ02 masses are excluded up to 700GeV, 380GeV, 345GeV, or 148GeV respectively, for a massless ˜χ01.
Resumo:
BACKGROUND A cost-effective strategy to increase the density of available markers within a population is to sequence a small proportion of the population and impute whole-genome sequence data for the remaining population. Increased densities of typed markers are advantageous for genome-wide association studies (GWAS) and genomic predictions. METHODS We obtained genotypes for 54 602 SNPs (single nucleotide polymorphisms) in 1077 Franches-Montagnes (FM) horses and Illumina paired-end whole-genome sequencing data for 30 FM horses and 14 Warmblood horses. After variant calling, the sequence-derived SNP genotypes (~13 million SNPs) were used for genotype imputation with the software programs Beagle, Impute2 and FImpute. RESULTS The mean imputation accuracy of FM horses using Impute2 was 92.0%. Imputation accuracy using Beagle and FImpute was 74.3% and 77.2%, respectively. In addition, for Impute2 we determined the imputation accuracy of all individual horses in the validation population, which ranged from 85.7% to 99.8%. The subsequent inclusion of Warmblood sequence data further increased the correlation between true and imputed genotypes for most horses, especially for horses with a high level of admixture. The final imputation accuracy of the horses ranged from 91.2% to 99.5%. CONCLUSIONS Using Impute2, the imputation accuracy was higher than 91% for all horses in the validation population, which indicates that direct imputation of 50k SNP-chip data to sequence level genotypes is feasible in the FM population. The individual imputation accuracy depended mainly on the applied software and the level of admixture.
Resumo:
1. Positive interactions among plants can increase species richness by relaxing environmental filters and providing more heterogeneous environments. However, it is not known if facilitation could affect coexistence through other mechanisms. Most studies on plant coexistence focus on negative frequency-dependent mechanisms (decreasing the abundance of common species); here, we test if facilitation can enhance coexistence by giving species an advantage when rare. 2. To test our hypothesis, we used a global data set from drylands and alpine environments and measured the intensity of facilitation (based on co-occurrences with nurse plants) for 48 species present in at least 4 different sites and with a range of abundances in the field. We compared these results with the degree of facilitation experienced by species which are globally rare or common (according to the IUCN Red List), and with a larger data base including over 1200 co-occurrences of target species with their nurses. 3. Facilitation was stronger for rare species (i.e. those having lower local abundances or considered endangered by the IUCN) than for common species, and strongly decreased with the abundance of the facilitated species. These results hold after accounting for the distance of each species from its ecological optimum (i.e. the degree of functional stress it experiences). 4. Synthesis. Our results highlight that nurse plants not only increase the number of species able to colonize a given site, but may also promote species coexistence by preventing the local extinction of rare species. Our findings illustrate the role that nurse plants play in conserving endangered species and link the relationship between facilitation and diversity with coexistence theory. As such, they provide further mechanistic understanding on how facilitation maintains plant diversity.
Resumo:
METHODS Spirometry datasets from South-Asian children were collated from four centres in India and five within the UK. Records with transcription errors, missing values for height or spirometry, and implausible values were excluded(n = 110). RESULTS Following exclusions, cross-sectional data were available from 8,124 children (56.3% male; 5-17 years). When compared with GLI-predicted values from White Europeans, forced expired volume in 1s (FEV1) and forced vital capacity (FVC) in South-Asian children were on average 15% lower, ranging from 4-19% between centres. By contrast, proportional reductions in FEV1 and FVC within all but two datasets meant that the FEV1/FVC ratio remained independent of ethnicity. The 'GLI-Other' equation fitted data from North India reasonably well while 'GLI-Black' equations provided a better approximation for South-Asian data than the 'GLI-White' equation. However, marked discrepancies in the mean lung function z-scores between centres especially when examined according to socio-economic conditions precluded derivation of a single South-Asian GLI-adjustment. CONCLUSION Until improved and more robust prediction equations can be derived, we recommend the use of 'GLI-Black' equations for interpreting most South-Asian data, although 'GLI-Other' may be more appropriate for North Indian data. Prospective data collection using standardised protocols to explore potential sources of variation due to socio-economic circumstances, secular changes in growth/predictors of lung function and ethnicities within the South-Asian classification are urgently required.
Resumo:
Logistic regression is one of the most important tools in the analysis of epidemiological and clinical data. Such data often contain missing values for one or more variables. Common practice is to eliminate all individuals for whom any information is missing. This deletion approach does not make efficient use of available information and often introduces bias.^ Two methods were developed to estimate logistic regression coefficients for mixed dichotomous and continuous covariates including partially observed binary covariates. The data were assumed missing at random (MAR). One method (PD) used predictive distribution as weight to calculate the average of the logistic regressions performing on all possible values of missing observations, and the second method (RS) used a variant of resampling technique. Additional seven methods were compared with these two approaches in a simulation study. They are: (1) Analysis based on only the complete cases, (2) Substituting the mean of the observed values for the missing value, (3) An imputation technique based on the proportions of observed data, (4) Regressing the partially observed covariates on the remaining continuous covariates, (5) Regressing the partially observed covariates on the remaining continuous covariates conditional on response variable, (6) Regressing the partially observed covariates on the remaining continuous covariates and response variable, and (7) EM algorithm. Both proposed methods showed smaller standard errors (s.e.) for the coefficient involving the partially observed covariate and for the other coefficients as well. However, both methods, especially PD, are computationally demanding; thus for analysis of large data sets with partially observed covariates, further refinement of these approaches is needed. ^
Resumo:
My dissertation focuses mainly on Bayesian adaptive designs for phase I and phase II clinical trials. It includes three specific topics: (1) proposing a novel two-dimensional dose-finding algorithm for biological agents, (2) developing Bayesian adaptive screening designs to provide more efficient and ethical clinical trials, and (3) incorporating missing late-onset responses to make an early stopping decision. Treating patients with novel biological agents is becoming a leading trend in oncology. Unlike cytotoxic agents, for which toxicity and efficacy monotonically increase with dose, biological agents may exhibit non-monotonic patterns in their dose-response relationships. Using a trial with two biological agents as an example, we propose a phase I/II trial design to identify the biologically optimal dose combination (BODC), which is defined as the dose combination of the two agents with the highest efficacy and tolerable toxicity. A change-point model is used to reflect the fact that the dose-toxicity surface of the combinational agents may plateau at higher dose levels, and a flexible logistic model is proposed to accommodate the possible non-monotonic pattern for the dose-efficacy relationship. During the trial, we continuously update the posterior estimates of toxicity and efficacy and assign patients to the most appropriate dose combination. We propose a novel dose-finding algorithm to encourage sufficient exploration of untried dose combinations in the two-dimensional space. Extensive simulation studies show that the proposed design has desirable operating characteristics in identifying the BODC under various patterns of dose-toxicity and dose-efficacy relationships. Trials of combination therapies for the treatment of cancer are playing an increasingly important role in the battle against this disease. To more efficiently handle the large number of combination therapies that must be tested, we propose a novel Bayesian phase II adaptive screening design to simultaneously select among possible treatment combinations involving multiple agents. Our design is based on formulating the selection procedure as a Bayesian hypothesis testing problem in which the superiority of each treatment combination is equated to a single hypothesis. During the trial conduct, we use the current values of the posterior probabilities of all hypotheses to adaptively allocate patients to treatment combinations. Simulation studies show that the proposed design substantially outperforms the conventional multi-arm balanced factorial trial design. The proposed design yields a significantly higher probability for selecting the best treatment while at the same time allocating substantially more patients to efficacious treatments. The proposed design is most appropriate for the trials combining multiple agents and screening out the efficacious combination to be further investigated. The proposed Bayesian adaptive phase II screening design substantially outperformed the conventional complete factorial design. Our design allocates more patients to better treatments while at the same time providing higher power to identify the best treatment at the end of the trial. Phase II trial studies usually are single-arm trials which are conducted to test the efficacy of experimental agents and decide whether agents are promising to be sent to phase III trials. Interim monitoring is employed to stop the trial early for futility to avoid assigning unacceptable number of patients to inferior treatments. We propose a Bayesian single-arm phase II design with continuous monitoring for estimating the response rate of the experimental drug. To address the issue of late-onset responses, we use a piece-wise exponential model to estimate the hazard function of time to response data and handle the missing responses using the multiple imputation approach. We evaluate the operating characteristics of the proposed method through extensive simulation studies. We show that the proposed method reduces the total length of the trial duration and yields desirable operating characteristics for different physician-specified lower bounds of response rate with different true response rates.
Commercial Sexual Exploitation and Missing Children in the Coastal Region of Sao Paulo State, Brazil
Resumo:
The commercial sexual exploitation of children (CSEC) has emerged as one of the world’s most heinous crimes. The problem affects millions of children worldwide and no country or community is fully immune from its effects. This paper reports first generation research of the relationship that exists between CSEC and the phenomenon of missing children living in and around the coastal regions of the state of Sao Paulo, Brazil, the country’s richest State. Data are reported from interviews and case records of 64 children and adolescents, who were receiving care through a major youth serving non-governmental organization (NGO) located in the coastal city of Sao Vicente. Also, data about missing children and adolescents were collected from Police Reports – a total of 858 Police Reports. In Brazil, prostitution is not a crime itself, however, the exploitation of prostitution is a crime. Therefore, the police have no information about children or adolescents in this situation, they only have information about the clients and exploiters. Thus, this investigation sought to accomplish two objectives: 1) to establish the relationship between missing and sexual exploited children; and 2) to sensitize police and child-serving authorities in both the governmental and nongovernmental sectors to the nature, extent, and seriousness of many unrecognized cases of CSEC and missing children that come to their attention. The observed results indicated that the missing children police report are significantly underestimated. They do not represent the number of children that run away and/or are involved in commercial sexual exploitation.