979 resultados para Methods : Statistical
Resumo:
Aims. We studied four young star clusters to characterise their anomalous extinction or variable reddening and asses whether they could be due to contamination by either dense clouds or circumstellar effects. Methods. We evaluated the extinction law (R-V) by adopting two methods: (i) the use of theoretical expressions based on the colour-excess of stars with known spectral type; and (ii) the analysis of two-colour diagrams, where the slope of the observed colour distribution was compared to the normal distribution. An algorithm to reproduce the zero-age main-sequence (ZAMS) reddened colours was developed to derive the average visual extinction (A(V)) that provides the closest fit to the observational data. The structure of the clouds was evaluated by means of a statistical fractal analysis, designed to compare their geometric structure with the spatial distribution of the cluster members. Results. The cluster NGC 6530 is the only object of our sample affected by anomalous extinction. On average, the other clusters suffer normal extinction, but several of their members, mainly in NGC 2264, seem to have high R-V, probably because of circumstellar effects. The ZAMS fitting provides A(V) values that are in good agreement with those found in the literature. The fractal analysis shows that NGC 6530 has a centrally concentrated distribution of stars that differs from the substructures found in the density distribution of the cloud projected in the A(V) map, suggesting that the original cloud was changed by the cluster formation. However, the fractal dimension and statistical parameters of Berkeley 86, NGC 2244, and NGC 2264 indicate that there is a good cloud-cluster correlation, when compared to other works based on an artificial distribution of points.
Resumo:
Context. Convergent point (CP) search methods are important tools for studying the kinematic properties of open clusters and young associations whose members share the same spatial motion. Aims. We present a new CP search strategy based on proper motion data. We test the new algorithm on synthetic data and compare it with previous versions of the CP search method. As an illustration and validation of the new method we also present an application to the Hyades open cluster and a comparison with independent results. Methods. The new algorithm rests on the idea of representing the stellar proper motions by great circles over the celestial sphere and visualizing their intersections as the CP of the moving group. The new strategy combines a maximum-likelihood analysis for simultaneously determining the CP and selecting the most likely group members and a minimization procedure that returns a refined CP position and its uncertainties. The method allows one to correct for internal motions within the group and takes into account that the stars in the group lie at different distances. Results. Based on Monte Carlo simulations, we find that the new CP search method in many cases returns a more precise solution than its previous versions. The new method is able to find and eliminate more field stars in the sample and is not biased towards distant stars. The CP solution for the Hyades open cluster is in excellent agreement with previous determinations.
Resumo:
We have completed a high-contrast direct imaging survey for giant planets around 57 debris disk stars as part of the Gemini NICI Planet-Finding Campaign. We achieved median H-band contrasts of 12.4 mag at 0.''5 and 14.1 mag at 1'' separation. Follow-up observations of the 66 candidates with projected separation <500 AU show that all of them are background objects. To establish statistical constraints on the underlying giant planet population based on our imaging data, we have developed a new Bayesian formalism that incorporates (1) non-detections, (2) single-epoch candidates, (3) astrometric and (4) photometric information, and (5) the possibility of multiple planets per star to constrain the planet population. Our formalism allows us to include in our analysis the previously known β Pictoris and the HR 8799 planets. Our results show at 95% confidence that <13% of debris disk stars have a ≥5 M Jup planet beyond 80 AU, and <21% of debris disk stars have a ≥3 M Jup planet outside of 40 AU, based on hot-start evolutionary models. We model the population of directly imaged planets as d 2 N/dMdavpropm α a β, where m is planet mass and a is orbital semi-major axis (with a maximum value of a max). We find that β < –0.8 and/or α > 1.7. Likewise, we find that β < –0.8 and/or a max < 200 AU. For the case where the planet frequency rises sharply with mass (α > 1.7), this occurs because all the planets detected to date have masses above 5 M Jup, but planets of lower mass could easily have been detected by our search. If we ignore the β Pic and HR 8799 planets (should they belong to a rare and distinct group), we find that <20% of debris disk stars have a ≥3 M Jup planet beyond 10 AU, and β < –0.8 and/or α < –1.5. Likewise, β < –0.8 and/or a max < 125 AU. Our Bayesian constraints are not strong enough to reveal any dependence of the planet frequency on stellar host mass. Studies of transition disks have suggested that about 20% of stars are undergoing planet formation; our non-detections at large separations show that planets with orbital separation >40 AU and planet masses >3 M Jup do not carve the central holes in these disks.
Resumo:
OBJECTIVES: The aim of the study was to assess whether prospective follow-up data within the Swiss HIV Cohort Study can be used to predict patients who stop smoking; or among smokers who stop, those who start smoking again. METHODS: We built prediction models first using clinical reasoning ('clinical models') and then by selecting from numerous candidate predictors using advanced statistical methods ('statistical models'). Our clinical models were based on literature that suggests that motivation drives smoking cessation, while dependence drives relapse in those attempting to stop. Our statistical models were based on automatic variable selection using additive logistic regression with component-wise gradient boosting. RESULTS: Of 4833 smokers, 26% stopped smoking, at least temporarily; because among those who stopped, 48% started smoking again. The predictive performance of our clinical and statistical models was modest. A basic clinical model for cessation, with patients classified into three motivational groups, was nearly as discriminatory as a constrained statistical model with just the most important predictors (the ratio of nonsmoking visits to total visits, alcohol or drug dependence, psychiatric comorbidities, recent hospitalization and age). A basic clinical model for relapse, based on the maximum number of cigarettes per day prior to stopping, was not as discriminatory as a constrained statistical model with just the ratio of nonsmoking visits to total visits. CONCLUSIONS: Predicting smoking cessation and relapse is difficult, so that simple models are nearly as discriminatory as complex ones. Patients with a history of attempting to stop and those known to have stopped recently are the best candidates for an intervention.
Resumo:
Objectives. The purpose of this study was to elucidate behavioral determinants (prevailing attitudes and beliefs) of hand hygiene practices among undergraduate dental students in a dental school. ^ Methods. Statistical modeling using the Integrative Behavioral Model (IBM) prediction was utilized to develop a questionnaire for evaluating behavioral perceptions of hand hygiene practices by dental school students. Self-report questionnaires were given to second, third and fourth year undergraduate dental students. Models representing two distinct hand hygiene practices, termed "elective in-dental school hand hygiene practice" and "inherent in-dental school hand hygiene practice" were tested using linear regression analysis. ^ Results. 58 responses were received (24.5%); the sample mean age was 26.6 years old and females comprised 51%. In our models, elective in-dental school hand hygiene practice and inherent in-dental school hand hygiene practice, explained 40% and 28%, respectively, of the variance in behavioral intention. Translation of community hand hygiene practice to the dental school setting is the predominant driver of elective hand hygiene practice. Intended elective in-school hand hygiene practice is further significantly predicted by students' self-efficacy. Students' attitudes, peer pressure of other dental students and clinic administrators, and role modeling had minimal effects. Inherent hand hygiene intent was strongly predicted by students' beliefs in the benefits of the activity and, to a lesser extent, role modeling. Inherent and elective community behaviors were insignificant. ^ Conclusions. This study provided significant insights into dental student's hand hygiene behavior and can form the basis for an effective behavioral intervention program designed to improve hand hygiene compliance.^
Resumo:
The H I Parkes All Sky Survey (HIPASS) is a blind extragalactic H I 21-cm emission-line survey covering the whole southern sky from declination -90degrees to +25degrees. The HIPASS catalogue (HICAT), containing 4315 H I-selected galaxies from the region south of declination +2degrees, is presented in Meyer et al. (Paper I). This paper describes in detail the completeness and reliability of HICAT, which are calculated from the recovery rate of synthetic sources and follow-up observations, respectively. HICAT is found to be 99 per cent complete at a peak flux of 84 mJy and an integrated flux of 9.4 Jy km. s(-1). The overall reliability is 95 per cent, but rises to 99 per cent for sources with peak fluxes >58 mJy or integrated flux >8.2 Jy km s(-1). Expressions are derived for the uncertainties on the most important HICAT parameters: peak flux, integrated flux, velocity width and recessional velocity. The errors on HICAT parameters are dominated by the noise in the HIPASS data, rather than by the parametrization procedure.
Resumo:
An important and common problem in microarray experiments is the detection of genes that are differentially expressed in a given number of classes. As this problem concerns the selection of significant genes from a large pool of candidate genes, it needs to be carried out within the framework of multiple hypothesis testing. In this paper, we focus on the use of mixture models to handle the multiplicity issue. With this approach, a measure of the local false discovery rate is provided for each gene, and it can be implemented so that the implied global false discovery rate is bounded as with the Benjamini-Hochberg methodology based on tail areas. The latter procedure is too conservative, unless it is modified according to the prior probability that a gene is not differentially expressed. An attractive feature of the mixture model approach is that it provides a framework for the estimation of this probability and its subsequent use in forming a decision rule. The rule can also be formed to take the false negative rate into account.
Resumo:
We present a new algorithm for detecting intercluster galaxy filaments based upon the assumption that the orientations of constituent galaxies along such filaments are non-isotropic. We apply the algorithm to the 2dF Galaxy Redshift Survey catalogue and find that it readily detects many straight filaments between close cluster pairs. At large intercluster separations (> 15 h(-1) Mpc), we find that the detection efficiency falls quickly, as it also does with more complex filament morphologies. We explore the underlying assumptions and suggest that it is only in the case of close cluster pairs that we can expect galaxy orientations to be significantly correlated with filament direction.
Resumo:
We consider the statistical problem of catalogue matching from a machine learning perspective with the goal of producing probabilistic outputs, and using all available information. A framework is provided that unifies two existing approaches to producing probabilistic outputs in the literature, one based on combining distribution estimates and the other based on combining probabilistic classifiers. We apply both of these to the problem of matching the HI Parkes All Sky Survey radio catalogue with large positional uncertainties to the much denser SuperCOSMOS catalogue with much smaller positional uncertainties. We demonstrate the utility of probabilistic outputs by a controllable completeness and efficiency trade-off and by identifying objects that have high probability of being rare. Finally, possible biasing effects in the output of these classifiers are also highlighted and discussed.
Resumo:
Fine-fraction (<63 µm) grain-size analyses of 530 samples from Holes 1095A, 1095B, and 1095D allow assessment of the downhole grain-size distribution at Drift 7. A variety of data processing methods, statistical treatment, and display techniques were used to describe this data set. The downhole fine-fraction grain-size distribution documents significant variations in the average grain-size composition and its cyclic pattern, revealed in five prominent intervals: (1) between 0 and 40 meters composite depth (mcd) (0 and 1.3 Ma), (2) between 40 and 80 mcd (1.3 and 2.4 Ma), (3) between 80 and 220 mcd (2.4 and 6 Ma), (4) between 220 and 360 mcd, and (5) below 360 mcd (prior to 8.1 Ma). In an approach designed to characterize depositional processes at Drift 7, we used statistical parameters determined by the method of moments for the sortable silt fraction to distinguish groups in the grainsize data set. We found three distinct grain-size populations and used these for a tentative environmental interpretation. Population 1 is related to a process in which glacially eroded shelf material was redeposited by turbidites with an ice-rafted debris influence. Population 2 is composed of interglacial turbidites. Population 3 is connected to depositional sequence tops linked to bioturbated sections that, in turn, are influenced by contourite currents and pelagic background sedimentation.
Resumo:
Introdução: A perda transitória da consciência e tónus postural seguido de rápida recuperação é definida como síncope. Tem sido dada atenção a uma síncope de origem central com descida da pressão sistémica conhecida por síncope vasovagal (SVV). Objetivos: A análise da variabilidade da frequência cardíaca (HRV) é uma das principais estratégias para estudar a SVV através de protocolos padrão (por exemplo tilt test). O principal objetivo deste trabalho é compreender a importância relativa de diversas variáveis, tais como pressão arterial diastólica e sistólica, (dBP) e (sBP), volume sistólico (SV) e resistência periférica total (TPR) na HRV. Métodos: Foram usados modelos estatísticos mistos para modelar o comportamento das variáveis acima descritas na HRV. Analisaram-se mais de mil e quinhentas observações de quatro pacientes com SVV, previamente testados com análise espectral clássica para a fase basal (LF/HF=3.01) e fases de tilt (LF/HF=0.64), indicando uma predominância vagal no período tilt. Resultados: O modelo 1 revelou o papel importante da dBP e uma baixa influência de SV, na fase de tilt, relativos à HRV. No modelo 2 a TPR revelou uma baixa influência na HRV na fase de tilt entre os pacientes. Conclusões: Verificou-se que a HRV é influenciada por um conjunto de variáveis fisiológicas, cuja contribuição individual pode ser usada para compreender as flutuações cardíacas. O uso de modelos estatísticos salientou a importância de estudar o papel da dBP e SV na SVV.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia do Ambiente
Resumo:
Dissertação apresentada como requisito parcial para obtenção do grau de Mestre em Estatística e Gestão de Informação
Resumo:
This paper presents and discusses the use of Bayesian procedures - introduced through the use of Bayesian networks in Part I of this series of papers - for 'learning' probabilities from data. The discussion will relate to a set of real data on characteristics of black toners commonly used in printing and copying devices. Particular attention is drawn to the incorporation of the proposed procedures as an integral part in probabilistic inference schemes (notably in the form of Bayesian networks) that are intended to address uncertainties related to particular propositions of interest (e.g., whether or not a sample originates from a particular source). The conceptual tenets of the proposed methodologies are presented along with aspects of their practical implementation using currently available Bayesian network software.