995 resultados para binomial distribution
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Pós-graduação em Zootecnia - FCAV
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Pós-graduação em Agronomia (Entomologia Agrícola) - FCAV
Resumo:
We investigated the Amblyomma fuscum load on a pullulating wild rodent population and the environmental and biological factors influencing the tick load on the hosts. One hundred and three individuals of Thrichomys laurentius were caught in an Atlantic forest fragment in northeastern Brazil, as part of a longitudinal survey on ticks infesting non-volant small mammals. Ticks (n = 342) were found on 45 individuals and the overall mean intensity of infestation was 7.6 ticks per infested rodent. Ticks were highly aggregated in the host population and the negative binomial distribution model provides a statistically satisfactory fit. The aggregated distribution was influenced by sex and age of the host. The microhabitat preference by T. laurentius probably increases contact opportunities between hosts and aggregated infesting stages of the ticks and represents important clues about the habitat suitability for A. fuscum.
Resumo:
The recent advent of Next-generation sequencing technologies has revolutionized the way of analyzing the genome. This innovation allows to get deeper information at a lower cost and in less time, and provides data that are discrete measurements. One of the most important applications with these data is the differential analysis, that is investigating if one gene exhibit a different expression level in correspondence of two (or more) biological conditions (such as disease states, treatments received and so on). As for the statistical analysis, the final aim will be statistical testing and for modeling these data the Negative Binomial distribution is considered the most adequate one especially because it allows for "over dispersion". However, the estimation of the dispersion parameter is a very delicate issue because few information are usually available for estimating it. Many strategies have been proposed, but they often result in procedures based on plug-in estimates, and in this thesis we show that this discrepancy between the estimation and the testing framework can lead to uncontrolled first-type errors. We propose a mixture model that allows each gene to share information with other genes that exhibit similar variability. Afterwards, three consistent statistical tests are developed for differential expression analysis. We show that the proposed method improves the sensitivity of detecting differentially expressed genes with respect to the common procedures, since it is the best one in reaching the nominal value for the first-type error, while keeping elevate power. The method is finally illustrated on prostate cancer RNA-seq data.
Resumo:
Primates as a taxonomic Order have the largest brains corrected for body size in the animal kingdom. These large brains have allowed primates to evolve the capacity to demonstrate advanced cognitive processes across a wide array of abilities. Nonhuman primates are particularly adept at social learning, defined as the modification of behavior by observing the actions of others. Additionally, primates often exploit resources differently depending on their social context. In this study, capuchin monkeys (Cebus apella) were tested on a cognitive task in three social contexts to determine if social context influenced their performance on the task. The three social contexts included: alone, having a dominant individual in an adjacent compartment, and having a subordinate individual in the adjacent compartment. The benefits to this design were thatthe social context was the only variable influencing performance, whereas in previous studies investigating audience effects other animals could physically and directly influence a subject's performance in an open testing situation. Based on past studies, Ipredicted that the presence of a dominant individual would reduce cognitive task performance compared to the other conditions. The cognitive test used was a match-tosample discrimination task in which animals matched combinations of eight geometric shapes. Animals were trained on this task in an isolated context until they reached a baseline level of proficiency and were then tested in the three social contexts in a random order multiple times. Two subjects (Mt and Dv) have successfully completed trials under all conditions. Results indicated that there were no significant difference in taskperformance across the three conditions (Dv x^2 (1) = 0.42, p=0.58; Mt x^2 (1) = 0.02, p=0.88). In all conditions, subjects performed significantly above chance (i.e., 39/60 trials determined by a binomial distribution). Results are contrary to previous studies thatreport low status monkeys 'play dumb' when testing in a mixed social context, possibly because other studies did not account for aggressive interference by dominants while testing. Results of this study suggest that the mere presence of a dominant individualdoes not necessarily affect performance on a cognitive task, but rather the imminence of physical aggression is the most important factor influencing testing in a social context.
Resumo:
Calcium levels in spines play a significant role in determining the sign and magnitude of synaptic plasticity. The magnitude of calcium influx into spines is highly dependent on influx through N-methyl D-aspartate (NMDA) receptors, and therefore depends on the number of postsynaptic NMDA receptors in each spine. We have calculated previously how the number of postsynaptic NMDA receptors determines the mean and variance of calcium transients in the postsynaptic density, and how this alters the shape of plasticity curves. However, the number of postsynaptic NMDA receptors in the postsynaptic density is not well known. Anatomical methods for estimating the number of NMDA receptors produce estimates that are very different than those produced by physiological techniques. The physiological techniques are based on the statistics of synaptic transmission and it is difficult to experimentally estimate their precision. In this paper we use stochastic simulations in order to test the validity of a physiological estimation technique based on failure analysis. We find that the method is likely to underestimate the number of postsynaptic NMDA receptors, explain the source of the error, and re-derive a more precise estimation technique. We also show that the original failure analysis as well as our improved formulas are not robust to small estimation errors in key parameters.
Resumo:
BACKGROUND Estimating the prevalence of comorbidities and their associated costs in patients with diabetes is fundamental to optimizing health care management. This study assesses the prevalence and health care costs of comorbid conditions among patients with diabetes compared with patients without diabetes. Distinguishing potentially diabetes- and nondiabetes-related comorbidities in patients with diabetes, we also determined the most frequent chronic conditions and estimated their effect on costs across different health care settings in Switzerland. METHODS Using health care claims data from 2011, we calculated the prevalence and average health care costs of comorbidities among patients with and without diabetes in inpatient and outpatient settings. Patients with diabetes and comorbid conditions were identified using pharmacy-based cost groups. Generalized linear models with negative binomial distribution were used to analyze the effect of comorbidities on health care costs. RESULTS A total of 932,612 persons, including 50,751 patients with diabetes, were enrolled. The most frequent potentially diabetes- and nondiabetes-related comorbidities in patients older than 64 years were cardiovascular diseases (91%), rheumatologic conditions (55%), and hyperlipidemia (53%). The mean total health care costs for diabetes patients varied substantially by comorbidity status (US$3,203-$14,223). Patients with diabetes and more than two comorbidities incurred US$10,584 higher total costs than patients without comorbidity. Costs were significantly higher in patients with diabetes and comorbid cardiovascular disease (US$4,788), hyperlipidemia (US$2,163), hyperacidity disorders (US$8,753), and pain (US$8,324) compared with in those without the given disease. CONCLUSION Comorbidities in patients with diabetes are highly prevalent and have substantial consequences for medical expenditures. Interestingly, hyperacidity disorders and pain were the most costly conditions. Our findings highlight the importance of developing strategies that meet the needs of patients with diabetes and comorbidities. Integrated diabetes care such as used in the Chronic Care Model may represent a useful strategy.
Resumo:
Investigation into the medical care utilization of elderly Medicare enrollees in an HMO (Kaiser - Portland, Oregon): The specific research topics are: (1) The utilization of medical care by selected determinants such as: place of service, type of service, type of appointment, physician status, physician specialty and number of associated morbidities. (2) The attended prevalence of 3 chronic diseases: hypertension, diabetes and arthritis in addition to pneumonias as an example of acute diseases. The selection of these examples was based on their importance in morbidity/or mortality results among the elderly. The share of these diseases in outpatient and inpatient contacts was examined as an example of the relation between morbidity and medical care utilization. (3) The tendency of individual utilization patterns to persist in subsequent time periods. The concept of contagion or proneness was studied in a period of 2 years. Fitting the negative binomial and the Poisson distributions was applied to the utilization in the 2nd year conditional on that in the 1st year as regards outpatient and inpatient contacts.^ The present research is based on a longitudinal study of 20% random sample of elderly Medicare enrollees. The sample size is 1683 individuals during the period from August 1980-December 1982.^ The results of the research were: (1) The distribution of contacts by selected determinants did not reveal a consistent pattern between sexes and age groups. (2) The attended prevalence of hypertension and arthritis showed excess prevalence among females. For diabetes and pneumonias no female excess was noticed. Consistent increased prevalence with increasing age was not detected.^ There were important findings pertaining to the relatively big share of the combined 3 chronic diseases in utilization. They accounted for 20% of male outpatient contacts vs. 25% of female outpatients. For inpatient contacts, they consumed 20% in case of males vs. 24% in case of females. (3) Finding that the negative binomial distribution fit the utilization experience supported the research hypothesis concerning the concept of contagion in utilization. This important finding can be helpful in estimating liability functions needed for forecasting future utilization according to previous experience. Such information has its relevance to organization, administration and planning for medical care in general. (Abstract shortened with permission of author.) ^
Resumo:
My dissertation focuses on two aspects of RNA sequencing technology. The first is the methodology for modeling the overdispersion inherent in RNA-seq data for differential expression analysis. This aspect is addressed in three sections. The second aspect is the application of RNA-seq data to identify the CpG island methylator phenotype (CIMP) by integrating datasets of mRNA expression level and DNA methylation status. Section 1: The cost of DNA sequencing has reduced dramatically in the past decade. Consequently, genomic research increasingly depends on sequencing technology. However it remains elusive how the sequencing capacity influences the accuracy of mRNA expression measurement. We observe that accuracy improves along with the increasing sequencing depth. To model the overdispersion, we use the beta-binomial distribution with a new parameter indicating the dependency between overdispersion and sequencing depth. Our modified beta-binomial model performs better than the binomial or the pure beta-binomial model with a lower false discovery rate. Section 2: Although a number of methods have been proposed in order to accurately analyze differential RNA expression on the gene level, modeling on the base pair level is required. Here, we find that the overdispersion rate decreases as the sequencing depth increases on the base pair level. Also, we propose four models and compare them with each other. As expected, our beta binomial model with a dynamic overdispersion rate is shown to be superior. Section 3: We investigate biases in RNA-seq by exploring the measurement of the external control, spike-in RNA. This study is based on two datasets with spike-in controls obtained from a recent study. We observe an undiscovered bias in the measurement of the spike-in transcripts that arises from the influence of the sample transcripts in RNA-seq. Also, we find that this influence is related to the local sequence of the random hexamer that is used in priming. We suggest a model of the inequality between samples and to correct this type of bias. Section 4: The expression of a gene can be turned off when its promoter is highly methylated. Several studies have reported that a clear threshold effect exists in gene silencing that is mediated by DNA methylation. It is reasonable to assume the thresholds are specific for each gene. It is also intriguing to investigate genes that are largely controlled by DNA methylation. These genes are called “L-shaped” genes. We develop a method to determine the DNA methylation threshold and identify a new CIMP of BRCA. In conclusion, we provide a detailed understanding of the relationship between the overdispersion rate and sequencing depth. And we reveal a new bias in RNA-seq and provide a detailed understanding of the relationship between this new bias and the local sequence. Also we develop a powerful method to dichotomize methylation status and consequently we identify a new CIMP of breast cancer with a distinct classification of molecular characteristics and clinical features.
Resumo:
Conservative procedures in low-dose risk assessment are used to set safety standards for known or suspected carcinogens. However, the assumptions upon which the methods are based and the effects of these methods are not well understood.^ To minimize the number of false-negatives and to reduce the cost of bioassays, animals are given very high doses of potential carcinogens. Results must then be extrapolated to much smaller doses to set safety standards for risks such as one per million. There are a number of competing methods that add a conservative safety factor into these calculations.^ A method of quantifying the conservatism of these methods was described and tested on eight procedures used in setting low-dose safety standards. The results using these procedures were compared by computer simulation and by the use of data from a large scale animal study.^ The method consisted of determining a "true safe dose" (tsd) according to an assumed underlying model. If one assumed that Y = the probability of cancer = P(d), a known mathematical function of the dose, then by setting Y to some predetermined acceptable risk, one can solve for d, the model's "true safe dose".^ Simulations were generated, assuming a binomial distribution, for an artificial bioassay. The eight procedures were then used to determine a "virtual safe dose" (vsd) that estimates the tsd, assuming a risk of one per million. A ratio R = ((tsd-vsd)/vsd) was calculated for each "experiment" (simulation). The mean R of 500 simulations and the probability R $<$ 0 was used to measure the over and under conservatism of each procedure.^ The eight procedures included Weil's method, Hoel's method, the Mantel-Byran method, the improved Mantel-Byran, Gross's method, fitting a one-hit model, Crump's procedure, and applying Rai and Van Ryzin's method to a Weibull model.^ None of the procedures performed uniformly well for all types of dose-response curves. When the data were linear, the one-hit model, Hoel's method, or the Gross-Mantel method worked reasonably well. However, when the data were non-linear, these same methods were overly conservative. Crump's procedure and the Weibull model performed better in these situations. ^