972 resultados para hypothesis tests
Resumo:
Background: Research in epistasis or gene-gene interaction detection for human complex traits has grown over the last few years. It has been marked by promising methodological developments, improved translation efforts of statistical epistasis to biological epistasis and attempts to integrate different omics information sources into the epistasis screening to enhance power. The quest for gene-gene interactions poses severe multiple-testing problems. In this context, the maxT algorithm is one technique to control the false-positive rate. However, the memory needed by this algorithm rises linearly with the amount of hypothesis tests. Gene-gene interaction studies will require a memory proportional to the squared number of SNPs. A genome-wide epistasis search would therefore require terabytes of memory. Hence, cache problems are likely to occur, increasing the computation time. In this work we present a new version of maxT, requiring an amount of memory independent from the number of genetic effects to be investigated. This algorithm was implemented in C++ in our epistasis screening software MBMDR-3.0.3. We evaluate the new implementation in terms of memory efficiency and speed using simulated data. The software is illustrated on real-life data for Crohn’s disease. Results: In the case of a binary (affected/unaffected) trait, the parallel workflow of MBMDR-3.0.3 analyzes all gene-gene interactions with a dataset of 100,000 SNPs typed on 1000 individuals within 4 days and 9 hours, using 999 permutations of the trait to assess statistical significance, on a cluster composed of 10 blades, containing each four Quad-Core AMD Opteron(tm) Processor 2352 2.1 GHz. In the case of a continuous trait, a similar run takes 9 days. Our program found 14 SNP-SNP interactions with a multiple-testing corrected p-value of less than 0.05 on real-life Crohn’s disease (CD) data. Conclusions: Our software is the first implementation of the MB-MDR methodology able to solve large-scale SNP-SNP interactions problems within a few days, without using much memory, while adequately controlling the type I error rates. A new implementation to reach genome-wide epistasis screening is under construction. In the context of Crohn’s disease, MBMDR-3.0.3 could identify epistasis involving regions that are well known in the field and could be explained from a biological point of view. This demonstrates the power of our software to find relevant phenotype-genotype higher-order associations.
Resumo:
Euclidean distance matrix analysis (EDMA) methods are used to distinguish whether or not significant difference exists between conformational samples of antibody complementarity determining region (CDR) loops, isolated LI loop and LI in three-loop assembly (LI, L3 and H3) obtained from Monte Carlo simulation. After the significant difference is detected, the specific inter-Ca distance which contributes to the difference is identified using EDMA.The estimated and improved mean forms of the conformational samples of isolated LI loop and LI loop in three-loop assembly, CDR loops of antibody binding site, are described using EDMA and distance geometry (DGEOM). To the best of our knowledge, it is the first time the EDMA methods are used to analyze conformational samples of molecules obtained from Monte Carlo simulations. Therefore, validations of the EDMA methods using both positive control and negative control tests for the conformational samples of isolated LI loop and LI in three-loop assembly must be done. The EDMA-I bootstrap null hypothesis tests showed false positive results for the comparison of six samples of the isolated LI loop and true positive results for comparison of conformational samples of isolated LI loop and LI in three-loop assembly. The bootstrap confidence interval tests revealed true negative results for comparisons of six samples of the isolated LI loop, and false negative results for the conformational comparisons between isolated LI loop and LI in three-loop assembly. Different conformational sample sizes are further explored by combining the samples of isolated LI loop to increase the sample size, or by clustering the sample using self-organizing map (SOM) to narrow the conformational distribution of the samples being comparedmolecular conformations. However, there is no improvement made for both bootstrap null hypothesis and confidence interval tests. These results show that more work is required before EDMA methods can be used reliably as a method for comparison of samples obtained by Monte Carlo simulations.
Resumo:
En este artículo se estudia la posibilidad de introducir seguros de desempleo en Colombia. En una primera parte, se propone una revisión de literatura de los seguros de desempleo en la cual se exponen las ventajas generadas por una cobertura contra este riesgo, así como sus inconvenientes. En una segunda parte, se estudian varios escenarios para introducir seguros de desempleo en Colombia. Después de haber presentado el contexto del mercado laboral y de las normas que lo vigilan, se proponen varios diseños que abordan la gestión y la administración del riesgo de desempleo en Colombia. Igualmente se presentan algunas consideraciones teóricas para la valoración del costo del aseguramiento, las cuales incorporan los efectos del riesgo moral sobre la duración y la incidencia del desempleo.
Resumo:
Many well-established statistical methods in genetics were developed in a climate of severe constraints on computational power. Recent advances in simulation methodology now bring modern, flexible statistical methods within the reach of scientists having access to a desktop workstation. We illustrate the potential advantages now available by considering the problem of assessing departures from Hardy-Weinberg (HW) equilibrium. Several hypothesis tests of HW have been established, as well as a variety of point estimation methods for the parameter which measures departures from HW under the inbreeding model. We propose a computational, Bayesian method for assessing departures from HW, which has a number of important advantages over existing approaches. The method incorporates the effects-of uncertainty about the nuisance parameters--the allele frequencies--as well as the boundary constraints on f (which are functions of the nuisance parameters). Results are naturally presented visually, exploiting the graphics capabilities of modern computer environments to allow straightforward interpretation. Perhaps most importantly, the method is founded on a flexible, likelihood-based modelling framework, which can incorporate the inbreeding model if appropriate, but also allows the assumptions of the model to he investigated and, if necessary, relaxed. Under appropriate conditions, information can be shared across loci and, possibly, across populations, leading to more precise estimation. The advantages of the method are illustrated by application both to simulated data and to data analysed by alternative methods in the recent literature.
Resumo:
In this paper we introduce a parametric model for handling lifetime data where an early lifetime can be related to the infant-mortality failure or to the wear processes but we do not know which risk is responsible for the failure. The maximum likelihood approach and the sampling-based approach are used to get the inferences of interest. Some special cases of the proposed model are studied via Monte Carlo methods for size and power of hypothesis tests. To illustrate the proposed methodology, we introduce an example consisting of a real data set.
Resumo:
O objetivo desta dissertação é identificar se persistência de performance positiva um fenômeno presente entre os fundos de investimento ativos em relação ao índice da bolsa de valores de São Paulo (IBOVESPA). Para tanto, foram utilizadas informações em bases diárias, para período de janeiro de 1997 dezembro de 2001, de 191 fundos que tiveram para cada ano sua performance calculada. Em cada ano sob estudo definiu-se, pelo menos, dois grupos: um em que se alinhavam os fundos que apresentavam os melhores indicadores outro em que estavam presentes aqueles cujos desempenhos foram considerados piores. Os fundos que lograram sucesso em permanecer no primeiro grupo em pares de anos consecutivos foram denominados persistentes. Para que se pudesse medir significância dos resultados apurados, fez-se uso do instrumental de testes estatísticos, testes de hipóteses, que buscaram identificar se as proporções de fundos persistentes forneciam evidências suficientes para que se garantisse ocorrência do fenômeno no mercado brasileiro para período pesquisado.
Resumo:
This master´s thesis presents a reliability study conducted among onshore oil fields in the Potiguar Basin (RN/CE) of Petrobras company, Brazil. The main study objective was to build a regression model to predict the risk of failures that impede production wells to function properly using the information of explanatory variables related to wells such as the elevation method, the amount of water produced in the well (BSW), the ratio gas-oil (RGO), the depth of the production bomb, the operational unit of the oil field, among others. The study was based on a retrospective sample of 603 oil columns from all that were functioning between 2000 and 2006. Statistical hypothesis tests under a Weibull regression model fitted to the failure data allowed the selection of some significant predictors in the set considered to explain the first failure time in the wells
Resumo:
This research aims to investigate the evolution presented during three decades (1980, 1990 and 2000) of using the tools of Operations Research (OR) as a suport to decision making in Production Operation Management (POM). Hypothesis tests were made to verify the proportional growth of a given area over the decades to the detriment of the areas of facility layout, capacity planning, production scheduling and inventory management. Six journals were selected and from them more than 800 articles were used for classification and analysis in the grounds of review. It also discussed possible ways for future research and comparisons are made with other papers of literature review. As a result, it was found that areas of heuristics and simulation showed a greater quantity of contributions in all POM areas of this study
Resumo:
This work purposes the application of a methodology to optimize the implantation cost of an wind-solar hybrid system for oil pumping. The developed model is estimated the implantation cost of system through Multiple Linear Regression technique, on the basis of the previous knowledge of variables: necessary capacity of storage, total daily energy demand, wind power, module power and module number. These variables are gotten by means of sizing. The considered model not only can be applied to the oil pumping, but also for any other purposes of electric energy generation for conversion of solar, wind or solar-wind energy, that demand short powers. Parametric statistical T-student tests had been used to detect the significant difference in the average of total cost to being considered the diameter of the wind, F by Snedecor in the variance analysis to test if the coefficients of the considered model are significantly different of zero and test not-parametric statistical by Friedman, toverify if there is difference in the system cost, by being considered the photovoltaic module powers. In decision of hypothesis tests was considered a 5%-significant level. The configurations module powers showed significant differences in total cost of investment by considering an electrical motor of 3 HP. The configurations module powers showed significant differences in total cost of investment by considering an electrical motor of 5 HP only to wind speed of 4m/s and 6 m/s in wind of 3 m, 4m and 5 m of diameter. There was not significant difference in costs to diameters of winds of 3 m and 4m. The mathematical model and the computational program may be used to others applications which require electrical between 2.250 W and 3.750 W. A computational program was developed to assist the study of several configurations that optimizes the implantation cost of an wind-solar system through considered mathematical model
Resumo:
Survival models deals with the modeling of time to event data. However in some situations part of the population may be no longer subject to the event. Models that take this fact into account are called cure rate models. There are few studies about hypothesis tests in cure rate models. Recently a new test statistic, the gradient statistic, has been proposed. It shares the same asymptotic properties with the classic large sample tests, the likelihood ratio, score and Wald tests. Some simulation studies have been carried out to explore the behavior of the gradient statistic in fi nite samples and compare it with the classic statistics in diff erent models. The main objective of this work is to study and compare the performance of gradient test and likelihood ratio test in cure rate models. We first describe the models and present the main asymptotic properties of the tests. We perform a simulation study based on the promotion time model with Weibull distribution to assess the performance of the tests in finite samples. An application is presented to illustrate the studied concepts
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
This paper presents an approach for structural health monitoring (SHM) by using adaptive filters. The experimental signals from different structural conditions provided by piezoelectric actuators/sensors bonded in the test structure are modeled by a discrete-time recursive least square (RLS) filter. The biggest advantage to use a RLS filter is the clear possibility to perform an online SHM procedure since that the identification is also valid for non-stationary linear systems. An online damage-sensitive index feature is computed based on autoregressive (AR) portion of coefficients normalized by the square root of the sum of the square of them. The proposed method is then utilized in a laboratory test involving an aeronautical panel coupled with piezoelectric sensors/actuators (PZTs) in different positions. A hypothesis test employing the t-test is used to obtain the damage decision. The proposed algorithm was able to identify and localize the damages simulated in the structure. The results have shown the applicability and drawbacks the method and the paper concludes with suggestions to improve it. ©2010 Society for Experimental Mechanics Inc.
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
To design connections between timber parts with metal dowels (nails or bolts), two phenomena must be considered: bending of the metal dowel and the embedment strength in the wood. The Brazilian Standard ABNT NBR 7190:1997 proposes the methods used in the laboratory tests to determine the embedment strength with the metal dowell and, in the absence of testing, specifies relations to calculate the embedment strength from the compression strength. The aim of this research was to compare the embedment strength values obtained by experimental tests and calculated by the relationships established by the Brazilian Standard ABNT NBR 7190:1997. By the results of the hypothesis tests, it was possible to conclude that the estimate of the embedment strength parallel to the grain proposed by the Brazilian standard ABNT NBR 7190:1997, establishing equivalence with the strength compression results in the same direction, must be effective to the wood species Pinus taeda L., however, the same was not found in the perpendicular direction with respect to the grain, possibly explained by the alpha(e) values in the equation to calculed fe(90)
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)