967 resultados para hypothesis testing
Resumo:
This Master Thesis consists of one theoretical article and one empirical article on the field of Microeconometrics. The first chapter\footnote{We also thank useful suggestions by Marinho Bertanha, Gabriel Cepaluni, Brigham Frandsen, Dalia Ghanem, Ricardo Masini, Marcela Mello, Áureo de Paula, Cristine Pinto, Edson Severnini and seminar participants at São Paulo School of Economics, the California Econometrics Conference 2015 and the 37\textsuperscript{th} Brazilian Meeting of Econometrics.}, called \emph{Synthetic Control Estimator: A Generalized Inference Procedure and Confidence Sets}, contributes to the literature about inference techniques of the Synthetic Control Method. This methodology was proposed to answer questions involving counterfactuals when only one treated unit and a few control units are observed. Although this method was applied in many empirical works, the formal theory behind its inference procedure is still an open question. In order to fulfill this lacuna, we make clear the sufficient hypotheses that guarantee the adequacy of Fisher's Exact Hypothesis Testing Procedure for panel data, allowing us to test any \emph{sharp null hypothesis} and, consequently, to propose a new way to estimate Confidence Sets for the Synthetic Control Estimator by inverting a test statistic, the first confidence set when we have access only to finite sample, aggregate level data whose cross-sectional dimension may be larger than its time dimension. Moreover, we analyze the size and the power of the proposed test with a Monte Carlo experiment and find that test statistics that use the synthetic control method outperforms test statistics commonly used in the evaluation literature. We also extend our framework for the cases when we observe more than one outcome of interest (simultaneous hypothesis testing) or more than one treated unit (pooled intervention effect) and when heteroskedasticity is present. The second chapter, called \emph{Free Economic Area of Manaus: An Impact Evaluation using the Synthetic Control Method}, is an empirical article. We apply the synthetic control method for Brazilian city-level data during the 20\textsuperscript{th} Century in order to evaluate the economic impact of the Free Economic Area of Manaus (FEAM). We find that this enterprise zone had positive significant effects on Real GDP per capita and Services Total Production per capita, but it also had negative significant effects on Agriculture Total Production per capita. Our results suggest that this subsidy policy achieve its goal of promoting regional economic growth, even though it may have provoked mis-allocation of resources among economic sectors.
Resumo:
Background: The tectum is a structure localized in the roof of the midbrain in vertebrates, and is taken to be highly conserved in evolution. The present article assessed three hypotheses concerning the evolution of lamination and citoarchitecture of the tectum of nontetrapod animals: 1) There is a significant degree of phylogenetic inertia in both traits studied (number of cellular layers and number of cell classes in tectum); 2) Both traits are positively correlated accross evolution after correction for phylogeny; and 3) Different developmental pathways should generate different patterns of lamination and cytoarchitecture.Methodology/Principal Findings: The hypotheses were tested using analytical-computational tools for phylogenetic hypothesis testing. Both traits presented a considerably large phylogenetic signal and were positively associated. However, no difference was found between two clades classified as per the general developmental pathways of their brains.Conclusions/Significance: The evidence amassed points to more variation in the tectum than would be expected by phylogeny in three species from the taxa analysed; this variation is not better explained by differences in the main course of development, as would be predicted by the developmental clade hypothesis. Those findings shed new light on the evolution of an functionally important structure in nontetrapods, the most basal radiations of vertebrates.
Resumo:
The equations and extrapolation use to localities whose characteristics of soil and climate, even if partial, distinguish the town to which they were generated, still permeate in studies to estimate the rainfall erosivity (EI 30). This work has objective to propose and validate mathematical equations to estimate the rainfall erosivity of two cities of Sao Paulo State's. The adjusted to estimate obtaining and validate data of equations of erosivity (EI 30) according to values of coefficient of rain (Rc) were obtained from pluviographic and pluviometric rainfall data, respectively, using of distinct historical rainfall series. Mutiple comparisions test and confidence intervals were performed to compare absolute average of EI 30, pluviometric data (Pp), and Rc. The correlation between EI 30 and Rc was verified by of Pearson correlation coefficient. Test of the hypothesis of equality between population variance was used to compare the equations. Pluviometrics data of historical series rainfall data different than those that the models were generated were used to validate and to assess the performance of the equations, proposed of this study and compare them with another equation already consolidated in literature. The results show that for the conditions under which the study was conducted, the simple linear equations, shown to be the most appropriate to estimate the rainfall erosivity these two cities. According to the test of the hypothesis of equality variances between populations, the equations adjusted for each city differ statistically so that the rainfall erosivity of each city must be estimated by their respective equation.
Resumo:
Copepod assemblages from two cascade reservoirs were analyzed during two consecutive years. The upstream reservoir (Chavantes) is a storage system with a high water retention time (WRT of 400 days), and the downstream one (Salto Grande) is a run-of-river system with only 1. 5 days WRT. Copepod composition, richness, abundance, and diversity were correlated with the limnological variables and the hydrological and morphometric features. Standard methods were employed for zooplankton sampling and analysis (vertical 50-μm net hauls and counting under a stereomicroscope). Two hypotheses were postulated and confirmed through the data obtained: (1) compartmentalization is more pronounced in the storage reservoir and determines the differences in the copepod assemblage structure; and (2) the assemblages are more homogeneous in the run-of-river reservoir, where the abundance decreases because of the predominance of washout effects. For both reservoirs, the upstream zone is more distinctive. In addition, in the smaller reservoir the influence of the input from tributaries is stronger (turbid waters). Richness did not differ significantly among seasons, but abundance was higher in the run-of-river reservoir during summer. © 2012 Springer Science+Business Media Dordrecht.
Resumo:
Predation is a primary driver of tadpole assemblages, and the activity rate is a good predictor of the tadpoles' tolerance for predation risk. The conflicting demands between activity and exposure to predation can generate suboptimal behaviours. Because morphological components, such as body colouration, may affect the activity of tadpoles, we predict that environmental features that enhance or match the tadpole colouration should affect their survival or activity rate in the presence of a predator. We tested this prediction experimentally by assessing the mortality rate of tadpoles of Rhinella schneideri and Eupemphix nattereri and the active time on two artificial background types: one bright-coloured and one black-coloured. We found no difference in tadpole mortality due to the background type. However, R. schneideri tadpoles were more active than E. nattereri tadpoles, and the activity of R. schneideri was reduced less in the presence of the predator than that of E. nattereri. Although the background colouration did not affect the tadpole mortality rate, it was a stimulus that elicited behavioural responses in the tadpoles, leading them to adjust their activity rate to the type of background colour. © 2013 Dipartimento di Biologia, Università degli Studi di Firenze, Italia.
Resumo:
Background: Arboviral diseases are major global public health threats. Yet, our understanding of infection risk factors is, with a few exceptions, considerably limited. A crucial shortcoming is the widespread use of analytical methods generally not suited for observational data - particularly null hypothesis-testing (NHT) and step-wise regression (SWR). Using Mayaro virus (MAYV) as a case study, here we compare information theory-based multimodel inference (MMI) with conventional analyses for arboviral infection risk factor assessment. Methodology/Principal Findings: A cross-sectional survey of anti-MAYV antibodies revealed 44% prevalence (n = 270 subjects) in a central Amazon rural settlement. NHT suggested that residents of village-like household clusters and those using closed toilet/latrines were at higher risk, while living in non-village-like areas, using bednets, and owning fowl, pigs or dogs were protective. The "minimum adequate" SWR model retained only residence area and bednet use. Using MMI, we identified relevant covariates, quantified their relative importance, and estimated effect-sizes (beta +/- SE) on which to base inference. Residence area (beta(Village) = 2.93 +/- 0.41; beta(Upland) = -0.56 +/- 0.33, beta(Riverbanks) = -2.37 +/- 0.55) and bednet use (beta = -0.95 +/- 0.28) were the most important factors, followed by crop-plot ownership (beta = 0.39 +/- 0.22) and regular use of a closed toilet/latrine (beta = 0.19 +/- 0.13); domestic animals had insignificant protective effects and were relatively unimportant. The SWR model ranked fifth among the 128 models in the final MMI set. Conclusions/Significance: Our analyses illustrate how MMI can enhance inference on infection risk factors when compared with NHT or SWR. MMI indicates that forest crop-plot workers are likely exposed to typical MAYV cycles maintained by diurnal, forest dwelling vectors; however, MAYV might also be circulating in nocturnal, domestic-peridomestic cycles in village-like areas. This suggests either a vector shift (synanthropic mosquitoes vectoring MAYV) or a habitat/habits shift (classical MAYV vectors adapting to densely populated landscapes and nocturnal biting); any such ecological/adaptive novelty could increase the likelihood of MAYV emergence in Amazonia.
Resumo:
The issue of assessing variance components is essential in deciding on the inclusion of random effects in the context of mixed models. In this work we discuss this problem by supposing nonlinear elliptical models for correlated data by using the score-type test proposed in Silvapulle and Silvapulle (1995). Being asymptotically equivalent to the likelihood ratio test and only requiring the estimation under the null hypothesis, this test provides a fairly easy computable alternative for assessing one-sided hypotheses in the context of the marginal model. Taking into account the possible non-normal distribution, we assume that the joint distribution of the response variable and the random effects lies in the elliptical class, which includes light-tailed and heavy-tailed distributions such as Student-t, power exponential, logistic, generalized Student-t, generalized logistic, contaminated normal, and the normal itself, among others. We compare the sensitivity of the score-type test under normal, Student-t and power exponential models for the kinetics data set discussed in Vonesh and Carter (1992) and fitted using the model presented in Russo et al. (2009). Also, a simulation study is performed to analyze the consequences of the kurtosis misspecification.
Resumo:
To estimate causal relationships, time series econometricians must be aware of spurious correlation, a problem first mentioned by Yule (1926). To deal with this problem, one can work either with differenced series or multivariate models: VAR (VEC or VECM) models. These models usually include at least one cointegration relation. Although the Bayesian literature on VAR/VEC is quite advanced, Bauwens et al. (1999) highlighted that "the topic of selecting the cointegrating rank has not yet given very useful and convincing results". The present article applies the Full Bayesian Significance Test (FBST), especially designed to deal with sharp hypotheses, to cointegration rank selection tests in VECM time series models. It shows the FBST implementation using both simulated and available (in the literature) data sets. As illustration, standard non informative priors are used.
Resumo:
In this paper, we present approximate distributions for the ratio of the cumulative wavelet periodograms considering stationary and non-stationary time series generated from independent Gaussian processes. We also adapt an existing procedure to use this statistic and its approximate distribution in order to test if two regularly or irregularly spaced time series are realizations of the same generating process. Simulation studies show good size and power properties for the test statistic. An application with financial microdata illustrates the test usefulness. We conclude advocating the use of these approximate distributions instead of the ones obtained through randomizations, mainly in the case of irregular time series. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
Aim: to evaluate the association of antenatal depressive symptomatology (AD) with life events and coping styles, the hypothesis was that certain coping strategies are associated to depressive symptomatology. Methods: we performed a cross sectional study of 312 women attending a private clinic in the city of Osasco, Sao Paulo from 27/05/1998 to 13/05/2002. The following instruments were used: Beck Depression Inventory (BDI), Holmes and Rahe Schedule of Recent Events (SSRS), Folkman and Lazarus Ways of Coping Questionnaire and questionnaire with social-demographic and obstetric data. Inclusion criteria: women with 110 past history of depression, psychiatric treatment, alcohol or drug abuse and no clinical-obstetrical complications. Odds ratios and 95% CI were used to examine the association between AD (according to BDI) and exposures variables. Hypothesis testing was done with chi(2) tests and a p value < .05. Results: AD occurred in 21.1% of pregnant women. By the univariate analyses, education, number of pregnancies, previous abortion, husband income, situation of marriage and score of SSRS were associated with AD. All coping styles were associated with AD, except seeking support and positive reappraisal. By the multivariate analyses, four coping styles were kept in the final model: confront (p = .039), accepting responsibility (p < .001), escape-avoidance (p = .002), problem-solving (p = .005). Conclusions: AD was highly prevalent and was associated with maladaptive coping styles.
Resumo:
[EN]In this paper the authors show that techniques employed in the prediction of chaotic time series" can also be applied to detection of outliers. A definition of outlier" lS provided and a theorem on hypothesis testing is also proved.
Resumo:
Model-based calibration of steady-state engine operation is commonly performed with highly parameterized empirical models that are accurate but not very robust, particularly when predicting highly nonlinear responses such as diesel smoke emissions. To address this problem, and to boost the accuracy of more robust non-parametric methods to the same level, GT-Power was used to transform the empirical model input space into multiple input spaces that simplified the input-output relationship and improved the accuracy and robustness of smoke predictions made by three commonly used empirical modeling methods: Multivariate Regression, Neural Networks and the k-Nearest Neighbor method. The availability of multiple input spaces allowed the development of two committee techniques: a 'Simple Committee' technique that used averaged predictions from a set of 10 pre-selected input spaces chosen by the training data and the "Minimum Variance Committee" technique where the input spaces for each prediction were chosen on the basis of disagreement between the three modeling methods. This latter technique equalized the performance of the three modeling methods. The successively increasing improvements resulting from the use of a single best transformed input space (Best Combination Technique), Simple Committee Technique and Minimum Variance Committee Technique were verified with hypothesis testing. The transformed input spaces were also shown to improve outlier detection and to improve k-Nearest Neighbor performance when predicting dynamic emissions with steady-state training data. An unexpected finding was that the benefits of input space transformation were unaffected by changes in the hardware or the calibration of the underlying GT-Power model.
Resumo:
Estimation of breastmilk infectivity in HIV-1 infected mothers is difficult because transmission can occur while the fetus is in-utero, during delivery, or through breastfeeding. Since transmission can only be detected through periodic testing, however, it may be impossible to determine the actual mode of transmission in any individual child. In this paper we develop a model to estimate breastmilk infectivity as well as the probabilities of in-utero and intrapartum transmission. In addition, the model allows separate estimation of early and late breastmilk infectivity and individual variation in maternal infectivity. Methods for hypothesis testing of binary risk factors and a method for assessing goodness of fit are also described. Data from a randomized trial of breastfeeding versus formula feeding among HIV-1 infected mothers in Nairobi, Kenya are used to illustrate the methods.
Resumo:
Bioequivalence trials are abbreviated clinical trials whereby a generic drug or new formulation is evaluated to determine if it is "equivalent" to a corresponding previously approved brand-name drug or formulation. In this manuscript, we survey the process of testing bioequivalence and advocate the likelihood paradigm for representing the resulting data as evidence. We emphasize the unique conflicts between hypothesis testing and confidence intervals in this area - which we believe are indicative of the existence of the systemic defects in the frequentist approach - that the likelihood paradigm avoids. We suggest the direct use of profile likelihoods for evaluating bioequivalence and examine the main properties of profile likelihoods and estimated likelihoods under simulation. This simulation study shows that profile likelihoods are a reasonable alternative to the (unknown) true likelihood for a range of parameters commensurate with bioequivalence research. Our study also shows that the standard methods in the current practice of bioequivalence trials offers only weak evidence from the evidential point of view.