14 resultados para Instrumental variable regression

em Duke University


Relevância:

80.00% 80.00%

Publicador:

Resumo:

We investigate the applicability of the present-value asset pricing model to fishing quota markets by applying instrumental variable panel data estimation techniques to 15 years of market transactions from New Zealand's individual transferable quota (ITQ) market. In addition to the influence of current fishing rents, we explore the effect of market interest rates, risk, and expected changes in future rents on quota asset prices. The results indicate that quota asset prices are positively related to declines in interest rates, lower levels of risk, expected increases in future fish prices, and expected cost reductions from rationalization under the quota system. © 2007 American Agricultural Economics Association.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

My dissertation has three chapters which develop and apply microeconometric tech- niques to empirically relevant problems. All the chapters examines the robustness issues (e.g., measurement error and model misspecification) in the econometric anal- ysis. The first chapter studies the identifying power of an instrumental variable in the nonparametric heterogeneous treatment effect framework when a binary treat- ment variable is mismeasured and endogenous. I characterize the sharp identified set for the local average treatment effect under the following two assumptions: (1) the exclusion restriction of an instrument and (2) deterministic monotonicity of the true treatment variable in the instrument. The identification strategy allows for general measurement error. Notably, (i) the measurement error is nonclassical, (ii) it can be endogenous, and (iii) no assumptions are imposed on the marginal distribution of the measurement error, so that I do not need to assume the accuracy of the measure- ment. Based on the partial identification result, I provide a consistent confidence interval for the local average treatment effect with uniformly valid size control. I also show that the identification strategy can incorporate repeated measurements to narrow the identified set, even if the repeated measurements themselves are endoge- nous. Using the the National Longitudinal Study of the High School Class of 1972, I demonstrate that my new methodology can produce nontrivial bounds for the return to college attendance when attendance is mismeasured and endogenous.

The second chapter, which is a part of a coauthored project with Federico Bugni, considers the problem of inference in dynamic discrete choice problems when the structural model is locally misspecified. We consider two popular classes of estimators for dynamic discrete choice models: K-step maximum likelihood estimators (K-ML) and K-step minimum distance estimators (K-MD), where K denotes the number of policy iterations employed in the estimation problem. These estimator classes include popular estimators such as Rust (1987)’s nested fixed point estimator, Hotz and Miller (1993)’s conditional choice probability estimator, Aguirregabiria and Mira (2002)’s nested algorithm estimator, and Pesendorfer and Schmidt-Dengler (2008)’s least squares estimator. We derive and compare the asymptotic distributions of K- ML and K-MD estimators when the model is arbitrarily locally misspecified and we obtain three main results. In the absence of misspecification, Aguirregabiria and Mira (2002) show that all K-ML estimators are asymptotically equivalent regardless of the choice of K. Our first result shows that this finding extends to a locally misspecified model, regardless of the degree of local misspecification. As a second result, we show that an analogous result holds for all K-MD estimators, i.e., all K- MD estimator are asymptotically equivalent regardless of the choice of K. Our third and final result is to compare K-MD and K-ML estimators in terms of asymptotic mean squared error. Under local misspecification, the optimally weighted K-MD estimator depends on the unknown asymptotic bias and is no longer feasible. In turn, feasible K-MD estimators could have an asymptotic mean squared error that is higher or lower than that of the K-ML estimators. To demonstrate the relevance of our asymptotic analysis, we illustrate our findings using in a simulation exercise based on a misspecified version of Rust (1987) bus engine problem.

The last chapter investigates the causal effect of the Omnibus Budget Reconcil- iation Act of 1993, which caused the biggest change to the EITC in its history, on unemployment and labor force participation among single mothers. Unemployment and labor force participation are difficult to define for a few reasons, for example, be- cause of marginally attached workers. Instead of searching for the unique definition for each of these two concepts, this chapter bounds unemployment and labor force participation by observable variables and, as a result, considers various competing definitions of these two concepts simultaneously. This bounding strategy leads to partial identification of the treatment effect. The inference results depend on the construction of the bounds, but they imply positive effect on labor force participa- tion and negligible effect on unemployment. The results imply that the difference- in-difference result based on the BLS definition of unemployment can be misleading

due to misclassification of unemployment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper provides a root-n consistent, asymptotically normal weighted least squares estimator of the coefficients in a truncated regression model. The distribution of the errors is unknown and permits general forms of unknown heteroskedasticity. Also provided is an instrumental variables based two-stage least squares estimator for this model, which can be used when some regressors are endogenous, mismeasured, or otherwise correlated with the errors. A simulation study indicates that the new estimators perform well in finite samples. Our limiting distribution theory includes a new asymptotic trimming result addressing the boundary bias in first-stage density estimation without knowledge of the support boundary. © 2007 Cambridge University Press.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We consider the problem of variable selection in regression modeling in high-dimensional spaces where there is known structure among the covariates. This is an unconventional variable selection problem for two reasons: (1) The dimension of the covariate space is comparable, and often much larger, than the number of subjects in the study, and (2) the covariate space is highly structured, and in some cases it is desirable to incorporate this structural information in to the model building process. We approach this problem through the Bayesian variable selection framework, where we assume that the covariates lie on an undirected graph and formulate an Ising prior on the model space for incorporating structural information. Certain computational and statistical problems arise that are unique to such high-dimensional, structured settings, the most interesting being the phenomenon of phase transitions. We propose theoretical and computational schemes to mitigate these problems. We illustrate our methods on two different graph structures: the linear chain and the regular graph of degree k. Finally, we use our methods to study a specific application in genomics: the modeling of transcription factor binding sites in DNA sequences. © 2010 American Statistical Association.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper studies the multiplicity-correction effect of standard Bayesian variable-selection priors in linear regression. Our first goal is to clarify when, and how, multiplicity correction happens automatically in Bayesian analysis, and to distinguish this correction from the Bayesian Ockham's-razor effect. Our second goal is to contrast empirical-Bayes and fully Bayesian approaches to variable selection through examples, theoretical results and simulations. Considerable differences between the two approaches are found. In particular, we prove a theorem that characterizes a surprising aymptotic discrepancy between fully Bayes and empirical Bayes. This discrepancy arises from a different source than the failure to account for hyperparameter uncertainty in the empirical-Bayes estimate. Indeed, even at the extreme, when the empirical-Bayes estimate converges asymptotically to the true variable-inclusion probability, the potential for a serious difference remains. © Institute of Mathematical Statistics, 2010.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fitting statistical models is computationally challenging when the sample size or the dimension of the dataset is huge. An attractive approach for down-scaling the problem size is to first partition the dataset into subsets and then fit using distributed algorithms. The dataset can be partitioned either horizontally (in the sample space) or vertically (in the feature space), and the challenge arise in defining an algorithm with low communication, theoretical guarantees and excellent practical performance in general settings. For sample space partitioning, I propose a MEdian Selection Subset AGgregation Estimator ({\em message}) algorithm for solving these issues. The algorithm applies feature selection in parallel for each subset using regularized regression or Bayesian variable selection method, calculates the `median' feature inclusion index, estimates coefficients for the selected features in parallel for each subset, and then averages these estimates. The algorithm is simple, involves very minimal communication, scales efficiently in sample size, and has theoretical guarantees. I provide extensive experiments to show excellent performance in feature selection, estimation, prediction, and computation time relative to usual competitors.

While sample space partitioning is useful in handling datasets with large sample size, feature space partitioning is more effective when the data dimension is high. Existing methods for partitioning features, however, are either vulnerable to high correlations or inefficient in reducing the model dimension. In the thesis, I propose a new embarrassingly parallel framework named {\em DECO} for distributed variable selection and parameter estimation. In {\em DECO}, variables are first partitioned and allocated to m distributed workers. The decorrelated subset data within each worker are then fitted via any algorithm designed for high-dimensional problems. We show that by incorporating the decorrelation step, DECO can achieve consistent variable selection and parameter estimation on each subset with (almost) no assumptions. In addition, the convergence rate is nearly minimax optimal for both sparse and weakly sparse models and does NOT depend on the partition number m. Extensive numerical experiments are provided to illustrate the performance of the new framework.

For datasets with both large sample sizes and high dimensionality, I propose a new "divided-and-conquer" framework {\em DEME} (DECO-message) by leveraging both the {\em DECO} and the {\em message} algorithm. The new framework first partitions the dataset in the sample space into row cubes using {\em message} and then partition the feature space of the cubes using {\em DECO}. This procedure is equivalent to partitioning the original data matrix into multiple small blocks, each with a feasible size that can be stored and fitted in a computer in parallel. The results are then synthezied via the {\em DECO} and {\em message} algorithm in a reverse order to produce the final output. The whole framework is extremely scalable.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Consensus HIV-1 genes can decrease the genetic distances between candidate immunogens and field virus strains. To ensure the functionality and optimal presentation of immunologic epitopes, we generated two group-M consensus env genes that contain variable regions either from a wild-type B/C recombinant virus isolate (CON6) or minimal consensus elements (CON-S) in the V1, V2, V4, and V5 regions. C57BL/6 and BALB/c mice were primed twice with CON6, CON-S, and subtype control (92UG37_A and HXB2/Bal_B) DNA and boosted with recombinant vaccinia virus (rVV). Mean antibody titers against 92UG37_A, 89.6_B, 96ZM651_C, CON6, and CON-S Env protein were determined. Both CON6 and CON-S induced higher mean antibody titers against several of the proteins, as compared with the subtype controls. However, no significant differences were found in mean antibody titers in animals immunized with CON6 or CON-S. Cellular immune responses were measured by using five complete Env overlapping peptide sets: subtype A (92UG37_A), subtype B (MN_B, 89.6_B and SF162_B), and subtype C (Chn19_C). The intensity of the induced cellular responses was measured by using pooled Env peptides; T-cell epitopes were identified by using matrix peptide pools and individual peptides. No significant differences in T-cell immune-response intensities were noted between CON6 and CON-S immunized BALB/c and C57BL/6 mice. In BALB/c mice, 10 and eight nonoverlapping T-cell epitopes were identified in CON6 and CON-S, whereas eight epitopes were identified in 92UG37_A and HXB2/BAL_B. In C57BL/6 mice, nine and six nonoverlapping T-cell epitopes were identified after immunization with CON6 and CON-S, respectively, whereas only four and three were identified in 92UG37_A and HXB2/BAL_B, respectively. When combined together from both mouse strains, 18 epitopes were identified. The group M artificial consensus env genes, CON6 and CON-S, were equally immunogenic in breadth and intensity for inducing humoral and cellular immune responses.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We developed a ratiometric method capable of estimating total hemoglobin concentration from optically measured diffuse reflectance spectra. The three isosbestic wavelength ratio pairs that best correlated to total hemoglobin concentration independent of saturation and scattering were 545/390, 452/390, and 529/390 nm. These wavelength pairs were selected using forward Monte Carlo simulations which were used to extract hemoglobin concentration from experimental phantom measurements. Linear regression coefficients from the simulated data were directly applied to the phantom data, by calibrating for instrument throughput using a single phantom. Phantoms with variable scattering and hemoglobin saturation were tested with two different instruments, and the average percent errors between the expected and ratiometrically-extracted hemoglobin concentration were as low as 6.3%. A correlation of r = 0.88 between hemoglobin concentration extracted using the 529/390 nm isosbestic ratio and a scalable inverse Monte Carlo model was achieved for in vivo dysplastic cervical measurements (hemoglobin concentrations have been shown to be diagnostic for the detection of cervical pre-cancer by our group). These results indicate that use of such a simple ratiometric method has the potential to be used in clinical applications where tissue hemoglobin concentrations need to be rapidly quantified in vivo.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Indoor residual spraying (IRS) has become an increasingly popular method of insecticide use for malaria control, and many recent studies have reported on its effectiveness in reducing malaria burden in a single community or region. There is a need for systematic review and integration of the published literature on IRS and the contextual determining factors of its success in controlling malaria. This study reports the findings of a meta-regression analysis based on 13 published studies, which were chosen from more than 400 articles through a systematic search and selection process. The summary relative risk for reducing malaria prevalence was 0.38 (95% confidence interval = 0.31-0.46), which indicated a risk reduction of 62%. However, an excessive degree of heterogeneity was found between the studies. The meta-regression analysis indicates that IRS is more effective with high initial prevalence, multiple rounds of spraying, use of DDT, and in regions with a combination of Plasmodium falciparum and P. vivax malaria.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Patents for several blockbuster biological products are expected to expire soon. The Food and Drug Administration is examining whether biologies can and should be treated like pharmaceuticals with regard to generics. In contrast with pharmaceuticals, which are manufactured through chemical synthesis, biologies are manufactured through fermentation, a process that is more variable and costly. Regulators might require extensive clinical testing of generic biologies to demonstrate equivalence to the branded product. The focus of the debate on generic biologies has been on legal and health concerns, but there are important economic implications. We combine a theoretical model of generic biologies with regression estimates from generic pharmaceuticals to estimate market entry and prices in the generic biologic market. We find that generic biologies will have high fixed costs from clinical testing and from manufacturing, so there will be less entry than would be expected for generic pharmaceuticals. With fewer generic competitors, generic biologies will be relatively close in price to branded biologies. Policy makers should be prudent in estimating financial benefits of generic biologies for consumers and payers. We also examine possible government strategies to promote generic competition. Copyright © 2007 John Wiley & Sons, Ltd.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In regression analysis of counts, a lack of simple and efficient algorithms for posterior computation has made Bayesian approaches appear unattractive and thus underdeveloped. We propose a lognormal and gamma mixed negative binomial (NB) regression model for counts, and present efficient closed-form Bayesian inference; unlike conventional Poisson models, the proposed approach has two free parameters to include two different kinds of random effects, and allows the incorporation of prior information, such as sparsity in the regression coefficients. By placing a gamma distribution prior on the NB dispersion parameter r, and connecting a log-normal distribution prior with the logit of the NB probability parameter p, efficient Gibbs sampling and variational Bayes inference are both developed. The closed-form updates are obtained by exploiting conditional conjugacy via both a compound Poisson representation and a Polya-Gamma distribution based data augmentation approach. The proposed Bayesian inference can be implemented routinely, while being easily generalizable to more complex settings involving multivariate dependence structures. The algorithms are illustrated using real examples. Copyright 2012 by the author(s)/owner(s).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A shearing quotient (SQ) is a way of quantitatively representing the Phase I shearing edges on a molar tooth. Ordinary or phylogenetic least squares regression is fit to data on log molar length (independent variable) and log sum of measured shearing crests (dependent variable). The derived linear equation is used to generate an 'expected' shearing crest length from molar length of included individuals or taxa. Following conversion of all variables to real space, the expected value is subtracted from the observed value for each individual or taxon. The result is then divided by the expected value and multiplied by 100. SQs have long been the metric of choice for assessing dietary adaptations in fossil primates. Not all studies using SQ have used the same tooth position or crests, nor have all computed regression equations using the same approach. Here we focus on re-analyzing the data of one recent study to investigate the magnitude of effects of variation in 1) shearing crest inclusion, and 2) details of the regression setup. We assess the significance of these effects by the degree to which they improve or degrade the association between computed SQs and diet categories. Though altering regression parameters for SQ calculation has a visible effect on plots, numerous iterations of statistical analyses vary surprisingly little in the success of the resulting variables for assigning taxa to dietary preference. This is promising for the comparability of patterns (if not casewise values) in SQ between studies. We suggest that differences in apparent dietary fidelity of recent studies are attributable principally to tooth position examined.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Antigenically variable RNA viruses are significant contributors to the burden of infectious disease worldwide. One reason for their ubiquity is their ability to escape herd immunity through rapid antigenic evolution and thereby to reinfect previously infected hosts. However, the ways in which these viruses evolve antigenically are highly diverse. Some have only limited diversity in the long-run, with every emergence of a new antigenic variant coupled with a replacement of the older variant. Other viruses rapidly accumulate antigenic diversity over time. Others still exhibit dynamics that can be considered evolutionary intermediates between these two extremes. Here, we present a theoretical framework that aims to understand these differences in evolutionary patterns by considering a virus's epidemiological dynamics in a given host population. Our framework, based on a dimensionless number, probabilistically anticipates patterns of viral antigenic diversification and thereby quantifies a virus's evolutionary potential. It is therefore similar in spirit to the basic reproduction number, the well-known dimensionless number which quantifies a pathogen's reproductive potential. We further outline how our theoretical framework can be applied to empirical viral systems, using influenza A/H3N2 as a case study. We end with predictions of our framework and work that remains to be done to further integrate viral evolutionary dynamics with disease ecology.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

BACKGROUND AND PURPOSE: Obesity and physical inactivity are independently associated with physical and functional limitations in older adults. The current study examines the impact of physical activity on odds of physical and functional limitations in older adults with central and general obesity. METHODS: Data from 6279 community-dwelling adults aged 60 years or more from the Health and Retirement Study 2006 and 2008 waves were used to calculate prevalence and odds of physical and functional limitation among obese older adults with high waist circumference (waist circumference ≥88 cm in females and ≥102 cm in males) who were physically active versus inactive (engaging in moderate/vigorous activity less than once per week). Logistic regression models were adjusted for age, sex, race/ethnicity, education, smoking status, body mass index, and number of comorbidities. RESULTS: Physical activity was associated with lower odds of physical and functional limitations among older adults with high waist circumference (odds ratio [OR], 0.59; confidence interval [CI], 0.52-0.68, for physical limitations; OR, 0.52; CI, 0.44-0.62, for activities of daily living; and OR, 0.44; CI, 0.39-0.50, for instrumental activities of daily living). CONCLUSIONS: Physical activity is associated with significantly lower odds of physical and functional limitations in obese older adults regardless of how obesity is classified. Additional research is needed to determine whether physical activity moderates long-term physical and functional limitations.