995 resultados para Probit models


Relevância:

60.00% 60.00%

Publicador:

Resumo:

The thesis studies the economic and financial conditions of Italian households, by using microeconomic data of the Survey on Household Income and Wealth (SHIW) over the period 1998-2006. It develops along two lines of enquiry. First it studies the determinants of households holdings of assets and liabilities and estimates their correlation degree. After a review of the literature, it estimates two non-linear multivariate models on the interactions between assets and liabilities with repeated cross-sections. Second, it analyses households financial difficulties. It defines a quantitative measure of financial distress and tests, by means of non-linear dynamic probit models, whether the probability of experiencing financial difficulties is persistent over time. Chapter 1 provides a critical review of the theoretical and empirical literature on the estimation of assets and liabilities holdings, on their interactions and on households net wealth. The review stresses the fact that a large part of the literature explain households debt holdings as a function, among others, of net wealth, an assumption that runs into possible endogeneity problems. Chapter 2 defines two non-linear multivariate models to study the interactions between assets and liabilities held by Italian households. Estimation refers to a pooling of cross-sections of SHIW. The first model is a bivariate tobit that estimates factors affecting assets and liabilities and their degree of correlation with results coherent with theoretical expectations. To tackle the presence of non normality and heteroskedasticity in the error term, generating non consistent tobit estimators, semi-parametric estimates are provided that confirm the results of the tobit model. The second model is a quadrivariate probit on three different assets (safe, risky and real) and total liabilities; the results show the expected patterns of interdependence suggested by theoretical considerations. Chapter 3 reviews the methodologies for estimating non-linear dynamic panel data models, drawing attention to the problems to be dealt with to obtain consistent estimators. Specific attention is given to the initial condition problem raised by the inclusion of the lagged dependent variable in the set of explanatory variables. The advantage of using dynamic panel data models lies in the fact that they allow to simultaneously account for true state dependence, via the lagged variable, and unobserved heterogeneity via individual effects specification. Chapter 4 applies the models reviewed in Chapter 3 to analyse financial difficulties of Italian households, by using information on net wealth as provided in the panel component of the SHIW. The aim is to test whether households persistently experience financial difficulties over time. A thorough discussion is provided of the alternative approaches proposed by the literature (subjective/qualitative indicators versus quantitative indexes) to identify households in financial distress. Households in financial difficulties are identified as those holding amounts of net wealth lower than the value corresponding to the first quartile of net wealth distribution. Estimation is conducted via four different methods: the pooled probit model, the random effects probit model with exogenous initial conditions, the Heckman model and the recently developed Wooldridge model. Results obtained from all estimators accept the null hypothesis of true state dependence and show that, according with the literature, less sophisticated models, namely the pooled and exogenous models, over-estimate such persistence.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The dissertation is structured in three parts. The first part compares US and EU agricultural policies since the end of WWII. There is not enough evidence for claiming that agricultural support has a negative impact on obesity trends. I discuss the possibility of an exchange in best practices to fight obesity. There are relevant economic, societal and legal differences between the US and the EU. However, partnerships against obesity are welcomed. The second part presents a socio-ecological model of the determinants of obesity. I employ an interdisciplinary model because it captures the simultaneous influence of several variables. Obesity is an interaction of pre-birth, primary and secondary socialization factors. To test the significance of each factor, I use data from the National Longitudinal Survey of Adolescent Health. I compare the average body mass index across different populations. Differences in means are statistically significant. In the last part I use the National Survey of Children Health. I analyze the effect that family characteristics, built environment, cultural norms and individual factors have on the body mass index (BMI). I use Ordered Probit models and I calculate the marginal effects. I use State and ethnicity fixed effects to control for unobserved heterogeneity. I find that southern US States tend have on average a higher probability of being obese. On the ethnicity side, White Americans have a lower BMI respect to Black Americans, Hispanics and American Indians Native Islanders; being Asian is associated with a lower probability of being obese. In neighborhoods where trust level and safety perception are higher, children are less overweight and obese. Similar results are shown for higher level of parental income and education. Breastfeeding has a negative impact. Higher values of measures of behavioral disorders have a positive and significant impact on obesity, as predicted by the theory.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Objectives. Latinos are the nation's largest minority group and will double in size by 2050. Their size coupled with the fact that Latinos do not constitute a separate race raises questions about Latinos' incorporation into the U. S. racial hierarchy. This article explores patterns of Latino racial identity formation, examining the determinants of racial identity. Methods. Using the 2006 Latino National Survey, I estimate multinomial logit and ordered probit models of identification choices. Results. Latino racial identity is strongly associated with several factors, including socioeconomic status, measures of perceived discrimination and commonality, and measures of acculturation/assimilation. Most Latinos have a broader, more complex understanding of race. Furthermore, some Latinos do believe that they occupy a unique position in the racial hierarchy. Conclusions. The results suggest that the color line W. E. DuBois argued has long divided our nation may eventually shift.

Relevância:

60.00% 60.00%

Publicador:

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This article studies the services exchanged in a particular Spanish time bank. Using data from users and transactions, we analyse the users’ profile as well as the determinants of providing and receiving different services. Our results show that the representative user is a Spanish female, not married, middle aged, highly educated and unemployed. We also find differences in the personal characteristics driving the supply and demand of services.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In this thesis, we explore the relationship between absorptive capacity and alliances, and their influence on firms’ competitive advantage in the US and European biopharmaceutical sectors. The study undertaken in this thesis is based on data from a large-scale international survey of over 2,500 biopharmaceutical firms in the US, the UK, Germany, France and Ireland. The thesis advanced a conceptual framework, which integrated the multi-dimensions of absorptive capacity, exploration-exploitation alliances, and competitive advantage, into a biopharmaceutical firm’s new product development process. The proposed framework is then tested in the empirical analysis, using truncated models to estimate firms’ sales growth, with zero-inflated negative binominal models capturing the number of alliances in which firms engage, and aspects of realised absorptive capacity analysed by ordinal probit models. The empirical results suggest that both skill-based and exploitation-based absorptive capacity play crucial roles in shaping firms’ competitive advantage, while neither exploratory nor exploitation alliances contribute to the improvement in firms’ competitive position. In terms of the interaction between firms’ absorptive capacity and alliance behaviour, the results suggest that engagement with exploratory alliances depends more strongly on firms’ assimilation capability (skills levels and continuity of R&D activities), while participation in exploitation alliances is more conditional on firms’ relevant knowledge monitoring capability. The results highlight the major differences between the determinants of firms’ alliance behaviour, and competitive advantage in the US and Europe – in the US firms’ skill levels prove more significant in determining firms’ engagement with exploratory alliances, whereas in Europe continuity of R&D proves more important. Correspondingly, while US firms’ engagement with exploitation alliances depends on market monitoring capability, that in Europe is more strongly linked to exploitation-based absorptive capacity. In respect of the determinants of firms’ competitive advantage – in Europe, market monitoring capability, engagement with exploitation alliances, and continuous R&D activities, prove more important, while in the US, it is firms’ market characteristics that matter most.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Value creation is the result of the continuous innovation activity of the entrepreneur, which is carried out mainly in form of open innovation among the agri-food SMEs. However value creation is not the ultimate goal of the enterprises. They are more interested in increased appropriation of the created value. Although the value creation (innovation) is very well explored and cultivated area of research, there are some voids in the field of agriculture and food industry: the behavioural aspect of open innovation is very rare. The value capturing is even much less studied, therefor our research approach is largely explorative one. Data are drawn from a survey carried out in Hungary among the agri-food SMEs in 2014. We use Structural Equation Modelling as well as ordered probit and semi-non parametric ordered probit models for analysing the data. Our results show that there is positive relationship between the knowledge sharing with chain partners and the innovativeness. We could explore that size of the firm, absorptive capacity and openness to foreign trade ambiguously affects value capturing. However trust in chain partners, reciprocity in knowledge sharing with chain partners and willingness to cooperate with buyers positively influence the appropriation of the created value.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This dissertation studies newly founded U.S. firms' survival using three different releases of the Kauffman Firm Survey. I study firms' survival from a different perspective in each chapter. ^ The first essay studies firms' survival through an analysis of their initial state at startup and the current state of the firms as they gain maturity. The probability of survival is determined using three probit models, using both firm-specific variables and an industry scale variable to control for the environment of operation. The firm's specific variables include size, experience and leverage as a debt-to-value ratio. The results indicate that size and relevant experience are both positive predictors for the initial and current states. Debt appears to be a predictor of exit if not justified wisely by acquiring assets. As suggested previously in the literature, entering a smaller-scale industry is a positive predictor of survival from birth. Finally, a smaller-scale industry diminishes the negative effects of debt. ^ The second essay makes use of a hazard model to confirm that new service-providing (SP) firms are more likely to survive than new product providers (PPs). I investigate the possible explanations for the higher survival rate of SPs using a Cox proportional hazard model. I examine six hypotheses (variations in capital per worker, expenses per worker, owners' experience, industry wages, assets and size), none of which appear to explain why SPs are more likely than PPs to survive. Two other possibilities are discussed: tax evasion and human/social relations, but these could not be tested due to lack of data. ^ The third essay investigates women-owned firms' higher failure rates using a Cox proportional hazard on two models. I make use of a never-before used variable that proxies for owners' confidence. This variable represents the owners' self-evaluated competitive advantage. ^ The first empirical model allows me to compare women's and men's hazard rates for each variable. In the second model I successively add the variables that could potentially explain why women have a higher failure rate. Unfortunately, I am not able to fully explain the gender effect on the firms' survival. Nonetheless, the second empirical approach allows me to confirm that social and psychological differences among genders are important in explaining the higher likelihood to fail in women-owned firms.^

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Many modern applications fall into the category of "large-scale" statistical problems, in which both the number of observations n and the number of features or parameters p may be large. Many existing methods focus on point estimation, despite the continued relevance of uncertainty quantification in the sciences, where the number of parameters to estimate often exceeds the sample size, despite huge increases in the value of n typically seen in many fields. Thus, the tendency in some areas of industry to dispense with traditional statistical analysis on the basis that "n=all" is of little relevance outside of certain narrow applications. The main result of the Big Data revolution in most fields has instead been to make computation much harder without reducing the importance of uncertainty quantification. Bayesian methods excel at uncertainty quantification, but often scale poorly relative to alternatives. This conflict between the statistical advantages of Bayesian procedures and their substantial computational disadvantages is perhaps the greatest challenge facing modern Bayesian statistics, and is the primary motivation for the work presented here.

Two general strategies for scaling Bayesian inference are considered. The first is the development of methods that lend themselves to faster computation, and the second is design and characterization of computational algorithms that scale better in n or p. In the first instance, the focus is on joint inference outside of the standard problem of multivariate continuous data that has been a major focus of previous theoretical work in this area. In the second area, we pursue strategies for improving the speed of Markov chain Monte Carlo algorithms, and characterizing their performance in large-scale settings. Throughout, the focus is on rigorous theoretical evaluation combined with empirical demonstrations of performance and concordance with the theory.

One topic we consider is modeling the joint distribution of multivariate categorical data, often summarized in a contingency table. Contingency table analysis routinely relies on log-linear models, with latent structure analysis providing a common alternative. Latent structure models lead to a reduced rank tensor factorization of the probability mass function for multivariate categorical data, while log-linear models achieve dimensionality reduction through sparsity. Little is known about the relationship between these notions of dimensionality reduction in the two paradigms. In Chapter 2, we derive several results relating the support of a log-linear model to nonnegative ranks of the associated probability tensor. Motivated by these findings, we propose a new collapsed Tucker class of tensor decompositions, which bridge existing PARAFAC and Tucker decompositions, providing a more flexible framework for parsimoniously characterizing multivariate categorical data. Taking a Bayesian approach to inference, we illustrate empirical advantages of the new decompositions.

Latent class models for the joint distribution of multivariate categorical, such as the PARAFAC decomposition, data play an important role in the analysis of population structure. In this context, the number of latent classes is interpreted as the number of genetically distinct subpopulations of an organism, an important factor in the analysis of evolutionary processes and conservation status. Existing methods focus on point estimates of the number of subpopulations, and lack robust uncertainty quantification. Moreover, whether the number of latent classes in these models is even an identified parameter is an open question. In Chapter 3, we show that when the model is properly specified, the correct number of subpopulations can be recovered almost surely. We then propose an alternative method for estimating the number of latent subpopulations that provides good quantification of uncertainty, and provide a simple procedure for verifying that the proposed method is consistent for the number of subpopulations. The performance of the model in estimating the number of subpopulations and other common population structure inference problems is assessed in simulations and a real data application.

In contingency table analysis, sparse data is frequently encountered for even modest numbers of variables, resulting in non-existence of maximum likelihood estimates. A common solution is to obtain regularized estimates of the parameters of a log-linear model. Bayesian methods provide a coherent approach to regularization, but are often computationally intensive. Conjugate priors ease computational demands, but the conjugate Diaconis--Ylvisaker priors for the parameters of log-linear models do not give rise to closed form credible regions, complicating posterior inference. In Chapter 4 we derive the optimal Gaussian approximation to the posterior for log-linear models with Diaconis--Ylvisaker priors, and provide convergence rate and finite-sample bounds for the Kullback-Leibler divergence between the exact posterior and the optimal Gaussian approximation. We demonstrate empirically in simulations and a real data application that the approximation is highly accurate, even in relatively small samples. The proposed approximation provides a computationally scalable and principled approach to regularized estimation and approximate Bayesian inference for log-linear models.

Another challenging and somewhat non-standard joint modeling problem is inference on tail dependence in stochastic processes. In applications where extreme dependence is of interest, data are almost always time-indexed. Existing methods for inference and modeling in this setting often cluster extreme events or choose window sizes with the goal of preserving temporal information. In Chapter 5, we propose an alternative paradigm for inference on tail dependence in stochastic processes with arbitrary temporal dependence structure in the extremes, based on the idea that the information on strength of tail dependence and the temporal structure in this dependence are both encoded in waiting times between exceedances of high thresholds. We construct a class of time-indexed stochastic processes with tail dependence obtained by endowing the support points in de Haan's spectral representation of max-stable processes with velocities and lifetimes. We extend Smith's model to these max-stable velocity processes and obtain the distribution of waiting times between extreme events at multiple locations. Motivated by this result, a new definition of tail dependence is proposed that is a function of the distribution of waiting times between threshold exceedances, and an inferential framework is constructed for estimating the strength of extremal dependence and quantifying uncertainty in this paradigm. The method is applied to climatological, financial, and electrophysiology data.

The remainder of this thesis focuses on posterior computation by Markov chain Monte Carlo. The Markov Chain Monte Carlo method is the dominant paradigm for posterior computation in Bayesian analysis. It has long been common to control computation time by making approximations to the Markov transition kernel. Comparatively little attention has been paid to convergence and estimation error in these approximating Markov Chains. In Chapter 6, we propose a framework for assessing when to use approximations in MCMC algorithms, and how much error in the transition kernel should be tolerated to obtain optimal estimation performance with respect to a specified loss function and computational budget. The results require only ergodicity of the exact kernel and control of the kernel approximation accuracy. The theoretical framework is applied to approximations based on random subsets of data, low-rank approximations of Gaussian processes, and a novel approximating Markov chain for discrete mixture models.

Data augmentation Gibbs samplers are arguably the most popular class of algorithm for approximately sampling from the posterior distribution for the parameters of generalized linear models. The truncated Normal and Polya-Gamma data augmentation samplers are standard examples for probit and logit links, respectively. Motivated by an important problem in quantitative advertising, in Chapter 7 we consider the application of these algorithms to modeling rare events. We show that when the sample size is large but the observed number of successes is small, these data augmentation samplers mix very slowly, with a spectral gap that converges to zero at a rate at least proportional to the reciprocal of the square root of the sample size up to a log factor. In simulation studies, moderate sample sizes result in high autocorrelations and small effective sample sizes. Similar empirical results are observed for related data augmentation samplers for multinomial logit and probit models. When applied to a real quantitative advertising dataset, the data augmentation samplers mix very poorly. Conversely, Hamiltonian Monte Carlo and a type of independence chain Metropolis algorithm show good mixing on the same dataset.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Purpose: The purpose of this paper is to analyse differences in the drivers of firm innovation performance across sectors. The literature often makes the assumption that firms in different sectors differ in their propensity to innovate but not in the drivers of innovation. The authors empirically assess whether this assumption is accurate through a series of econometric estimations and tests. Design/methodology/approach: The data used are derived from the Irish Community Innovation Survey 2004-2006. A series of multivariate probit models are estimated and the resulting coefficients are tested for parameter stability across sectors using likelihood ratio tests. Findings: The results indicate that there is a strong degree of heterogeneity in the drivers of innovation across sectors. The determinants of process, organisational, new to firm and new to market innovation varies across sectors suggesting that the pooling of sectors in an innovation production function may lead to biased inferences. Research limitations/implications: The implications of the results are that innovation policies targeted at stimulating innovation need to be tailored to particular industries. One size fits all policies would seem inappropriate given the large degree of heterogeneity observed across the drivers of innovation in different sectors. Originality/value: The value of this paper is that it provides an empirical test as to whether it is suitable to group sectoral data when estimating innovation production functions. Most papers simply include sectoral dummies, implying that only the propensity to innovate differs across sectors and that the slope of the coefficient estimates are in fact consistent across sectors.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Mestrado em Economia Monetária e Financeira