15 resultados para ESTIMATORS
em Aston University Research Archive
Resumo:
Firm-level innovation is investigated using three probit panel estimators, which control for unobserved heterogeneity, and a standard probit estimator. Results indicate the standard probit model is misspecified and that inter-firm networks are important for innovation.
Resumo:
Neural network learning rules can be viewed as statistical estimators. They should be studied in Bayesian framework even if they are not Bayesian estimators. Generalisation should be measured by the divergence between the true distribution and the estimated distribution. Information divergences are invariant measurements of the divergence between two distributions. The posterior average information divergence is used to measure the generalisation ability of a network. The optimal estimators for multinomial distributions with Dirichlet priors are studied in detail. This confirms that the definition is compatible with intuition. The results also show that many commonly used methods can be put under this unified framework, by assume special priors and special divergences.
Resumo:
A family of measurements of generalisation is proposed for estimators of continuous distributions. In particular, they apply to neural network learning rules associated with continuous neural networks. The optimal estimators (learning rules) in this sense are Bayesian decision methods with information divergence as loss function. The Bayesian framework guarantees internal coherence of such measurements, while the information geometric loss function guarantees invariance. The theoretical solution for the optimal estimator is derived by a variational method. It is applied to the family of Gaussian distributions and the implications are discussed. This is one in a series of technical reports on this topic; it generalises the results of ¸iteZhu95:prob.discrete to continuous distributions and serve as a concrete example of a larger picture ¸iteZhu95:generalisation.
Resumo:
Neural networks are statistical models and learning rules are estimators. In this paper a theory for measuring generalisation is developed by combining Bayesian decision theory with information geometry. The performance of an estimator is measured by the information divergence between the true distribution and the estimate, averaged over the Bayesian posterior. This unifies the majority of error measures currently in use. The optimal estimators also reveal some intricate interrelationships among information geometry, Banach spaces and sufficient statistics.
Resumo:
The problem of evaluating different learning rules and other statistical estimators is analysed. A new general theory of statistical inference is developed by combining Bayesian decision theory with information geometry. It is coherent and invariant. For each sample a unique ideal estimate exists and is given by an average over the posterior. An optimal estimate within a model is given by a projection of the ideal estimate. The ideal estimate is a sufficient statistic of the posterior, so practical learning rules are functions of the ideal estimator. If the sole purpose of learning is to extract information from the data, the learning rule must also approximate the ideal estimator. This framework is applicable to both Bayesian and non-Bayesian methods, with arbitrary statistical models, and to supervised, unsupervised and reinforcement learning schemes.
Resumo:
We derive a mean field algorithm for binary classification with Gaussian processes which is based on the TAP approach originally proposed in Statistical Physics of disordered systems. The theory also yields an approximate leave-one-out estimator for the generalization error which is computed with no extra computational cost. We show that from the TAP approach, it is possible to derive both a simpler 'naive' mean field theory and support vector machines (SVM) as limiting cases. For both mean field algorithms and support vectors machines, simulation results for three small benchmark data sets are presented. They show 1. that one may get state of the art performance by using the leave-one-out estimator for model selection and 2. the built-in leave-one-out estimators are extremely precise when compared to the exact leave-one-out estimate. The latter result is a taken as a strong support for the internal consistency of the mean field approach.
Resumo:
This paper examines the relationship between the transfer of ownership between the public and private sectors of Chinese industry, and its impacts on performance. We link ownership changes to productivity growth, and demonstrate that privatisation contributes significantly. We offer an extension that is generally ignored in the literature, in looking at firms that are taken back into state ownership, and evaluating the productivity growth effects of this. Further, we highlight the well-understood simultaneity problems, and demonstrate the hazard of ignoring the issue by comparing various estimators, including the modified control function approach. In general, the results stress the importance of allowing for such endogeneity when evaluating the productivity effects of ownership change.
Resumo:
Automatically generating maps of a measured variable of interest can be problematic. In this work we focus on the monitoring network context where observations are collected and reported by a network of sensors, and are then transformed into interpolated maps for use in decision making. Using traditional geostatistical methods, estimating the covariance structure of data collected in an emergency situation can be difficult. Variogram determination, whether by method-of-moment estimators or by maximum likelihood, is very sensitive to extreme values. Even when a monitoring network is in a routine mode of operation, sensors can sporadically malfunction and report extreme values. If this extreme data destabilises the model, causing the covariance structure of the observed data to be incorrectly estimated, the generated maps will be of little value, and the uncertainty estimates in particular will be misleading. Marchant and Lark [2007] propose a REML estimator for the covariance, which is shown to work on small data sets with a manual selection of the damping parameter in the robust likelihood. We show how this can be extended to allow treatment of large data sets together with an automated approach to all parameter estimation. The projected process kriging framework of Ingram et al. [2007] is extended to allow the use of robust likelihood functions, including the two component Gaussian and the Huber function. We show how our algorithm is further refined to reduce the computational complexity while at the same time minimising any loss of information. To show the benefits of this method, we use data collected from radiation monitoring networks across Europe. We compare our results to those obtained from traditional kriging methodologies and include comparisons with Box-Cox transformations of the data. We discuss the issue of whether to treat or ignore extreme values, making the distinction between the robust methods which ignore outliers and transformation methods which treat them as part of the (transformed) process. Using a case study, based on an extreme radiological events over a large area, we show how radiation data collected from monitoring networks can be analysed automatically and then used to generate reliable maps to inform decision making. We show the limitations of the methods and discuss potential extensions to remedy these.
Resumo:
This paper presents a greedy Bayesian experimental design criterion for heteroscedastic Gaussian process models. The criterion is based on the Fisher information and is optimal in the sense of minimizing parameter uncertainty for likelihood based estimators. We demonstrate the validity of the criterion under different noise regimes and present experimental results from a rabies simulator to demonstrate the effectiveness of the resulting approximately optimal designs.
Resumo:
The detection of signals in the presence of noise is one of the most basic and important problems encountered by communication engineers. Although the literature abounds with analyses of communications in Gaussian noise, relatively little work has appeared dealing with communications in non-Gaussian noise. In this thesis several digital communication systems disturbed by non-Gaussian noise are analysed. The thesis is divided into two main parts. In the first part, a filtered-Poisson impulse noise model is utilized to calulate error probability characteristics of a linear receiver operating in additive impulsive noise. Firstly the effect that non-Gaussian interference has on the performance of a receiver that has been optimized for Gaussian noise is determined. The factors affecting the choice of modulation scheme so as to minimize the deterimental effects of non-Gaussian noise are then discussed. In the second part, a new theoretical model of impulsive noise that fits well with the observed statistics of noise in radio channels below 100 MHz has been developed. This empirical noise model is applied to the detection of known signals in the presence of noise to determine the optimal receiver structure. The performance of such a detector has been assessed and is found to depend on the signal shape, the time-bandwidth product, as well as the signal-to-noise ratio. The optimal signal to minimize the probability of error of; the detector is determined. Attention is then turned to the problem of threshold detection. Detector structure, large sample performance and robustness against errors in the detector parameters are examined. Finally, estimators of such parameters as. the occurrence of an impulse and the parameters in an empirical noise model are developed for the case of an adaptive system with slowly varying conditions.
Resumo:
This article compares the importance of agglomerations of local firms, and inward FDI as drivers of regional development. The empirical analysis exploits a unique panel dataset of the Italian manufacturing sector at the regional and industry levels. We explore whether FDI and firm agglomeration can be drivers of total factor productivity (separately and jointly), with this effect being robust to different estimators, and different assumptions about inter-regional effects. In particular, we isolate one form of firm agglomeration that is especially relevant in the Italian context, industrial districts, in order to ascertain their impact on productivity. In so doing, we distinguish standard agglomeration and localization economies from industrial districts to understand what additional impact the latter has on standard agglomeration effects. Interaction effects between FDI spillovers and different types of agglomeration economies shed some light on the heterogeneity of regional development patterns as well as on the opportunity to fine tune policy measures to specific regional contexts.
Resumo:
This paper presents an empirical study based on a survey of 399 owners of small and medium size companies in Lithuania. Applying bivariate and ordered probit estimators, we investigate why some business owners expect their firms to expand, while others do not. Our main findings provide evidence that SME owner's generic and specific human capital matter. Those with higher education and 'learning by doing' attributes, either through previous job experience or additional entrepreneurial experience, expect their businesses to expand. The expectations of growth are positively related to exporting and non-monotonically to enterprise size. In addition, we analyse the link between the perceptions of constraints to business activities and growth expectations and find that the factors, which are perceived as main business barriers, are not necessary those which are associated with reduced growth expectations. In particular, perceptions of both corruption and of inadequate tax systems seem to affect growth expectations the most.
Resumo:
Estimation of economic relationships often requires imposition of constraints such as positivity or monotonicity on each observation. Methods to impose such constraints, however, vary depending upon the estimation technique employed. We describe a general methodology to impose (observation-specific) constraints for the class of linear regression estimators using a method known as constraint weighted bootstrapping. While this method has received attention in the nonparametric regression literature, we show how it can be applied for both parametric and nonparametric estimators. A benefit of this method is that imposing numerous constraints simultaneously can be performed seamlessly. We apply this method to Norwegian dairy farm data to estimate both unconstrained and constrained parametric and nonparametric models.
Resumo:
Background - The objective of this study was to investigate the association between ethnicity and health related quality of life (HRQoL) in patients with type 2 diabetes. Methods - The EuroQol EQ-5D measure was administered to 1,978 patients with type 2 diabetes in the UK Asian Diabetes Study (UKADS): 1,486 of south Asian origin (Indian, Pakistani, Bangladeshi or other south Asian) and 492 of white European origin. Multivariate regression using ordinary least square (OLS), Tobit, fractional logit and Censored Least Absolutes Deviations estimators was used to estimate the impact of ethnicity on both visual analogue scale (VAS) and utility scores for the EuroQol EQ-5D. Results - Mean EQ-5D VAS and utility scores were lower among south Asians with diabetes compared to the white European population; the unadjusted effect on the mean EQ-5D VAS score was −7.82 (Standard error [SE] = 1.06, p < 0.01) and on the EQ-5D utility score was −0.06 (SE = 0.02, p < 0.01) (OLS estimator). After controlling for socio-demographic and clinical confounders, the adjusted effect on the EQ-5D VAS score was −9.35 (SE = 2.46, p < 0.01) and on the EQ-5D utility score was 0.06 (SE = 0.04), although the latter was not statistically significant. Conclusions - There was a large and statistically significant association between south Asian ethnicity and lower EQ-5D VAS scores. In contrast, there was no significant difference in EQ-5D utility scores between the south Asian and white European sub-groups. Further research is needed to explain the differences in effects on subjective EQ-5D VAS scores and population-weighted EQ-5D utility scores in this context.
Resumo:
We test for departures from normal and independent and identically distributed (NIID) log returns, for log returns under the alternative hypothesis that are self-affine and either long-range dependent, or drawn randomly from an L-stable distribution with infinite higher-order moments. The finite sample performance of estimators of the two forms of self-affinity is explored in a simulation study. In contrast to rescaled range analysis and other conventional estimation methods, the variant of fluctuation analysis that considers finite sample moments only is able to identify both forms of self-affinity. When log returns are self-affine and long-range dependent under the alternative hypothesis, however, rescaled range analysis has higher power than fluctuation analysis. The techniques are illustrated by means of an analysis of the daily log returns for the indices of 11 stock markets of developed countries. Several of the smaller stock markets by capitalization exhibit evidence of long-range dependence in log returns. © 2012 Elsevier Inc. All rights reserved.