970 resultados para mean-variance estimation
Resumo:
Cette thèse examine le rôle du pouvoir de marché dans le marché bancaire. L’emphase est mis sur la prise de risque, les économies d’échelle, l’efficacité économique du marché et la transmission des chocs. Le premier chapitre présente un modèle d’équilibre général dynamique stochastique en économie ouverte comprenant un marché bancaire en concurrence monopolistique. Suivant l’hypothèse de Krugman (1979, 1980) sur la relation entre les économies d’échelle et les exportations, les banques doivent défrayer un coût de transaction pour échanger à l’étranger qui diminue à mesure que le volume de leurs activités locales augmente. Cela incite les banques à réduire leur marge locale afin de profiter davantage du marché extérieur. Le modèle est solutionné et simulé pour divers degrés de concentration dans le marché bancaire. Les résultats obtenus indiquent que deux forces contraires, les économies d’échelle et le pouvoir de marché, s’affrontent lorsque le marché se concentre. La concentration permet aussi aux banques d’accroître leurs activités étrangères, ce qui les rend en contrepartie plus vulnérables aux chocs extérieurs. Le deuxième chapitre élabore un cadre de travail semblable, mais à l’intérieur duquel les banques font face à un risque de crédit. Celui-ci est partiellement assuré par un collatéral fourni par les entrepreneurs et peut être limité à l’aide d’un effort financier. Le modèle est solutionné et simulé pour divers degrés de concentration dans le marché bancaire. Les résultats montrent qu’un plus grand pouvoir de marché réduit la taille du marché financier et de la production à l’état stationnaire, mais incite les banques à prendre moins de risques. De plus, les économies dont le marché bancaire est fortement concentré sont moins sensibles à certains chocs puisque les marges plus élevés donnent initialement de la marge de manoeuvre aux banques en cas de chocs négatifs. Cet effet modérateur est éliminé lorsqu’il est possible pour les banques d’entrer et de sortir librement du marché. Une autre extension avec économies d’échelle montre que sous certaines conditions, un marché moyennement concentré est optimal pour l’économie. Le troisième chapitre utilise un modèle en analyse de portefeuille de type Moyenne-Variance afin de représenter une banque détenant du pouvoir de marché. Le rendement des dépôts et des actifs peut varier selon la quantité échangée, ce qui modifie le choix de portefeuille de la banque. Celle-ci tend à choisir un portefeuille dont la variance est plus faible lorsqu’elle est en mesure d’obtenir un rendement plus élevé sur un actif. Le pouvoir de marché sur les dépôts amène un résultat sembable pour un pouvoir de marché modéré, mais la variance finit par augmenter une fois un certain niveau atteint. Les résultats sont robustes pour différentes fonctions de demandes.
Resumo:
In this study cross-section data was used to analyze the effect of farmers’ demographic, socioeconomic and institutional setting, market access and physical attributes on the probability and intensity of tissue culture banana (TCB) adoption. The study was carried out between July 2011 and November 2011. Both descriptive (mean, variance, promotions) and regression analysis were used in the analysis. A double hurdle regression model was fitted on the data. Using multistage sampling technique, four counties and eight sub-locations were randomly selected. Using random sampling technique, three hundred and thirty farmers were selected from a list of banana households in the selected sub-locations. The adoption level of tissue culture banana (TCB) was about 32%. The results also revealed that the likelihood of TCB adoption was significantly influenced by: availability of TCB planting material, proportion of banana income to the total farm income, per capita household expenditure and the location of the farmer in Kisii County; while those that significantly influenced the intensity of TCB adoption were: occupation of farmers, family size, labour source, farm size, soil fertility, availability/access of TCB plantlets to farmers, distance to banana market, use of manure in planting banana, access to agricultural extension services and index of TCB/non-TCB banana cultivar attributes which were scored by farmers. Compared to West Pokot County, farmers located in Bungoma County are more significantly and likely to adopt TCB technology. Therefore, the results of the study suggest that the probability of adoption and intensity of the use of TCB should be enhanced. This can be done by taking cognizance of these variables in order to meet the priority needs of the smallholder farmers who were the target group. This would lead to alleviating banana shortage in the region for enhanced food security. Subsequently, actors along the banana value chain are encouraged to target the intervention strategies based on the identified farmer, farm and institutional characteristics for enhanced impact on food provision. Opening up more TCB multiplication centres in different regions will make farmers access the TCB technology for enhanced impact on the target population.
Resumo:
Mestrado em Finanças
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Faculdade de Tecnologia, Departamento de Engenharia Civil e Ambiental, 2016.
Resumo:
There has been a resurgence of interest in the mean trace length estimator of Pahl for window sampling of traces. The estimator has been dealt with by Mauldon and Zhang and Einstein in recent publications. The estimator is a very useful one in that it is non-parametric. However, despite some discussion regarding the statistical distribution of the estimator, none of the recent works or the original work by Pahl provide a rigorous basis for the determination a confidence interval for the estimator or a confidence region for the estimator and the corresponding estimator of trace spatial intensity in the sampling window. This paper shows, by consideration of a simplified version of the problem but without loss of generality, that the estimator is in fact the maximum likelihood estimator (MLE) and that it can be considered essentially unbiased. As the MLE, it possesses the least variance of all estimators and confidence intervals or regions should therefore be available through application of classical ML theory. It is shown that valid confidence intervals can in fact be determined. The results of the work and the calculations of the confidence intervals are illustrated by example. (C) 2003 Elsevier Science Ltd. All rights reserved.
Resumo:
We discuss the estimation of the expected value of the quality-adjusted survival, based on multistate models. We generalize an earlier work, considering the sojourn times in health states are not identically distributed, for a given vector of covariates. Approaches based on semiparametric and parametric (exponential and Weibull distributions) methodologies are considered. A simulation study is conducted to evaluate the performance of the proposed estimator and the jackknife resampling method is used to estimate the variance of such estimator. An application to a real data set is also included.
Resumo:
In clinical trials, it may be of interest taking into account physical and emotional well-being in addition to survival when comparing treatments. Quality-adjusted survival time has the advantage of incorporating information about both survival time and quality-of-life. In this paper, we discuss the estimation of the expected value of the quality-adjusted survival, based on multistate models for the sojourn times in health states. Semiparametric and parametric (with exponential distribution) approaches are considered. A simulation study is presented to evaluate the performance of the proposed estimator and the jackknife resampling method is used to compute bias and variance of the estimator. (C) 2007 Elsevier B.V. All rights reserved.
Resumo:
Com este trabalho objetivou-se determinar parâmetros genéticos para peso corporal de perdizes em cativeiro. Foram utilizados modelos de regressão aleatória na análise dos dados considerando os efeitos genéticos aditivos diretos (AD) e de ambiente permanente de animal (AP) como aleatórios. As variâncias residuais foram modeladas utilizando-se funções de variância de ordem 5. A curva média da população foi ajustada por polinômios ortogonais de Legendre de ordem 6. Os efeitos genéticos aditivos diretos e de ambiente permanente de animal foram modelados utilizando-se polinômios de Legendre de segunda a nona ordem. Os melhores resultados foram obtidos pelos modelos de ordem 6 de ajuste para os efeitos genéticos aditivos diretos e de ordem 3 para os de ambiente permanente pelo Critério de Informação de Akaike e ordem 3 para ambos os efeitos pelos Critério de Informação Bayesiano de Schwartz e Teste de Razão de Verossimilhança. As herdabilidades estimadas variaram de 0,02 a 0,57. O primeiro autovalor respondeu por 94 e 90% da variação decorrente de efeitos aditivos diretos e de ambiente permanente, respectivamente. A seleção de perdizes para peso corporal é mais efetiva a partir de 112 dias de idade.
Resumo:
Asymmetric discrete triangular distributions are introduced in order to extend the symmetric ones serving for discrete associated kernels in the nonparametric estimation for discrete functions. The extension from one to two orders around the mode provides a large family of discrete distributions having a finite support. Establishing a bridge between Dirac and discrete uniform distributions, some different shapes are also obtained and their properties are investigated. In particular, the mean and variance are pointed out. Applications to discrete kernel estimators are given with a solution to a boundary bias problem. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
Background From the mid-1980s to mid-1990s, the WHO MONICA Project monitored coronary events and classic risk factors for coronary heart disease (CHD) in 38 populations from 21 countries. We assessed the extent to which changes in these risk factors explain the variation in the trends in coronary-event rates across the populations. Methods In men and women aged 35-64 years, non-fatal myocardial infarction and coronary deaths were registered continuously to assess trends in rates of coronary events. We carried out population surveys to estimate trends in risk factors. Trends in event rates were regressed on trends in risk score and in individual risk factors. Findings Smoking rates decreased in most male populations but trends were mixed in women; mean blood pressures and cholesterol concentrations decreased, body-mass index increased, and overall risk scores and coronary-event rates decreased. The model of trends in 10-year coronary-event rates against risk scores and single risk factors showed a poor fit, but this was improved with a 4-year time lag for coronary events. The explanatory power of the analyses was limited by imprecision of the estimates and homogeneity of trends in the study populations. Interpretation Changes in the classic risk factors seem to partly explain the variation in population trends in CHD. Residual variance is attributable to difficulties in measurement and analysis, including time lag, and to factors that were not included, such as medical interventions. The results support prevention policies based on the classic risk factors but suggest potential for prevention beyond these.
Resumo:
Preface The starting point for this work and eventually the subject of the whole thesis was the question: how to estimate parameters of the affine stochastic volatility jump-diffusion models. These models are very important for contingent claim pricing. Their major advantage, availability T of analytical solutions for characteristic functions, made them the models of choice for many theoretical constructions and practical applications. At the same time, estimation of parameters of stochastic volatility jump-diffusion models is not a straightforward task. The problem is coming from the variance process, which is non-observable. There are several estimation methodologies that deal with estimation problems of latent variables. One appeared to be particularly interesting. It proposes the estimator that in contrast to the other methods requires neither discretization nor simulation of the process: the Continuous Empirical Characteristic function estimator (EGF) based on the unconditional characteristic function. However, the procedure was derived only for the stochastic volatility models without jumps. Thus, it has become the subject of my research. This thesis consists of three parts. Each one is written as independent and self contained article. At the same time, questions that are answered by the second and third parts of this Work arise naturally from the issues investigated and results obtained in the first one. The first chapter is the theoretical foundation of the thesis. It proposes an estimation procedure for the stochastic volatility models with jumps both in the asset price and variance processes. The estimation procedure is based on the joint unconditional characteristic function for the stochastic process. The major analytical result of this part as well as of the whole thesis is the closed form expression for the joint unconditional characteristic function for the stochastic volatility jump-diffusion models. The empirical part of the chapter suggests that besides a stochastic volatility, jumps both in the mean and the volatility equation are relevant for modelling returns of the S&P500 index, which has been chosen as a general representative of the stock asset class. Hence, the next question is: what jump process to use to model returns of the S&P500. The decision about the jump process in the framework of the affine jump- diffusion models boils down to defining the intensity of the compound Poisson process, a constant or some function of state variables, and to choosing the distribution of the jump size. While the jump in the variance process is usually assumed to be exponential, there are at least three distributions of the jump size which are currently used for the asset log-prices: normal, exponential and double exponential. The second part of this thesis shows that normal jumps in the asset log-returns should be used if we are to model S&P500 index by a stochastic volatility jump-diffusion model. This is a surprising result. Exponential distribution has fatter tails and for this reason either exponential or double exponential jump size was expected to provide the best it of the stochastic volatility jump-diffusion models to the data. The idea of testing the efficiency of the Continuous ECF estimator on the simulated data has already appeared when the first estimation results of the first chapter were obtained. In the absence of a benchmark or any ground for comparison it is unreasonable to be sure that our parameter estimates and the true parameters of the models coincide. The conclusion of the second chapter provides one more reason to do that kind of test. Thus, the third part of this thesis concentrates on the estimation of parameters of stochastic volatility jump- diffusion models on the basis of the asset price time-series simulated from various "true" parameter sets. The goal is to show that the Continuous ECF estimator based on the joint unconditional characteristic function is capable of finding the true parameters. And, the third chapter proves that our estimator indeed has the ability to do so. Once it is clear that the Continuous ECF estimator based on the unconditional characteristic function is working, the next question does not wait to appear. The question is whether the computation effort can be reduced without affecting the efficiency of the estimator, or whether the efficiency of the estimator can be improved without dramatically increasing the computational burden. The efficiency of the Continuous ECF estimator depends on the number of dimensions of the joint unconditional characteristic function which is used for its construction. Theoretically, the more dimensions there are, the more efficient is the estimation procedure. In practice, however, this relationship is not so straightforward due to the increasing computational difficulties. The second chapter, for example, in addition to the choice of the jump process, discusses the possibility of using the marginal, i.e. one-dimensional, unconditional characteristic function in the estimation instead of the joint, bi-dimensional, unconditional characteristic function. As result, the preference for one or the other depends on the model to be estimated. Thus, the computational effort can be reduced in some cases without affecting the efficiency of the estimator. The improvement of the estimator s efficiency by increasing its dimensionality faces more difficulties. The third chapter of this thesis, in addition to what was discussed above, compares the performance of the estimators with bi- and three-dimensional unconditional characteristic functions on the simulated data. It shows that the theoretical efficiency of the Continuous ECF estimator based on the three-dimensional unconditional characteristic function is not attainable in practice, at least for the moment, due to the limitations on the computer power and optimization toolboxes available to the general public. Thus, the Continuous ECF estimator based on the joint, bi-dimensional, unconditional characteristic function has all the reasons to exist and to be used for the estimation of parameters of the stochastic volatility jump-diffusion models.
Resumo:
This paper focused on four alternatives of analysis of experiments in square lattice as far as the estimation of variance components and some genetic parameters are concerned: 1) intra-block analysis with adjusted treatment and blocks within unadjusted repetitions; 2) lattice analysis as complete randomized blocks; 3) intrablock analysis with unadjusted treatment and blocks within adjusted repetitions; 4) lattice analysis as complete randomized blocks, by utilizing the adjusted means of treatments, obtained from the analysis with recovery of interblock information, having as mean square of the error the mean effective variance of this same analysis with recovery of inter-block information. For the four alternatives of analysis, the estimators and estimates were obtained for the variance components and heritability coefficients. The classification of material was also studied. The present study suggests that for each experiment and depending of the objectives of the analysis, one should observe which alternative of analysis is preferable, mainly in cases where a negative estimate is obtained for the variance component due to effects of blocks within adjusted repetitions.
Resumo:
In the first part of the study, nine estimators of the first-order autoregressive parameter are reviewed and a new estimator is proposed. The relationships and discrepancies between the estimators are discussed in order to achieve a clear differentiation. In the second part of the study, the precision in the estimation of autocorrelation is studied. The performance of the ten lag-one autocorrelation estimators is compared in terms of Mean Square Error (combining bias and variance) using data series generated by Monte Carlo simulation. The results show that there is not a single optimal estimator for all conditions, suggesting that the estimator ought to be chosen according to sample size and to the information available of the possible direction of the serial dependence. Additionally, the probability of labelling an actually existing autocorrelation as statistically significant is explored using Monte Carlo sampling. The power estimates obtained are quite similar among the tests associated with the different estimators. These estimates evidence the small probability of detecting autocorrelation in series with less than 20 measurement times.
Resumo:
Image registration has been proposed as an automatic method for recovering cardiac displacement fields from Tagged Magnetic Resonance Imaging (tMRI) sequences. Initially performed as a set of pairwise registrations, these techniques have evolved to the use of 3D+t deformation models, requiring metrics of joint image alignment (JA). However, only linear combinations of cost functions defined with respect to the first frame have been used. In this paper, we have applied k-Nearest Neighbors Graphs (kNNG) estimators of the -entropy (H ) to measure the joint similarity between frames, and to combine the information provided by different cardiac views in an unified metric. Experiments performed on six subjects showed a significantly higher accuracy (p < 0.05) with respect to a standard pairwise alignment (PA) approach in terms of mean positional error and variance with respect to manually placed landmarks. The developed method was used to study strains in patients with myocardial infarction, showing a consistency between strain, infarction location, and coronary occlusion. This paper also presentsan interesting clinical application of graph-based metric estimators, showing their value for solving practical problems found in medical imaging.