948 resultados para [JEL:C79] Mathematical and Quantitative Methods - Game Theory and Bargaining Theory - Other
Resumo:
In this paper, generalizing results in Alòs, León and Vives (2007b), we see that the dependence of jumps in the volatility under a jump-diffusion stochastic volatility model, has no effect on the short-time behaviour of the at-the-money implied volatility skew, although the corresponding Hull and White formula depends on the jumps. Towards this end, we use Malliavin calculus techniques for Lévy processes based on Løkka (2004), Petrou (2006), and Solé, Utzet and Vives (2007).
Resumo:
Species' geographic ranges are usually considered as basic units in macroecology and biogeography, yet it is still difficult to measure them accurately for many reasons. About 20 years ago, researchers started using local data on species' occurrences to estimate broad scale ranges, thereby establishing the niche modeling approach. However, there are still many problems in model evaluation and application, and one of the solutions is to find a consensus solution among models derived from different mathematical and statistical models for niche modeling, climatic projections and variable combination, all of which are sources of uncertainty during niche modeling. In this paper, we discuss this approach of ensemble forecasting and propose that it can be divided into three phases with increasing levels of complexity. Phase I is the simple combination of maps to achieve a consensual and hopefully conservative solution. In Phase II, differences among the maps used are described by multivariate analyses, and Phase III consists of the quantitative evaluation of the relative magnitude of uncertainties from different sources and their mapping. To illustrate these developments, we analyzed the occurrence data of the tiger moth, Utetheisa ornatrix (Lepidoptera, Arctiidae), a Neotropical moth species, and modeled its geographic range in current and future climates.
Resumo:
Donors often rely on local intermediaries to deliver benefits to target beneficiaries. Each selected recipient observes if the intermediary under-delivers to them, so they serve as natural monitors. However, they may withhold complaints when feeling unentitled or grateful to the intermediary for selecting them. Furthermore, the intermediary may distort selection (e.g. by picking richer recipients who feel less entitled) to reduce complaints. We design an experimental game representing the donor s problem. In one treatment, the intermediary selects recipients. In the other, selection is random - as by an uninformed donor. In our data, random selection dominates delegation of the selection task to the intermediary. Selection distortions are similar, but intermediaries embezzle more when they have selection power and (correctly) expect fewer complaints.
Resumo:
As the prevalence of smoking has decreased to below 20%, health practitioners interest has shifted towards theprevalence of obesity, and reducing it is one of the major health challenges in decades to come. In this paper westudy the impact that the final product of the anti-smoking campaign, that is, smokers quitting the habit, had onaverage weight in the population. To these ends, we use data from the Behavioral Risk Factors Surveillance System,a large series of independent representative cross-sectional surveys. We construct a synthetic panel that allows us tocontrol for unobserved heterogeneity and we exploit the exogenous changes in taxes and regulations to instrumentthe endogenous decision to give up the habit of smoking. Our estimates, are very close to estimates issued in the 90sby the US Department of Health, and indicate that a 10% decrease in the incidence of smoking leads to an averageweight increase of 2.2 to 3 pounds, depending on choice of specification. In addition, we find evidence that the effectovershoots in the short run, although a significant part remains even after two years. However, when we split thesample between men and women, we only find a significant effect for men. Finally, the implicit elasticity of quittingsmoking to the probability of becoming obese is calculated at 0.58. This implies that the net benefit from reducingthe incidence of smoking by 1% is positive even though the cost to society is $0.6 billions.
Resumo:
We present a simple randomized procedure for the prediction of a binary sequence. The algorithm uses ideas from recent developments of the theory of the prediction of individual sequences. We show that if thesequence is a realization of a stationary and ergodic random process then the average number of mistakes converges, almost surely, to that of the optimum, given by the Bayes predictor.
Resumo:
The generalization of simple (two-variable) correspondence analysis to more than two categorical variables, commonly referred to as multiple correspondence analysis, is neither obvious nor well-defined. We present two alternative ways of generalizing correspondence analysis, one based on the quantification of the variables and intercorrelation relationships, and the other based on the geometric ideas of simple correspondence analysis. We propose a version of multiple correspondence analysis, with adjusted principal inertias, as the method of choice for the geometric definition, since it contains simple correspondence analysis as an exact special case, which is not the situation of the standard generalizations. We also clarify the issue of supplementary point representation and the properties of joint correspondence analysis, a method that visualizes all two-way relationships between the variables. The methodology is illustrated using data on attitudes to science from the International Social Survey Program on Environment in 1993.
Resumo:
We investigate on-line prediction of individual sequences. Given a class of predictors, the goal is to predict as well as the best predictor in the class, where the loss is measured by the self information (logarithmic) loss function. The excess loss (regret) is closely related to the redundancy of the associated lossless universal code. Using Shtarkov's theorem and tools from empirical process theory, we prove a general upper bound on the best possible (minimax) regret. The bound depends on certain metric properties of the class of predictors. We apply the bound to both parametric and nonparametric classes ofpredictors. Finally, we point out a suboptimal behavior of the popular Bayesian weighted average algorithm.
Resumo:
We develop a general error analysis framework for the Monte Carlo simulationof densities for functionals in Wiener space. We also study variancereduction methods with the help of Malliavin derivatives. For this, wegive some general heuristic principles which are applied to diffusionprocesses. A comparison with kernel density estimates is made.
Resumo:
This paper analyzes the relationship between ethnic fractionalization, polarization, and conflict. In recent years many authors have found empirical evidence that ethnic fractionalization has a negative effect on growth. One mechanism that can explain this nexus is the effect of ethnic heterogeneity on rent-seeking activities and the increase in potential conflict, which is negative for investment. However the empirical evidence supporting the effect of ethnic fractionalization on the incidence of civil conflicts is very weak. Although ethnic fractionalization may be important for growth, we argue that the channel is not through an increase in potential ethnic conflict. We discuss the appropriateness of indices of polarization to capture conflictive dimensions. We develop a new measure of ethnic heterogeneity that satisfies the basic properties associated with the concept of polarization. The empirical section shows that this index of ethnic polarization is a significant variable in the explanation of the incidence of civil wars. This result is robust to the presence of other indicators of ethnic heterogeneity, other sources of data for the construction of the index, and other data structures.
Resumo:
I discuss the identifiability of a structural New Keynesian Phillips curve when it is embedded in a small scale dynamic stochastic general equilibrium model. Identification problems emerge because not all the structural parameters are recoverable from the semi-structural ones and because the objective functions I consider are poorly behaved. The solution and the moment mappings are responsible for the problems.
Resumo:
Dual scaling of a subjects-by-objects table of dominance data (preferences,paired comparisons and successive categories data) has been contrasted with correspondence analysis, as if the two techniques were somehow different. In this note we show that dual scaling of dominance data is equivalent to the correspondence analysis of a table which is doubled with respect to subjects. We also show that the results of both methods can be recovered from a principal components analysis of the undoubled dominance table which is centred with respect to subject means.
Resumo:
I study monotonicity and uniqueness of the equilibrium strategies in a two-person first price auction with affiliated signals. I show thatwhen the game is symmetric there is a unique Nash equilibrium thatsatisfies a regularity condition requiring that the equilibrium strategies be{\sl piecewise monotone}. Moreover, when the signals are discrete-valued, the equilibrium is unique. The central part of the proof consists of showing that at any regular equilibrium the bidders' strategies must be monotone increasing within the support of winning bids. The monotonicity result derived in this paper provides the missing link for the analysis of uniqueness in two-person first price auctions. Importantly, this result extends to asymmetric auctions.
Resumo:
Structural equation models (SEM) are commonly used to analyze the relationship between variables some of which may be latent, such as individual ``attitude'' to and ``behavior'' concerning specific issues. A number of difficulties arise when we want to compare a large number of groups, each with large sample size, and the manifest variables are distinctly non-normally distributed. Using an specific data set, we evaluate the appropriateness of the following alternative SEM approaches: multiple group versus MIMIC models, continuous versus ordinal variables estimation methods, and normal theory versus non-normal estimation methods. The approaches are applied to the ISSP-1993 Environmental data set, with the purpose of exploring variation in the mean level of variables of ``attitude'' to and ``behavior''concerning environmental issues and their mutual relationship across countries. Issues of both theoretical and practical relevance arise in the course of this application.
Resumo:
In 2007 the first Quality Enhancement Meeting on sampling in the European SocialSurvey (ESS) took place. The discussion focused on design effects and inteviewereffects in face-to-face interviews. Following the recomendations of this meeting theSpanish ESS team studied the impact of interviewers as a new element in the designeffect in the response s variance using the information of the correspondent SampleDesign Data Files. Hierarchical multilevel and cross-classified multilevel analysis areconducted in order to estimate the amount of responses variation due to PSU and tointerviewers for different questions in the survey. Factor such as the age of theinterviewer, gender, workload, training and experience and respondent characteristicssuch as age, gender, renuance to participate and their possible interactions are alsoincluded in the analysis of some specific questions like trust in politicians and trustin legal system . Some recomendations related to future sampling designs and thecontents of the briefing sessions are derived from this initial research.
Resumo:
We consider an agent who has to repeatedly make choices in an uncertainand changing environment, who has full information of the past, who discountsfuture payoffs, but who has no prior. We provide a learning algorithm thatperforms almost as well as the best of a given finite number of experts orbenchmark strategies and does so at any point in time, provided the agentis sufficiently patient. The key is to find the appropriate degree of forgettingdistant past. Standard learning algorithms that treat recent and distant pastequally do not have the sequential epsilon optimality property.