979 resultados para Fonction rénale


Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we introduce a new approach for volatility modeling in discrete and continuous time. We follow the stochastic volatility literature by assuming that the variance is a function of a state variable. However, instead of assuming that the loading function is ad hoc (e.g., exponential or affine), we assume that it is a linear combination of the eigenfunctions of the conditional expectation (resp. infinitesimal generator) operator associated to the state variable in discrete (resp. continuous) time. Special examples are the popular log-normal and square-root models where the eigenfunctions are the Hermite and Laguerre polynomials respectively. The eigenfunction approach has at least six advantages: i) it is general since any square integrable function may be written as a linear combination of the eigenfunctions; ii) the orthogonality of the eigenfunctions leads to the traditional interpretations of the linear principal components analysis; iii) the implied dynamics of the variance and squared return processes are ARMA and, hence, simple for forecasting and inference purposes; (iv) more importantly, this generates fat tails for the variance and returns processes; v) in contrast to popular models, the variance of the variance is a flexible function of the variance; vi) these models are closed under temporal aggregation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In spatial environments, we consider social welfare functions satisfying Arrow's requirements. i.e., weak Pareto and independence of irrelevant alternatives. When the policy space os a one-dimensional continuum, such a welfare function is determined by a collection of 2n strictly quasi-concave preferences and a tie-breaking rule. As a corrollary, we obtain that when the number of voters is odd, simple majority voting is transitive if and only if each voter's preference is strictly quasi-concave. When the policy space is multi-dimensional, we establish Arrow's impossibility theorem. Among others, we show that weak Pareto, independence of irrelevant alternatives, and non-dictatorship are inconsistent if the set of alternatives has a non-empty interior and it is compact and convex.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The focus of the paper is the nonparametric estimation of an instrumental regression function P defined by conditional moment restrictions stemming from a structural econometric model : E[Y-P(Z)|W]=0 and involving endogenous variables Y and Z and instruments W. The function P is the solution of an ill-posed inverse problem and we propose an estimation procedure based on Tikhonov regularization. The paper analyses identification and overidentification of this model and presents asymptotic properties of the estimated nonparametric instrumental regression function.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Consistency of a binary relation requires any preference cycle to involve indifference only. As shown by Suzumura (1976b), consistency is necessary and sufficient for the existence of an ordering extension of a relation. Because of this important role of consistency, it is of interest to examine the rationalizability of choice functions by means of consistent relations. We describe the logical relationships between the different notions of rationalizability obtained if reflexivity or completeness are added to consistency, both for greatest-element rationalizability and for maximal-element rationalizability. All but one notion of consistent rationalizability are characterized for general domains, and all of them are characterized for domains that contain all two-element subsets of the universal set.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Ce Texte Constitue un Survol des Differentes Approches Destines a Mesurer le Progres Technique. Nous Utilisons une Notation Uniforme Tout au Long des Demonstrations Mathematiques et Nous Faisons Ressortir les Hypotheses Qui Rendent L'application des Methodes Proposees Envisageable et Qui En Limitent la Portee. les Diverses Approches Sont Regroupees D'apres une Classification Suggeree Par Diewert (1981) Selon Laquelle Deux Groupes Sont a Distinguer. le Premier Groupe Contient Toutes les Methodes Definissant le Progres Technique Comme le Taux de Croissance D'un Indice des Outputs Divise Par un Indice des Inputs (Approche de Divisia). L'autre Groupe Inclut Toutes les Methodes Definissant le Progres Technique Comme Etant le Deplacement D'une Fonction Representant la Technologie (Production, Cout, Distance). Ce Second Groupe Est Subdivise Entre L'approche Econometrique,La Theorie des Nombres Indices et L 'Approche Non Parametrique. une Liste des Pricipaux Economistes a Qui L'on Doit les Diverses Approches Est Fournie. Cependant Ce Survol Est Suffisamment Detaille Pour Etre Lu Sans Se Referer aux Articles Originaux.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Dans Cet Article J'ai Cherche a Demontrer les Liens Qui Existent Entre la Theorie Quantitative de la Monnaie, la Theorie du \"Markup\" et L'inflation. Bien Qu'il Ne Soit Pas Necessaire D'admettre L'equilibre et les Courbes Is-Lm, Ma Theorie du Capital Fictif Est Compatible Avec le Q de Tobin. le Principal Avantage de la Theorie du \"Markup\" Flexible Est de Montrer Comment L'inflation Est Fonction Non Seulement du Prix et de la Productivite du Travail, Mais Aussi du Prix de la Productivite du Capital, de Son Taux D'amortissement et de Son Taux de Financement. les Nouveaux Resultats Econometriques Obtenus a Partir des Donnees Annuelles Canadiennes Illustrent Hors de Tout Doute le Bien Fonde de la Relation Entre Capital Fictif et Inflation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this article we study the effect of uncertainty on an entrepreneur who must choose the capacity of his business before knowing the demand for his product. The unit profit of operation is known with certainty but there is no flexibility in our one-period framework. We show how the introduction of global uncertainty reduces the investment of the risk neutral entrepreneur and, even more, that the risk averse one. We also show how marginal increases in risk reduce the optimal capacity of both the risk neutral and the risk averse entrepreneur, without any restriction on the concave utility function and with limited restrictions on the definition of a mean preserving spread. These general results are explained by the fact that the newsboy has a piecewise-linear, and concave, monetary payoff witha kink endogenously determined at the level of optimal capacity. Our results are compared with those in the two literatures on price uncertainty and demand uncertainty, and particularly, with the recent contributions of Eeckhoudt, Gollier and Schlesinger (1991, 1995).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La présente étude analyse les effets dynamiques de la dévaluation du franc CFA, à l’aide d’un modèle monétaire d’équilibre général, intertemporel et multisectoriel. L’accent est particulièrement mis sur les interactions entre dévaluation et accumulation du capital. Dans le modèle, les effets du changement de parité passent par le marché du travail qui se caractérise par l’inertie du salaire nominal. Les résultats montrent que la dévaluation relance l’investissement, avec des effets expansionnistes sur l’activité économique. Le choc monétaire n’a eu qu’un impact limité sur les soldes budgétaire et commercial. Une mesure d’accompagnement telle que la réduction des salaires de la fonction publique améliore ces deux soldes mais déclenche un processus récessif.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper develops a general stochastic framework and an equilibrium asset pricing model that make clear how attitudes towards intertemporal substitution and risk matter for option pricing. In particular, we show under which statistical conditions option pricing formulas are not preference-free, in other words, when preferences are not hidden in the stock and bond prices as they are in the standard Black and Scholes (BS) or Hull and White (HW) pricing formulas. The dependence of option prices on preference parameters comes from several instantaneous causality effects such as the so-called leverage effect. We also emphasize that the most standard asset pricing models (CAPM for the stock and BS or HW preference-free option pricing) are valid under the same stochastic setting (typically the absence of leverage effect), regardless of preference parameter values. Even though we propose a general non-preference-free option pricing formula, we always keep in mind that the BS formula is dominant both as a theoretical reference model and as a tool for practitioners. Another contribution of the paper is to characterize why the BS formula is such a benchmark. We show that, as soon as we are ready to accept a basic property of option prices, namely their homogeneity of degree one with respect to the pair formed by the underlying stock price and the strike price, the necessary statistical hypotheses for homogeneity provide BS-shaped option prices in equilibrium. This BS-shaped option-pricing formula allows us to derive interesting characterizations of the volatility smile, that is, the pattern of BS implicit volatilities as a function of the option moneyness. First, the asymmetry of the smile is shown to be equivalent to a particular form of asymmetry of the equivalent martingale measure. Second, this asymmetry appears precisely when there is either a premium on an instantaneous interest rate risk or on a generalized leverage effect or both, in other words, whenever the option pricing formula is not preference-free. Therefore, the main conclusion of our analysis for practitioners should be that an asymmetric smile is indicative of the relevance of preference parameters to price options.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper exploits the term structure of interest rates to develop testable economic restrictions on the joint process of long-term interest rates and inflation when the latter is subject to a targeting policy by the Central Bank. Two competing models that econometrically describe agents’ inferences about inflation targets are developed and shown to generate distinct predictions on the behavior of interest rates. In an empirical application to the Canadian inflation target zone, results indicate that agents perceive the band to be substantially narrower than officially announced and asymmetric around the stated mid-point. The latter result (i) suggests that the monetary authority attaches different weights to positive and negative deviations from the central target, and (ii) challenges on empirical grounds the assumption, frequently made in the literature, that the policy maker’s loss function is symmetric (usually a quadratic function) around a desired inflation value.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We extend the class of M-tests for a unit root analyzed by Perron and Ng (1996) and Ng and Perron (1997) to the case where a change in the trend function is allowed to occur at an unknown time. These tests M(GLS) adopt the GLS detrending approach of Dufour and King (1991) and Elliott, Rothenberg and Stock (1996) (ERS). Following Perron (1989), we consider two models : one allowing for a change in slope and the other for both a change in intercept and slope. We derive the asymptotic distribution of the tests as well as that of the feasible point optimal tests PT(GLS) suggested by ERS. The asymptotic critical values of the tests are tabulated. Also, we compute the non-centrality parameter used for the local GLS detrending that permits the tests to have 50% asymptotic power at that value. We show that the M(GLS) and PT(GLS) tests have an asymptotic power function close to the power envelope. An extensive simulation study analyzes the size and power in finite samples under various methods to select the truncation lag for the autoregressive spectral density estimator. An empirical application is also provided.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose finite sample tests and confidence sets for models with unobserved and generated regressors as well as various models estimated by instrumental variables methods. The validity of the procedures is unaffected by the presence of identification problems or \"weak instruments\", so no detection of such problems is required. We study two distinct approaches for various models considered by Pagan (1984). The first one is an instrument substitution method which generalizes an approach proposed by Anderson and Rubin (1949) and Fuller (1987) for different (although related) problems, while the second one is based on splitting the sample. The instrument substitution method uses the instruments directly, instead of generated regressors, in order to test hypotheses about the \"structural parameters\" of interest and build confidence sets. The second approach relies on \"generated regressors\", which allows a gain in degrees of freedom, and a sample split technique. For inference about general possibly nonlinear transformations of model parameters, projection techniques are proposed. A distributional theory is obtained under the assumptions of Gaussian errors and strictly exogenous regressors. We show that the various tests and confidence sets proposed are (locally) \"asymptotically valid\" under much weaker assumptions. The properties of the tests proposed are examined in simulation experiments. In general, they outperform the usual asymptotic inference methods in terms of both reliability and power. Finally, the techniques suggested are applied to a model of Tobin’s q and to a model of academic performance.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper considers various asymptotic approximations in the near-integrated firstorder autoregressive model with a non-zero initial condition. We first extend the work of Knight and Satchell (1993), who considered the random walk case with a zero initial condition, to derive the expansion of the relevant joint moment generating function in this more general framework. We also consider, as alternative approximations, the stochastic expansion of Phillips (1987c) and the continuous time approximation of Perron (1991). We assess how these alternative methods provide or not an adequate approximation to the finite-sample distribution of the least-squares estimator in a first-order autoregressive model. The results show that, when the initial condition is non-zero, Perron's (1991) continuous time approximation performs very well while the others only offer improvements when the initial condition is zero.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We provide a theoretical framework to explain the empirical finding that the estimated betas are sensitive to the sampling interval even when using continuously compounded returns. We suppose that stock prices have both permanent and transitory components. The permanent component is a standard geometric Brownian motion while the transitory component is a stationary Ornstein-Uhlenbeck process. The discrete time representation of the beta depends on the sampling interval and two components labelled \"permanent and transitory betas\". We show that if no transitory component is present in stock prices, then no sampling interval effect occurs. However, the presence of a transitory component implies that the beta is an increasing (decreasing) function of the sampling interval for more (less) risky assets. In our framework, assets are labelled risky if their \"permanent beta\" is greater than their \"transitory beta\" and vice versa for less risky assets. Simulations show that our theoretical results provide good approximations for the means and standard deviations of estimated betas in small samples. Our results can be perceived as indirect evidence for the presence of a transitory component in stock prices, as proposed by Fama and French (1988) and Poterba and Summers (1988).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This note investigates the adequacy of the finite-sample approximation provided by the Functional Central Limit Theorem (FCLT) when the errors are allowed to be dependent. We compare the distribution of the scaled partial sums of some data with the distribution of the Wiener process to which it converges. Our setup is purposely very simple in that it considers data generated from an ARMA(1,1) process. Yet, this is sufficient to bring out interesting conclusions about the particular elements which cause the approximations to be inadequate in even quite large sample sizes.