43 resultados para Mean-Reverting Jump-Diffusion

em Université de Montréal, Canada


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dans cette thèse, nous étudions quelques problèmes fondamentaux en mathématiques financières et actuarielles, ainsi que leurs applications. Cette thèse est constituée de trois contributions portant principalement sur la théorie de la mesure de risques, le problème de l’allocation du capital et la théorie des fluctuations. Dans le chapitre 2, nous construisons de nouvelles mesures de risque cohérentes et étudions l’allocation de capital dans le cadre de la théorie des risques collectifs. Pour ce faire, nous introduisons la famille des "mesures de risque entropique cumulatifs" (Cumulative Entropic Risk Measures). Le chapitre 3 étudie le problème du portefeuille optimal pour le Entropic Value at Risk dans le cas où les rendements sont modélisés par un processus de diffusion à sauts (Jump-Diffusion). Dans le chapitre 4, nous généralisons la notion de "statistiques naturelles de risque" (natural risk statistics) au cadre multivarié. Cette extension non-triviale produit des mesures de risque multivariées construites à partir des données financiéres et de données d’assurance. Le chapitre 5 introduit les concepts de "drawdown" et de la "vitesse d’épuisement" (speed of depletion) dans la théorie de la ruine. Nous étudions ces concepts pour des modeles de risque décrits par une famille de processus de Lévy spectrallement négatifs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Presently, conditions ensuring the validity of bootstrap methods for the sample mean of (possibly heterogeneous) near epoch dependent (NED) functions of mixing processes are unknown. Here we establish the validity of the bootstrap in this context, extending the applicability of bootstrap methods to a class of processes broadly relevant for applications in economics and finance. Our results apply to two block bootstrap methods: the moving blocks bootstrap of Künsch ( 989) and Liu and Singh ( 992), and the stationary bootstrap of Politis and Romano ( 994). In particular, the consistency of the bootstrap variance estimator for the sample mean is shown to be robust against heteroskedasticity and dependence of unknown form. The first order asymptotic validity of the bootstrap approximation to the actual distribution of the sample mean is also established in this heterogeneous NED context.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we provide both qualitative and quantitative measures of the cost of measuring the integrated volatility by the realized volatility when the frequency of observation is fixed. We start by characterizing for a general diffusion the difference between the realized and the integrated volatilities for a given frequency of observations. Then, we compute the mean and variance of this noise and the correlation between the noise and the integrated volatility in the Eigenfunction Stochastic Volatility model of Meddahi (2001a). This model has, as special examples, log-normal, affine, and GARCH diffusion models. Using some previous empirical works, we show that the standard deviation of the noise is not negligible with respect to the mean and the standard deviation of the integrated volatility, even if one considers returns at five minutes. We also propose a simple approach to capture the information about the integrated volatility contained in the returns through the leverage effect.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

By reporting his satisfaction with his job or any other experience, an individual does not communicate the number of utils that he feels. Instead, he expresses his posterior preference over available alternatives conditional on acquired knowledge of the past. This new interpretation of reported job satisfaction restores the power of microeconomic theory without denying the essential role of discrepancies between one’s situation and available opportunities. Posterior human wealth discrepancies are found to be the best predictor of reported job satisfaction. Static models of relative utility and other subjective well-being assumptions are all unambiguously rejected by the data, as well as an \"economic\" model in which job satisfaction is a measure of posterior human wealth. The \"posterior choice\" model readily explains why so many people usually report themselves as happy or satisfied, why both younger and older age groups are insensitive to current earning discrepancies, and why the past weighs more heavily than the present and the future.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The GARCH and Stochastic Volatility paradigms are often brought into conflict as two competitive views of the appropriate conditional variance concept : conditional variance given past values of the same series or conditional variance given a larger past information (including possibly unobservable state variables). The main thesis of this paper is that, since in general the econometrician has no idea about something like a structural level of disaggregation, a well-written volatility model should be specified in such a way that one is always allowed to reduce the information set without invalidating the model. To this respect, the debate between observable past information (in the GARCH spirit) versus unobservable conditioning information (in the state-space spirit) is irrelevant. In this paper, we stress a square-root autoregressive stochastic volatility (SR-SARV) model which remains true to the GARCH paradigm of ARMA dynamics for squared innovations but weakens the GARCH structure in order to obtain required robustness properties with respect to various kinds of aggregation. It is shown that the lack of robustness of the usual GARCH setting is due to two very restrictive assumptions : perfect linear correlation between squared innovations and conditional variance on the one hand and linear relationship between the conditional variance of the future conditional variance and the squared conditional variance on the other hand. By relaxing these assumptions, thanks to a state-space setting, we obtain aggregation results without renouncing to the conditional variance concept (and related leverage effects), as it is the case for the recently suggested weak GARCH model which gets aggregation results by replacing conditional expectations by linear projections on symmetric past innovations. Moreover, unlike the weak GARCH literature, we are able to define multivariate models, including higher order dynamics and risk premiums (in the spirit of GARCH (p,p) and GARCH in mean) and to derive conditional moment restrictions well suited for statistical inference. Finally, we are able to characterize the exact relationships between our SR-SARV models (including higher order dynamics, leverage effect and in-mean effect), usual GARCH models and continuous time stochastic volatility models, so that previous results about aggregation of weak GARCH and continuous time GARCH modeling can be recovered in our framework.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we consider testing marginal normal distributional assumptions. More precisely, we propose tests based on moment conditions implied by normality. These moment conditions are known as the Stein (1972) equations. They coincide with the first class of moment conditions derived by Hansen and Scheinkman (1995) when the random variable of interest is a scalar diffusion. Among other examples, Stein equation implies that the mean of Hermite polynomials is zero. The GMM approach we adopted is well suited for two reasons. It allows us to study in detail the parameter uncertainty problem, i.e., when the tests depend on unknown parameters that have to be estimated. In particular, we characterize the moment conditions that are robust against parameter uncertainty and show that Hermite polynomials are special examples. This is the main contribution of the paper. The second reason for using GMM is that our tests are also valid for time series. In this case, we adopt a Heteroskedastic-Autocorrelation-Consistent approach to estimate the weighting matrix when the dependence of the data is unspecified. We also make a theoretical comparison of our tests with Jarque and Bera (1980) and OPG regression tests of Davidson and MacKinnon (1993). Finite sample properties of our tests are derived through a comprehensive Monte Carlo study. Finally, three applications to GARCH and realized volatility models are presented.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we propose exact likelihood-based mean-variance efficiency tests of the market portfolio in the context of Capital Asset Pricing Model (CAPM), allowing for a wide class of error distributions which include normality as a special case. These tests are developed in the frame-work of multivariate linear regressions (MLR). It is well known however that despite their simple statistical structure, standard asymptotically justified MLR-based tests are unreliable. In financial econometrics, exact tests have been proposed for a few specific hypotheses [Jobson and Korkie (Journal of Financial Economics, 1982), MacKinlay (Journal of Financial Economics, 1987), Gib-bons, Ross and Shanken (Econometrica, 1989), Zhou (Journal of Finance 1993)], most of which depend on normality. For the gaussian model, our tests correspond to Gibbons, Ross and Shanken’s mean-variance efficiency tests. In non-gaussian contexts, we reconsider mean-variance efficiency tests allowing for multivariate Student-t and gaussian mixture errors. Our framework allows to cast more evidence on whether the normality assumption is too restrictive when testing the CAPM. We also propose exact multivariate diagnostic checks (including tests for multivariate GARCH and mul-tivariate generalization of the well known variance ratio tests) and goodness of fit tests as well as a set estimate for the intervening nuisance parameters. Our results [over five-year subperiods] show the following: (i) multivariate normality is rejected in most subperiods, (ii) residual checks reveal no significant departures from the multivariate i.i.d. assumption, and (iii) mean-variance efficiency tests of the market portfolio is not rejected as frequently once it is allowed for the possibility of non-normal errors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Affiliation: Paul Allard : Département de kinésiologie, Université de Montréal

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Affiliation: Faculté de médecine, Université de Montréal & CANVAC

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Affiliation: Margaret Cargo : Département de médecine sociale et préventive, Faculté de médecine, Université de Montréal

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ancré dans une perspective historique, ce mémoire cherche à mettre en application une relecture de la théorie wébérienne de la « rationalisation éthique » comme facteur explicatif de la reconfiguration moderne du rapport entretenu entre les individus et la religion. Un retour sur les changements survenus dans la pensée religieuse de la Renaissance — pensée mise en contraste avec la situation religieuse des populations du Moyen-Âge — permet de mettre en évidence le passage d’une religion syncrétique, ritualiste et imprégnée de magie, à un christianisme épuré, intériorisé et rationnel. L’étude de la pensée religieuse de l’humaniste Érasme de Rotterdam, pris comme « figure historique » porteuse de cette transformation, pointe vers la diffusion à la Renaissance d’un christianisme compris comme système philosophique compréhensif dépouillé de son caractère mystique. Cette diffusion d’un « esprit » chrétien, et l’importance accordée à la mise en œuvre d’une conduite de vie méthodique spécifiquement orientée vers le salut, participe au premier chef d’un processus de « quotidianisation » du charisme religieux, prélude essentiel, dans une perspective wébérienne, à la « rationalisation éthique » et à l’autonomisation de la sphère religieuse dans la vie sociale.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

"Mémoire présenté à la Faculté des études supérieures en vue de l'obtention du grade de maîtrise en droit des technologies de l'information"

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Un fichier intitulé Charbonneau_Nathalie_2008_AnimationAnnexeT accompagne la thèse. Il contient une séquence animée démontrant le type de parcours pouvant être effectué au sein des environnements numériques développés. Il s'agit d'un fichier .wmv qui a été compressé.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Travail de recherche réalisé à l’hiver 2009 dans le cadre du cours Arv1050 Introduction à l’archivistique sous la direction de la professeure Sabine Mas.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Comprendre le mode d’existence de l’organisation est certainement l’un des plus grands défis que se sont donnés les chercheurs qui s’intéressent à ce domaine d’étude. La littérature nous présente ainsi plusieurs images, métaphores et perspectives qui, combinées, dressent un portrait hybride de ce type de collectif. Je propose, dans cette thèse, de reconnaître et exploiter ce caractère hybride de l’organisation en partant d’une réflexion centrée sur l'espace. En m’inspirant particulièrement des travaux de la géographe Doreen Massey (1999, 2005), le concept d'espace auquel je souscris est celui d’un espace ouvert et dynamique (qui incorpore le temps), basé sur une relationalité matérielle et hétérogène, supposant des acteurs humains et non humains en interaction. L'espace peut donc être compris comme la coexistence d’ontologies hétérogènes, ce que Massey (2005) nomme une coexistence de trajectoires comme stories-so-far. Il s’agit ici d’une vision performative de l’espace organisationnel qui est constitué dans la relation de trajectoires distinctes qui coexistent, se rencontrent, s’affectent, entrent en conflit ou coopèrent (Massey, 1999). Je postule que pour assurer une certaine continuité et cohérence dans la coexistence de trajectoires hétérogènes, un travail d’alignement et d’ordonnancement est mis à l’oeuvre, et ce, par le suivi d’une trajectoire principale — ce que je nomme une trajectoire scriptée. Suivre cette trajectoire permet ainsi à l’organisation de s’étendre, de se rendre présente dans le temps et dans l’espace, sans pour autant perdre son identité : to be here and there at the same time, now and then at the same place. À partir de cette définition de l’espace, je propose d’« espacer l’organisation », et plus particulièrement d’« espacer » Explora, un programme d’éducation non formelle du gouvernement du Chili visant la diffusion et la valorisation de la science et de la technologie. Cette proposition est double : elle renvoie aux pratiques d’espacements — des pratiques hybrides, collectives et situées — des agents organisationnels (dans ce cas, aux pratiques des agents d’Explora impliqués dans l’organisation d’un projet, celui de la Semaine de la science 2006),mais aussi à une pratique de recherche. « Espacer l’organisation » veut donc dire déployer ces espaces pleins, déplier l’organisation, accroître la série des simultanéités-successions pour ainsi créer plus d’espace-temps.