896 resultados para model confidence set
Resumo:
The characteristics of a global set-up of the Finite-Element Sea-Ice Ocean Model under forcing of the period 1958-2004 are presented. The model set-up is designed to study the variability in the deep-water mass formation areas and was therefore regionally better resolved in the deep-water formation areas in the Labrador Sea, Greenland Sea, Weddell Sea and Ross Sea. The sea-ice model reproduces realistic sea-ice distributions and variabilities in the sea-ice extent of both hemispheres as well as sea-ice transport that compares well with observational data. Based on a comparison between model and ocean weather ship data in the North Atlantic, we observe that the vertical structure is well captured in areas with a high resolution. In our model set-up, we are able to simulate decadal ocean variability including several salinity anomaly events and corresponding fingerprint in the vertical hydrography. The ocean state of the model set-up features pronounced variability in the Atlantic Meridional Overturning Circulation as well as the associated mixed layer depth pattern in the North Atlantic deep-water formation areas.
Resumo:
Multivariate volatility forecasts are an important input in many financial applications, in particular portfolio optimisation problems. Given the number of models available and the range of loss functions to discriminate between them, it is obvious that selecting the optimal forecasting model is challenging. The aim of this thesis is to thoroughly investigate how effective many commonly used statistical (MSE and QLIKE) and economic (portfolio variance and portfolio utility) loss functions are at discriminating between competing multivariate volatility forecasts. An analytical investigation of the loss functions is performed to determine whether they identify the correct forecast as the best forecast. This is followed by an extensive simulation study examines the ability of the loss functions to consistently rank forecasts, and their statistical power within tests of predictive ability. For the tests of predictive ability, the model confidence set (MCS) approach of Hansen, Lunde and Nason (2003, 2011) is employed. As well, an empirical study investigates whether simulation findings hold in a realistic setting. In light of these earlier studies, a major empirical study seeks to identify the set of superior multivariate volatility forecasting models from 43 models that use either daily squared returns or realised volatility to generate forecasts. This study also assesses how the choice of volatility proxy affects the ability of the statistical loss functions to discriminate between forecasts. Analysis of the loss functions shows that QLIKE, MSE and portfolio variance can discriminate between multivariate volatility forecasts, while portfolio utility cannot. An examination of the effective loss functions shows that they all can identify the correct forecast at a point in time, however, their ability to discriminate between competing forecasts does vary. That is, QLIKE is identified as the most effective loss function, followed by portfolio variance which is then followed by MSE. The major empirical analysis reports that the optimal set of multivariate volatility forecasting models includes forecasts generated from daily squared returns and realised volatility. Furthermore, it finds that the volatility proxy affects the statistical loss functions’ ability to discriminate between forecasts in tests of predictive ability. These findings deepen our understanding of how to choose between competing multivariate volatility forecasts.
Resumo:
Forecasts of volatility and correlation are important inputs into many practical financial problems. Broadly speaking, there are two ways of generating forecasts of these variables. Firstly, time-series models apply a statistical weighting scheme to historical measurements of the variable of interest. The alternative methodology extracts forecasts from the market traded value of option contracts. An efficient options market should be able to produce superior forecasts as it utilises a larger information set of not only historical information but also the market equilibrium expectation of options market participants. While much research has been conducted into the relative merits of these approaches, this thesis extends the literature along several lines through three empirical studies. Firstly, it is demonstrated that there exist statistically significant benefits to taking the volatility risk premium into account for the implied volatility for the purposes of univariate volatility forecasting. Secondly, high-frequency option implied measures are shown to lead to superior forecasts of the intraday stochastic component of intraday volatility and that these then lead on to superior forecasts of intraday total volatility. Finally, the use of realised and option implied measures of equicorrelation are shown to dominate measures based on daily returns.
Resumo:
The performance of techniques for evaluating multivariate volatility forecasts are not yet as well understood as their univariate counterparts. This paper aims to evaluate the efficacy of a range of traditional statistical-based methods for multivariate forecast evaluation together with methods based on underlying considerations of economic theory. It is found that a statistical-based method based on likelihood theory and an economic loss function based on portfolio variance are the most effective means of identifying optimal forecasts of conditional covariance matrices.
Resumo:
The price formation of financial assets is a complex process. It extends beyond the standard economic paradigm of supply and demand to the understanding of the dynamic behavior of price variability, the price impact of information, and the implications of trading behavior of market participants on prices. In this thesis, I study aggregate market and individual assets volatility, liquidity dimensions, and causes of mispricing for US equities over a recent sample period. How volatility forecasts are modeled, what determines intradaily jumps and causes changes in intradaily volatility and what drives the premium of traded equity indexes? Are they induced, for example, by the information content of lagged volatility and return parameters or by macroeconomic news, changes in liquidity and volatility? Besides satisfying our intellectual curiosity, answers to these questions are of direct importance to investors developing trading strategies, policy makers evaluating macroeconomic policies and to arbitrageurs exploiting mispricing in exchange-traded funds. Results show that the leverage effect and lagged absolute returns improve forecasts of continuous components of daily realized volatility as well as jumps. Implied volatility does not subsume the information content of lagged returns in forecasting realized volatility and its components. The reported results are linked to the heterogeneous market hypothesis and demonstrate the validity of extending the hypothesis to returns. Depth shocks, signed order flow, the number of trades, and resiliency are the most important determinants of intradaily volatility. In contrast, spread shock and resiliency are predictive of signed intradaily jumps. There are fewer macroeconomic news announcement surprises that cause extreme price movements or jumps than those that elevate intradaily volatility. Finally, the premium of exchange-traded funds is significantly associated with momentum in net asset value and a number of liquidity parameters including the spread, traded volume, and illiquidity. The mispricing of industry exchange traded funds suggest that limits to arbitrage are driven by potential illiquidity.
Resumo:
The importance of modelling correlation has long been recognised in the field of portfolio management, with largedimensional multivariate problems increasingly becoming the focus of research. This paper provides a straightforward and commonsense approach toward investigating a number of models used to generate forecasts of the correlation matrix for large-dimensional problems.We find evidence in favour of assuming equicorrelation across various portfolio sizes, particularly during times of crisis. During periods of market calm, however, the suitability of the constant conditional correlation model cannot be discounted, especially for large portfolios. A portfolio allocation problem is used to compare forecasting methods. The global minimum variance portfolio and Model Confidence Set are used to compare methods, while portfolio weight stability and relative economic value are also considered.
Resumo:
O trabalho tem como objetivo comparar a eficácia das diferentes metodologias de projeção de inflação aplicadas ao Brasil. Serão comparados modelos de projeção que utilizam os dados agregados e desagregados do IPCA em um horizonte de até doze meses à frente. Foi utilizado o IPCA na base mensal, com início em janeiro de 1996 e fim em março de 2012. A análise fora da amostra foi feita para o período entre janeiro de 2008 e março de 2012. Os modelos desagregados serão estimados por SARIMA, pelo software X-12 ARIMA disponibilizado pelo US Census Bureau, e terão as aberturas do IPCA de grupos (9) e itens (52), assim como aberturas com sentido mais econômico utilizadas pelo Banco Central do Brasil como: serviços, administrados, alimentos e industrializados; duráveis, não duráveis, semiduráveis, serviços e administrados. Os modelos agregados serão estimados por técnicas como SARIMA, modelos estruturais em espaço-estado (Filtro de Kalman) e Markov-switching. Os modelos serão comparados pela técnica de seleção de modelo Model Confidence Set, introduzida por Hansen, Lunde e Nason (2010), e Dielbod e Mariano (1995), no qual encontramos evidências de ganhos de desempenho nas projeções dos modelos mais desagregados em relação aos modelos agregados.
Resumo:
O objetivo do presente trabalho é utilizar modelos econométricos de séries de tempo para previsão do comportamento da inadimplência agregada utilizando um conjunto amplo de informação, através dos métodos FAVAR (Factor-Augmented Vector Autoregressive) de Bernanke, Boivin e Eliasz (2005) e FAVECM (Factor-augmented Error Correction Models) de Baneerjee e Marcellino (2008). A partir disso, foram construídas previsões fora da amostra de modo a comparar a eficácia de projeção dos modelos contra modelos univariados mais simples - ARIMA - modelo auto-regressivo integrado de média móvel e SARIMA - modelo sazonal auto-regressivo integrado de média móvel. Para avaliação da eficácia preditiva foi utilizada a metodologia MCS (Model Confidence Set) de Hansen, Lunde e James (2011) Essa metodologia permite comparar a superioridade de modelos temporais vis-à-vis a outros modelos.
Resumo:
This work aims to compare the forecast efficiency of different types of methodologies applied to Brazilian Consumer inflation (IPCA). We will compare forecasting models using disaggregated and aggregated data over twelve months ahead. The disaggregated models were estimated by SARIMA and will have different levels of disaggregation. Aggregated models will be estimated by time series techniques such as SARIMA, state-space structural models and Markov-switching. The forecasting accuracy comparison will be made by the selection model procedure known as Model Confidence Set and by Diebold-Mariano procedure. We were able to find evidence of forecast accuracy gains in models using more disaggregated data
Resumo:
O presente estudo tem como objetivo comparar e combinar diferentes técnicas de projeção para o PIB trimestral brasileiro de 1991 ao segundo trimestre de 2014, utilizando dados agregados, e dados desagregados com pesos fixos e estocásticos. Os modelos desagregados univariados e multivariados, assim como os pesos estocásticos, foram estimados pelo algoritmo Autometrics criado por Doornik (2009), através dos níveis de desagregação disponibilizados pelo IBGE no Sistema de Contas Nacionais. Os modelos agregados foram estimados pelo Autometrics, por Markov-Switching e por modelos estruturais de espaço-estado. A metodologia de comparação de projeções utilizada foi o Model Confidence Set, desenvolvida por Hanse, Lunde e Nason (2011). Foram realizadas duas simulações, sendo a primeira com a análise fora da amostra a partir de 2008, e a segunda a partir de 2000, com horizonte de projeção de até 6 passos à frente. Os resultados sugerem que os modelos desagregados com pesos fixos desempenham melhor nos dois primeiros passos, enquanto nos períodos restantes os modelos da série agregada geram melhores previsões.
Resumo:
O presente trabalho tem como objetivo avaliar a capacidade preditiva de modelos econométricos de séries de tempo baseados em indicadores macroeconômicos na previsão da inflação brasileira (IPCA). Os modelos serão ajustados utilizando dados dentro da amostra e suas projeções ex-post serão acumuladas de um a doze meses à frente. As previsões serão comparadas a de modelos univariados como autoregressivo de primeira ordem - AR(1) - que nesse estudo será o benchmark escolhido. O período da amostra vai de janeiro de 2000 até agosto de 2015 para ajuste dos modelos e posterior avaliação. Ao todo foram avaliadas 1170 diferentes variáveis econômicas a cada período a ser projetado, procurando o melhor conjunto preditores para cada ponto no tempo. Utilizou-se o algoritmo Autometrics para a seleção de modelos. A comparação dos modelos foi feita através do Model Confidence Set desenvolvido por Hansen, Lunde e Nason (2010). Os resultados obtidos nesse ensaio apontam evidências de ganhos de desempenho dos modelos multivariados para períodos posteriores a 1 passo à frente.
Resumo:
O trabalho tem como objetivo verificar a existência e a relevância dos Efeitos Calendário em indicadores industriais. São explorados modelos univariados lineares para o indicador mensal da produção industrial brasileira e alguns de seus componentes. Inicialmente é realizada uma análise dentro da amostra valendo-se de modelos estruturais de espaço-estado e do algoritmo de seleção Autometrics, a qual aponta efeito significante da maioria das variáveis relacionadas ao calendário. Em seguida, através do procedimento de Diebold-Mariano (1995) e do Model Confidence Set, proposto por Hansen, Lunde e Nason (2011), são realizadas comparações de previsões de modelos derivados do Autometrics com um dispositivo simples de Dupla Diferença para um horizonte de até 24 meses à frente. Em geral, os modelos Autometrics que consideram as variáveis de calendário se mostram superiores nas projeções de 1 a 2 meses adiante e superam o modelo simples em todos os horizontes. Quando se agrega os componentes de categoria de uso para formar o índice industrial total, há evidências de ganhos nas projeções de prazo mais curto.
Resumo:
This dissertation contains four essays that all share a common purpose: developing new methodologies to exploit the potential of high-frequency data for the measurement, modeling and forecasting of financial assets volatility and correlations. The first two chapters provide useful tools for univariate applications while the last two chapters develop multivariate methodologies. In chapter 1, we introduce a new class of univariate volatility models named FloGARCH models. FloGARCH models provide a parsimonious joint model for low frequency returns and realized measures, and are sufficiently flexible to capture long memory as well as asymmetries related to leverage effects. We analyze the performances of the models in a realistic numerical study and on the basis of a data set composed of 65 equities. Using more than 10 years of high-frequency transactions, we document significant statistical gains related to the FloGARCH models in terms of in-sample fit, out-of-sample fit and forecasting accuracy compared to classical and Realized GARCH models. In chapter 2, using 12 years of high-frequency transactions for 55 U.S. stocks, we argue that combining low-frequency exogenous economic indicators with high-frequency financial data improves the ability of conditionally heteroskedastic models to forecast the volatility of returns, their full multi-step ahead conditional distribution and the multi-period Value-at-Risk. Using a refined version of the Realized LGARCH model allowing for time-varying intercept and implemented with realized kernels, we document that nominal corporate profits and term spreads have strong long-run predictive ability and generate accurate risk measures forecasts over long-horizon. The results are based on several loss functions and tests, including the Model Confidence Set. Chapter 3 is a joint work with David Veredas. We study the class of disentangled realized estimators for the integrated covariance matrix of Brownian semimartingales with finite activity jumps. These estimators separate correlations and volatilities. We analyze different combinations of quantile- and median-based realized volatilities, and four estimators of realized correlations with three synchronization schemes. Their finite sample properties are studied under four data generating processes, in presence, or not, of microstructure noise, and under synchronous and asynchronous trading. The main finding is that the pre-averaged version of disentangled estimators based on Gaussian ranks (for the correlations) and median deviations (for the volatilities) provide a precise, computationally efficient, and easy alternative to measure integrated covariances on the basis of noisy and asynchronous prices. Along these lines, a minimum variance portfolio application shows the superiority of this disentangled realized estimator in terms of numerous performance metrics. Chapter 4 is co-authored with Niels S. Hansen, Asger Lunde and Kasper V. Olesen, all affiliated with CREATES at Aarhus University. We propose to use the Realized Beta GARCH model to exploit the potential of high-frequency data in commodity markets. The model produces high quality forecasts of pairwise correlations between commodities which can be used to construct a composite covariance matrix. We evaluate the quality of this matrix in a portfolio context and compare it to models used in the industry. We demonstrate significant economic gains in a realistic setting including short selling constraints and transaction costs.
Resumo:
We study the problem of measuring the uncertainty of CGE (or RBC)-type model simulations associated with parameter uncertainty. We describe two approaches for building confidence sets on model endogenous variables. The first one uses a standard Wald-type statistic. The second approach assumes that a confidence set (sampling or Bayesian) is available for the free parameters, from which confidence sets are derived by a projection technique. The latter has two advantages: first, confidence set validity is not affected by model nonlinearities; second, we can easily build simultaneous confidence intervals for an unlimited number of variables. We study conditions under which these confidence sets take the form of intervals and show they can be implemented using standard methods for solving CGE models. We present an application to a CGE model of the Moroccan economy to study the effects of policy-induced increases of transfers from Moroccan expatriates.
Resumo:
Asset correlations are of critical importance in quantifying portfolio credit risk and economic capitalin financial institutions. Estimation of asset correlation with rating transition data has focusedon the point estimation of the correlation without giving any consideration to the uncertaintyaround these point estimates. In this article we use Bayesian methods to estimate a dynamicfactor model for default risk using rating data (McNeil et al., 2005; McNeil and Wendin, 2007).Bayesian methods allow us to formally incorporate human judgement in the estimation of assetcorrelation, through the prior distribution and fully characterize a confidence set for the correlations.Results indicate: i) a two factor model rather than the one factor model, as proposed bythe Basel II framework, better represents the historical default data. ii) importance of unobservedfactors in this type of models is reinforced and point out that the levels of the implied asset correlationscritically depend on the latent state variable used to capture the dynamics of default,as well as other assumptions on the statistical model. iii) the posterior distributions of the assetcorrelations show that the Basel recommended bounds, for this parameter, undermine the levelof systemic risk.