934 resultados para out-of-sample forecast


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The term structure of interest rates is often summarized using a handful of yield factors that capture shifts in the shape of the yield curve. In this paper, we develop a comprehensive model for volatility dynamics in the level, slope, and curvature of the yield curve that simultaneously includes level and GARCH effects along with regime shifts. We show that the level of the short rate is useful in modeling the volatility of the three yield factors and that there are significant GARCH effects present even after including a level effect. Further, we find that allowing for regime shifts in the factor volatilities dramatically improves the model’s fit and strengthens the level effect. We also show that a regime-switching model with level and GARCH effects provides the best out-of-sample forecasting performance of yield volatility. We argue that the auxiliary models often used to estimate term structure models with simulation-based estimation techniques should be consistent with the main features of the yield curve that are identified by our model.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background Road safety targets are widely used and provide a basis for evaluating progress in road safety outcomes against a quantified goal. In Australia, a reduction in fatalities from road traffic crashes (RTCs) is a public policy objective: a national target of no more than 5.6 fatalities per 100,000 population by 2010 was set in 2001. The purpose of this paper is to examine the progress Australia and its states and territories have made in reducing RTC fatalities, and to estimate when the 2010 target may be reached by the jurisdictions. Methods Following a descriptive analysis, univariate time-series models estimate past trends in fatality rates over recent decades. Data for differing time periods are analysed and different trend specifications estimated. Preferred models were selected on the basis of statistical criteria and the period covered by the data. The results of preferred regressions are used to determine out-of-sample forecasts of when the national target may be attained by the jurisdictions. Though there are limitations with the time series approach used, inadequate data precluded the estimation of a full causal/structural model. Results Statistically significant reductions in fatality rates since 1971 were found for all jurisdictions with the national rate decreasing on average, 3% per year since 1992. However the gains have varied across time and space, with percent changes in fatality rates ranging from an 8% increase in New South Wales 1972-1981 to a 46% decrease in Queensland 1982-1991. Based on an estimate of past trends, it is possible that the target set for 2010 may not be reached nationally, until 2016. Unsurprisingly, the analysis indicated a range of outcomes for the respective state/territory jurisdictions though these results should be interpreted with caution due to different assumptions and length of data. Conclusions Results indicate that while Australia has been successful over recent decades in reducing RTC mortality, an important gap between aspirations and achievements remains. Moreover, unless there are fairly radical ("trend-breaking") changes in the factors that affect the incidence of RTC fatalities, deaths from RTCs are likely to remain above the national target in some areas of Australia, for years to come.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Between-subject and within-subject variability is ubiquitous in biology and physiology and understanding and dealing with this is one of the biggest challenges in medicine. At the same time it is difficult to investigate this variability by experiments alone. A recent modelling and simulation approach, known as population of models (POM), allows this exploration to take place by building a mathematical model consisting of multiple parameter sets calibrated against experimental data. However, finding such sets within a high-dimensional parameter space of complex electrophysiological models is computationally challenging. By placing the POM approach within a statistical framework, we develop a novel and efficient algorithm based on sequential Monte Carlo (SMC). We compare the SMC approach with Latin hypercube sampling (LHS), a method commonly adopted in the literature for obtaining the POM, in terms of efficiency and output variability in the presence of a drug block through an in-depth investigation via the Beeler-Reuter cardiac electrophysiological model. We show improved efficiency via SMC and that it produces similar responses to LHS when making out-of-sample predictions in the presence of a simulated drug block.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis addresses modeling of financial time series, especially stock market returns and daily price ranges. Modeling data of this kind can be approached with so-called multiplicative error models (MEM). These models nest several well known time series models such as GARCH, ACD and CARR models. They are able to capture many well established features of financial time series including volatility clustering and leptokurtosis. In contrast to these phenomena, different kinds of asymmetries have received relatively little attention in the existing literature. In this thesis asymmetries arise from various sources. They are observed in both conditional and unconditional distributions, for variables with non-negative values and for variables that have values on the real line. In the multivariate context asymmetries can be observed in the marginal distributions as well as in the relationships of the variables modeled. New methods for all these cases are proposed. Chapter 2 considers GARCH models and modeling of returns of two stock market indices. The chapter introduces the so-called generalized hyperbolic (GH) GARCH model to account for asymmetries in both conditional and unconditional distribution. In particular, two special cases of the GARCH-GH model which describe the data most accurately are proposed. They are found to improve the fit of the model when compared to symmetric GARCH models. The advantages of accounting for asymmetries are also observed through Value-at-Risk applications. Both theoretical and empirical contributions are provided in Chapter 3 of the thesis. In this chapter the so-called mixture conditional autoregressive range (MCARR) model is introduced, examined and applied to daily price ranges of the Hang Seng Index. The conditions for the strict and weak stationarity of the model as well as an expression for the autocorrelation function are obtained by writing the MCARR model as a first order autoregressive process with random coefficients. The chapter also introduces inverse gamma (IG) distribution to CARR models. The advantages of CARR-IG and MCARR-IG specifications over conventional CARR models are found in the empirical application both in- and out-of-sample. Chapter 4 discusses the simultaneous modeling of absolute returns and daily price ranges. In this part of the thesis a vector multiplicative error model (VMEM) with asymmetric Gumbel copula is found to provide substantial benefits over the existing VMEM models based on elliptical copulas. The proposed specification is able to capture the highly asymmetric dependence of the modeled variables thereby improving the performance of the model considerably. The economic significance of the results obtained is established when the information content of the volatility forecasts derived is examined.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[ES] Los modelos implícitos constituyen uno de los enfoques de valoración de opciones alternativos al modelo de Black-Scholes que ha conocido un mayor desarrollo en los últimos años. Dentro de este planteamiento existen diferentes alternativas: los árboles implícitos, los modelos con función de volatilidad determinista y los modelos con función de volatilidad implícita. Todos ellos se construyen a partir de una estimación de la distribución de probabilidades riesgo-neutral del precio futuro del activo subyacente, congruente con los precios de mercado de las opciones negociadas. En consecuencia, los modelos implícitos proporcionan buenos resultados en la valoración de opciones dentro de la muestra. Sin embargo, su comportamiento como instrumento de predicción para opciones fuera de muestra no resulta satisfactorio. En este artículo se analiza la medida en la que este enfoque contribuye a la mejora de la valoración de opciones, tanto desde un punto de vista teórico como práctico.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the first part of the thesis we explore three fundamental questions that arise naturally when we conceive a machine learning scenario where the training and test distributions can differ. Contrary to conventional wisdom, we show that in fact mismatched training and test distribution can yield better out-of-sample performance. This optimal performance can be obtained by training with the dual distribution. This optimal training distribution depends on the test distribution set by the problem, but not on the target function that we want to learn. We show how to obtain this distribution in both discrete and continuous input spaces, as well as how to approximate it in a practical scenario. Benefits of using this distribution are exemplified in both synthetic and real data sets.

In order to apply the dual distribution in the supervised learning scenario where the training data set is fixed, it is necessary to use weights to make the sample appear as if it came from the dual distribution. We explore the negative effect that weighting a sample can have. The theoretical decomposition of the use of weights regarding its effect on the out-of-sample error is easy to understand but not actionable in practice, as the quantities involved cannot be computed. Hence, we propose the Targeted Weighting algorithm that determines if, for a given set of weights, the out-of-sample performance will improve or not in a practical setting. This is necessary as the setting assumes there are no labeled points distributed according to the test distribution, only unlabeled samples.

Finally, we propose a new class of matching algorithms that can be used to match the training set to a desired distribution, such as the dual distribution (or the test distribution). These algorithms can be applied to very large datasets, and we show how they lead to improved performance in a large real dataset such as the Netflix dataset. Their computational complexity is the main reason for their advantage over previous algorithms proposed in the covariate shift literature.

In the second part of the thesis we apply Machine Learning to the problem of behavior recognition. We develop a specific behavior classifier to study fly aggression, and we develop a system that allows analyzing behavior in videos of animals, with minimal supervision. The system, which we call CUBA (Caltech Unsupervised Behavior Analysis), allows detecting movemes, actions, and stories from time series describing the position of animals in videos. The method summarizes the data, as well as it provides biologists with a mathematical tool to test new hypotheses. Other benefits of CUBA include finding classifiers for specific behaviors without the need for annotation, as well as providing means to discriminate groups of animals, for example, according to their genetic line.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Semisupervised dimensionality reduction has been attracting much attention as it not only utilizes both labeled and unlabeled data simultaneously, but also works well in the situation of out-of-sample. This paper proposes an effective approach of semisupervised dimensionality reduction through label propagation and label regression. Different from previous efforts, the new approach propagates the label information from labeled to unlabeled data with a well-designed mechanism of random walks, in which outliers are effectively detected and the obtained virtual labels of unlabeled data can be well encoded in a weighted regression model. These virtual labels are thereafter regressed with a linear model to calculate the projection matrix for dimensionality reduction. By this means, when the manifold or the clustering assumption of data is satisfied, the labels of labeled data can be correctly propagated to the unlabeled data; and thus, the proposed approach utilizes the labeled and the unlabeled data more effectively than previous work. Experimental results are carried out upon several databases, and the advantage of the new approach is well demonstrated.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Orthogonal neighborhood-preserving projection (ONPP) is a recently developed orthogonal linear algorithm for overcoming the out-of-sample problem existing in the well-known manifold learning algorithm, i.e., locally linear embedding. It has been shown that ONPP is a strong analyzer of high-dimensional data. However, when applied to classification problems in a supervised setting, ONPP only focuses on the intraclass geometrical information while ignores the interaction of samples from different classes. To enhance the performance of ONPP in classification, a new algorithm termed discriminative ONPP (DONPP) is proposed in this paper. DONPP 1) takes into account both intraclass and interclass geometries; 2) considers the neighborhood information of interclass relationships; and 3) follows the orthogonality property of ONPP. Furthermore, DONPP is extended to the semisupervised case, i.e., semisupervised DONPP (SDONPP). This uses unlabeled samples to improve the classification accuracy of the original DONPP. Empirical studies demonstrate the effectiveness of both DONPP and SDONPP.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

针对非线性系统传感器故障诊断难以解决的问题,提出了一种新的基于局部嵌入映射(LLE)的方法,解决了非线性数据的特征映射问题。首先,改进了基于分形维估计的内在维数的估计,通过线性拟合解决了线性区域的自动确定。然后,将故障状态与空间分布结合起来,通过确定数据点在空间超球内的分布完成故障的检测,在这个过程中将超球的确定与LLE算法中基于核函数的样本外数据扩展结合起来,大大减少了计算量,提高了算法的实时性,从而为复杂非线性传感器的故障诊断提供了一种新的有效的方法。

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Using data on user attributes and interactions from an online dating site, we estimate mate preferences, and use the Gale-Shapley algorithm to predict stable matches. The predicted matches are similar to the actual matches achieved by the dating site, and the actual matches are approximately efficient. Out-of-sample predictions of offline matches, i.e., marriages, exhibit assortative mating patterns similar to those observed in actual marriages. Thus, mate preferences, without resort to search frictions, can generate sorting in marriages. However, we underpredict some of the correlation patterns; search frictions may play a role in explaining the discrepancy.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, we propose a framework for robust optimization that relaxes the standard notion of robustness by allowing the decision maker to vary the protection level in a smooth way across the uncertainty set. We apply our approach to the problem of maximizing the expected value of a payoff function when the underlying distribution is ambiguous and therefore robustness is relevant. Our primary objective is to develop this framework and relate it to the standard notion of robustness, which deals with only a single guarantee across one uncertainty set. First, we show that our approach connects closely to the theory of convex risk measures. We show that the complexity of this approach is equivalent to that of solving a small number of standard robust problems. We then investigate the conservatism benefits and downside probability guarantees implied by this approach and compare to the standard robust approach. Finally, we illustrate theme thodology on an asset allocation example consisting of historical market data over a 25-year investment horizon and find in every case we explore that relaxing standard robustness with soft robustness yields a seemingly favorable risk-return trade-off: each case results in a higher out-of-sample expected return for a relatively minor degradation of out-of-sample downside performance. © 2010 INFORMS.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The paper addresses the issue of choice of bandwidth in the application of semiparametric estimation of the long memory parameter in a univariate time series process. The focus is on the properties of forecasts from the long memory model. A variety of cross-validation methods based on out of sample forecasting properties are proposed. These procedures are used for the choice of bandwidth and subsequent model selection. Simulation evidence is presented that demonstrates the advantage of the proposed new methodology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Reducible diffusions (RDs) are nonlinear transformations of analytically solvable Basic Diffusions (BDs). Hence, by construction RDs are analytically tractable and flexible diffusion processes. Existing literature on RDs has mostly focused on time-homogeneous transformations, which to a significant extent fail to explore the full potential of RDs from both theoretical and practical points of view. In this paper, we propose flexible and economically justifiable time variations to the transformations of RDs. Concentrating on the Constant Elasticity Variance (CEV) RDs, we consider nonlinear dynamics for our time-varying transformations with both deterministic and stochastic designs. Such time variations can greatly enhance the flexibility of RDs while maintaining sufficient tractability of the resulting models. In the meantime, our modeling approach enjoys the benefits of classical inferential techniques such as the Maximum Likelihood (ML). Our application to the UK and the US short-term interest rates suggests that from an empirical point of view time-varying transformations are highly relevant and statistically significant. We expect that the proposed models can describe more truthfully the dynamic time-varying behavior of economic and financial variables and potentially improve out-of-sample forecasts significantly.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We develop a continuous-time asset price model to capture the timeseries momentum documented recently. The underlying stochastic delay differentialsystem facilitates the analysis of effects of different time horizons used bymomentum trading. By studying an optimal asset allocation problem, we find thatthe performance of time series momentum strategy can be significantly improvedby combining with market fundamentals and timing opportunity with respect tomarket trend and volatility. Furthermore, the results also hold for different timehorizons, the out-of-sample tests and with short-sale constraints. The outperformanceof the optimal strategy is immune to market states, investor sentiment andmarket volatility.