305 resultados para Stochastic nonlinear regression
Resumo:
This paper examines the impact of allowing for stochastic volatility and jumps (SVJ) in a structural model on corporate credit risk prediction. The results from a simulation study verify the better performance of the SVJ model compared with the commonly used Merton model, and three sources are provided to explain the superiority. The empirical analysis on two real samples further ascertains the importance of recognizing the stochastic volatility and jumps by showing that the SVJ model decreases bias in spread prediction from the Merton model, and better explains the time variation in actual CDS spreads. The improvements are found particularly apparent in small firms or when the market is turbulent such as the recent financial crisis.
Resumo:
This paper addresses the issue of output feedback model predictive control for linear systems with input constraints and stochastic disturbances. We show that the optimal policy uses the Kalman filter for state estimation, but the resultant state estimates are not utilized in a certainty equivalence control law
Resumo:
This paper proposes a nonlinear excitation controller to improve transient stability, oscillation damping and voltage regulation of the power system. The energy function of the predicted system states is used to obtain the desired flux for the next time step, which in turn is used to obtain a supplementary control input using an inverse filtering method. The inverse filtering technique enables the system to provide an additional input for the excitation system, which forces the system to track the desired flux. Synchronous generator flux saturation model is used in this paper. A single machine infinite bus (SMIB) test system is used to demonstrate the efficacy of the proposed control method using time-domain simulations. The robustness of the controller is assessed under different operating conditions.
Resumo:
Demand response can be used for providing regulation services in the electricity markets. The retailers can bid in a day-ahead market and respond to real-time regulation signal by load control. This paper proposes a new stochastic ranking method to provide regulation services via demand response. A pool of thermostatically controllable appliances (TCAs) such as air conditioners and water heaters are adjusted using direct load control method. The selection of appliances is based on a probabilistic ranking technique utilizing attributes such as temperature variation and statuses of TCAs. These attributes are stochastically forecasted for the next time step using day-ahead information. System performance is analyzed with a sample regulation signal. Network capability to provide regulation services under various seasons is analyzed. The effect of network size on the regulation services is also investigated.
Resumo:
To enhance the efficiency of regression parameter estimation by modeling the correlation structure of correlated binary error terms in quantile regression with repeated measurements, we propose a Gaussian pseudolikelihood approach for estimating correlation parameters and selecting the most appropriate working correlation matrix simultaneously. The induced smoothing method is applied to estimate the covariance of the regression parameter estimates, which can bypass density estimation of the errors. Extensive numerical studies indicate that the proposed method performs well in selecting an accurate correlation structure and improving regression parameter estimation efficiency. The proposed method is further illustrated by analyzing a dental dataset.
Resumo:
In the Bayesian framework a standard approach to model criticism is to compare some function of the observed data to a reference predictive distribution. The result of the comparison can be summarized in the form of a p-value, and it's well known that computation of some kinds of Bayesian predictive p-values can be challenging. The use of regression adjustment approximate Bayesian computation (ABC) methods is explored for this task. Two problems are considered. The first is the calibration of posterior predictive p-values so that they are uniformly distributed under some reference distribution for the data. Computation is difficult because the calibration process requires repeated approximation of the posterior for different data sets under the reference distribution. The second problem considered is approximation of distributions of prior predictive p-values for the purpose of choosing weakly informative priors in the case where the model checking statistic is expensive to compute. Here the computation is difficult because of the need to repeatedly sample from a prior predictive distribution for different values of a prior hyperparameter. In both these problems we argue that high accuracy in the computations is not required, which makes fast approximations such as regression adjustment ABC very useful. We illustrate our methods with several samples.
Resumo:
This thesis progresses Bayesian experimental design by developing novel methodologies and extensions to existing algorithms. Through these advancements, this thesis provides solutions to several important and complex experimental design problems, many of which have applications in biology and medicine. This thesis consists of a series of published and submitted papers. In the first paper, we provide a comprehensive literature review on Bayesian design. In the second paper, we discuss methods which may be used to solve design problems in which one is interested in finding a large number of (near) optimal design points. The third paper presents methods for finding fully Bayesian experimental designs for nonlinear mixed effects models, and the fourth paper investigates methods to rapidly approximate the posterior distribution for use in Bayesian utility functions.
Resumo:
This paper investigates stochastic analysis of transit segment hourly passenger load factor variation for transit capacity and quality of service (QoS) analysis using Automatic Fare Collection data for a premium radial bus route in Brisbane, Australia. It compares stochastic analysis to traditional peak hour factor (PHF) analysis to gain further insight into variability of transit route segments’ passenger loading during a study hour. It demonstrates that hourly design load factor is a useful method of modeling a route segment’s capacity and QoS time history across the study weekday. This analysis method is readily adaptable to different passenger load standards by adjusting design percentile, reflecting either a more relaxed or more stringent condition. This paper also considers hourly coefficient of variation of load factor as a capacity and QoS assessment measure, in particular through its relationships with hourly average and design load factors. Smaller value reflects uniform passenger loading, which is generally indicative of well dispersed passenger boarding demands and good schedule maintenance. Conversely, higher value may be indicative of pulsed or uneven passenger boarding demands, poor schedule maintenance, and/or bus bunching. An assessment table based on hourly coefficient of variation of load factor is developed and applied to this case study. Inferences are drawn for a selection of study hours across the weekday studied.
Resumo:
This study uses weekday Automatic Fare Collection (AFC) data on a premium bus line in Brisbane, Australia •Stochastic analysis is compared to peak hour factor (PHF) analysis for insight into passenger loading variability •Hourly design load factor (e.g. 88th percentile) is found to be a useful method of modeling a segment’s passenger demand time-history across a study weekday, for capacity and QoS assessment •Hourly coefficient of variation of load factor is found to be a useful QoS and operational assessment measure, particularly through its relationship with hourly average load factor, and with design load factor •An assessment table based on hourly coefficient of variation of load factor is developed from the case study
Resumo:
In this paper we propose a novel approach to multi-action recognition that performs joint segmentation and classification. This approach models each action using a Gaussian mixture using robust low-dimensional action features. Segmentation is achieved by performing classification on overlapping temporal windows, which are then merged to produce the final result. This approach is considerably less complicated than previous methods which use dynamic programming or computationally expensive hidden Markov models (HMMs). Initial experiments on a stitched version of the KTH dataset show that the proposed approach achieves an accuracy of 78.3%, outperforming a recent HMM-based approach which obtained 71.2%.