13 resultados para Partial Credit Model
em Aston University Research Archive
Resumo:
In this paper the performance of opening and closing returns, for the components of the FT-30 will be studied. It will be shown that for these stocks opening returns have higher volatility and a greater tendency towards negative serial correlation than closing returns. Unlike previous studies this contrasting performance cannot solely be attributed to differences in the trading mechanism across the trading day. All the stocks used in our sample trade thought the day using a uniform trading mechanism. In this paper, we suggest that it is differences in the speed that closing and opening returns adjust to new information that causes differences in return performance. By estimating the Amihud and Mendelson (1987) [Amihud, Yakov, & Mendelson, Haim (1987). Trading mechanisms and stock returns: An empirical investigation, Journal of Finance, 62 533-553.] partial adjustment model with noise, we show that opening returns have a tendency towards over-reaction, while closing returns have a tendency towards under-reaction. We suggest that it is these differences that cause a substantial proportion (although not all) of the asymmetric return patterns associated with opening and closing returns. © 2005 Elsevier Inc. All rights reserved.
Resumo:
In this article a partial-adjustment model, which shows how equity prices fail to adjust instantaneously to new information, is estimated using a Kalman filter. For the components of the Dow Jones Industrial 30 index I aim to identify whether overreaction or noise is the cause of serial correlation and high volatility associated with opening returns. I find that the tendency for overreaction in opening prices is much stronger than for closing prices; therefore, overreaction rather than noise may account for differences in the return behavior of opening and closing returns.
Resumo:
Purpose – On 29 January 2001, Euronext LIFFE introduced single security futures contracts on a range of global companies. The purpose of this paper is to examine the impact that the introduction of these futures contracts had on the behaviour of opening and closing UK equity returns. Design/methodology/approach – The paper models the price discovery process using the Amihud and Mendelson partial adjustment model which can be estimated using a Kalman filter. Findings – Empirical results show that during the pre-futures period both opening and closing returns under-react to new information. After the introduction of futures contracts opening returns over-react. A rise in the partial adjustment coefficient also takes place for closing returns but this is not large enough to cause over-reaction. Originality/value – This is the first study to examine the impact of a single security futures contract on the speed of spot market price discovery.
Resumo:
When applying multivariate analysis techniques in information systems and social science disciplines, such as management information systems (MIS) and marketing, the assumption that the empirical data originate from a single homogeneous population is often unrealistic. When applying a causal modeling approach, such as partial least squares (PLS) path modeling, segmentation is a key issue in coping with the problem of heterogeneity in estimated cause-and-effect relationships. This chapter presents a new PLS path modeling approach which classifies units on the basis of the heterogeneity of the estimates in the inner model. If unobserved heterogeneity significantly affects the estimated path model relationships on the aggregate data level, the methodology will allow homogenous groups of observations to be created that exhibit distinctive path model estimates. The approach will, thus, provide differentiated analytical outcomes that permit more precise interpretations of each segment formed. An application on a large data set in an example of the American customer satisfaction index (ACSI) substantiates the methodology’s effectiveness in evaluating PLS path modeling results.
Resumo:
Levels of lignin and hydroxycinnamic acid wall components in three genera of forage grasses (Lolium,Festuca and Dactylis) have been accurately predicted by Fourier-transform infrared spectroscopy using partial least squares models correlated to analytical measurements. Different models were derived that predicted the concentrations of acid detergent lignin, total hydroxycinnamic acids, total ferulate monomers plus dimers, p-coumarate and ferulate dimers in independent spectral test data from methanol extracted samples of perennial forage grass with accuracies of 92.8%, 86.5%, 86.1%, 59.7% and 84.7% respectively, and analysis of model projection scores showed that the models relied generally on spectral features that are known absorptions of these compounds. Acid detergent lignin was predicted in samples of two species of energy grass, (Phalaris arundinacea and Pancium virgatum) with an accuracy of 84.5%.
Resumo:
The development of new products in today's marketing environment is generally accepted as a requirement for the continual growth and prosperity of organisations. The literature is consequently rich with information on the development of various aspects of good products. In the case of service industries, it can be argued that new service product development is of as least equal importance as it is to organisations that produce tangible goods products. Unlike the new goods product literature, the literature on service marketing practices, and in particular, new service product development, is relatively sparse. The main purpose of this thesis is to examine a number of aspects of new service product development practice with respect to financial services and specifically, credit card financial services. The empirical investigation utilises both a case study and a survey approach, to examine aspects of new service product development industry practice relating specifically to gaps and deficiencies in the literature with respect to the financial service industry. The findings of the empirical work are subsequently examined in the context in which they provide guidance and support for a new normative new service product development model. The study examines the UK credit card financial service product sector as an industry case study and perspective. The findings of the field work reveal that the new service product development process is still evolving, and that in the case of credit card financial services can be seen as a well-structured and well-documented process. New product development can also be seen as an incremental, complex, interactive and continuous process which has been applied in a variety of ways. A number of inferences are subsequently presented.
Resumo:
Semantic Web Service, one of the most significant research areas within the Semantic Web vision, has attracted increasing attention from both the research community and industry. The Web Service Modelling Ontology (WSMO) has been proposed as an enabling framework for the total/partial automation of the tasks (e.g., discovery, selection, composition, mediation, execution, monitoring, etc.) involved in both intra- and inter-enterprise integration of Web services. To support the standardisation and tool support of WSMO, a formal model of the language is highly desirable. As several variants of WSMO have been proposed by the WSMO community, which are still under development, the syntax and semantics of WSMO should be formally defined to facilitate easy reuse and future development. In this paper, we present a formal Object-Z formal model of WSMO, where different aspects of the language have been precisely defined within one unified framework. This model not only provides a formal unambiguous model which can be used to develop tools and facilitate future development, but as demonstrated in this paper, can be used to identify and eliminate errors present in existing documentation.
Resumo:
This paper investigates the impact of HRM systems on organisational performance in a sample of 178 Greek manufacturing organisations. The results show strong support for the ‘universalistic’ model, highlighting that both resource-development and reward-relations systems are positively related with organisational performance. The results also show weak and partial support for the ‘contingency model’, i.e., resourcedevelopment and reward-relations systems are contingent on the business strategies of quality, innovation, and cost in determining organisational efficiency. The study concludes that the universalistic and contingency perspectives are not necessarily mutually exclusive but on the contrary are in some cases complementary.
Resumo:
We uncover high persistence in credit spread series that can obscure the relationship between the theoretical determinants of credit risk and observed credit spreads. We use a Markovswitching model, which also captures the stability (low frequency changes) of credit ratings, to show why credit spreads may continue to respond to past levels of credit risk, even though the state of the economy has changed. A bivariate model of credit spreads and either macroeconomic activity or equity market volatility detects large and significant correlations that are consistent with theory but have not been observed in previous studies. © 2010 Nova Science Publishers, Inc. All rights reserved.
Resumo:
The accurate identification of T-cell epitopes remains a principal goal of bioinformatics within immunology. As the immunogenicity of peptide epitopes is dependent on their binding to major histocompatibility complex (MHC) molecules, the prediction of binding affinity is a prerequisite to the reliable prediction of epitopes. The iterative self-consistent (ISC) partial-least-squares (PLS)-based additive method is a recently developed bioinformatic approach for predicting class II peptide−MHC binding affinity. The ISC−PLS method overcomes many of the conceptual difficulties inherent in the prediction of class II peptide−MHC affinity, such as the binding of a mixed population of peptide lengths due to the open-ended class II binding site. The method has applications in both the accurate prediction of class II epitopes and the manipulation of affinity for heteroclitic and competitor peptides. The method is applied here to six class II mouse alleles (I-Ab, I-Ad, I-Ak, I-As, I-Ed, and I-Ek) and included peptides up to 25 amino acids in length. A series of regression equations highlighting the quantitative contributions of individual amino acids at each peptide position was established. The initial model for each allele exhibited only moderate predictivity. Once the set of selected peptide subsequences had converged, the final models exhibited a satisfactory predictive power. Convergence was reached between the 4th and 17th iterations, and the leave-one-out cross-validation statistical terms - q2, SEP, and NC - ranged between 0.732 and 0.925, 0.418 and 0.816, and 1 and 6, respectively. The non-cross-validated statistical terms r2 and SEE ranged between 0.98 and 0.995 and 0.089 and 0.180, respectively. The peptides used in this study are available from the AntiJen database (http://www.jenner.ac.uk/AntiJen). The PLS method is available commercially in the SYBYL molecular modeling software package. The resulting models, which can be used for accurate T-cell epitope prediction, will be made freely available online (http://www.jenner.ac.uk/MHCPred).
Resumo:
Motivation: The immunogenicity of peptides depends on their ability to bind to MHC molecules. MHC binding affinity prediction methods can save significant amounts of experimental work. The class II MHC binding site is open at both ends, making epitope prediction difficult because of the multiple binding ability of long peptides. Results: An iterative self-consistent partial least squares (PLS)-based additive method was applied to a set of 66 pep- tides no longer than 16 amino acids, binding to DRB1*0401. A regression equation containing the quantitative contributions of the amino acids at each of the nine positions was generated. Its predictability was tested using two external test sets which gave r pred =0.593 and r pred=0.655, respectively. Furthermore, it was benchmarked using 25 known T-cell epitopes restricted by DRB1*0401 and we compared our results with four other online predictive methods. The additive method showed the best result finding 24 of the 25 T-cell epitopes. Availability: Peptides used in the study are available from http://www.jenner.ac.uk/JenPep. The PLS method is available commercially in the SYBYL molecular modelling software package. The final model for affinity prediction of peptides binding to DRB1*0401 molecule is available at http://www.jenner.ac.uk/MHCPred. Models developed for DRB1*0101 and DRB1*0701 also are available in MHC- Pred
Resumo:
Abstract A new LIBS quantitative analysis method based on analytical line adaptive selection and Relevance Vector Machine (RVM) regression model is proposed. First, a scheme of adaptively selecting analytical line is put forward in order to overcome the drawback of high dependency on a priori knowledge. The candidate analytical lines are automatically selected based on the built-in characteristics of spectral lines, such as spectral intensity, wavelength and width at half height. The analytical lines which will be used as input variables of regression model are determined adaptively according to the samples for both training and testing. Second, an LIBS quantitative analysis method based on RVM is presented. The intensities of analytical lines and the elemental concentrations of certified standard samples are used to train the RVM regression model. The predicted elemental concentration analysis results will be given with a form of confidence interval of probabilistic distribution, which is helpful for evaluating the uncertainness contained in the measured spectra. Chromium concentration analysis experiments of 23 certified standard high-alloy steel samples have been carried out. The multiple correlation coefficient of the prediction was up to 98.85%, and the average relative error of the prediction was 4.01%. The experiment results showed that the proposed LIBS quantitative analysis method achieved better prediction accuracy and better modeling robustness compared with the methods based on partial least squares regression, artificial neural network and standard support vector machine.
Resumo:
Data fluctuation in multiple measurements of Laser Induced Breakdown Spectroscopy (LIBS) greatly affects the accuracy of quantitative analysis. A new LIBS quantitative analysis method based on the Robust Least Squares Support Vector Machine (RLS-SVM) regression model is proposed. The usual way to enhance the analysis accuracy is to improve the quality and consistency of the emission signal, such as by averaging the spectral signals or spectrum standardization over a number of laser shots. The proposed method focuses more on how to enhance the robustness of the quantitative analysis regression model. The proposed RLS-SVM regression model originates from the Weighted Least Squares Support Vector Machine (WLS-SVM) but has an improved segmented weighting function and residual error calculation according to the statistical distribution of measured spectral data. Through the improved segmented weighting function, the information on the spectral data in the normal distribution will be retained in the regression model while the information on the outliers will be restrained or removed. Copper elemental concentration analysis experiments of 16 certified standard brass samples were carried out. The average value of relative standard deviation obtained from the RLS-SVM model was 3.06% and the root mean square error was 1.537%. The experimental results showed that the proposed method achieved better prediction accuracy and better modeling robustness compared with the quantitative analysis methods based on Partial Least Squares (PLS) regression, standard Support Vector Machine (SVM) and WLS-SVM. It was also demonstrated that the improved weighting function had better comprehensive performance in model robustness and convergence speed, compared with the four known weighting functions.