973 resultados para score test information matrix artificial regression
Resumo:
Using the classical Parzen window (PW) estimate as the desired response, the kernel density estimation is formulated as a regression problem and the orthogonal forward regression technique is adopted to construct sparse kernel density (SKD) estimates. The proposed algorithm incrementally minimises a leave-one-out test score to select a sparse kernel model, and a local regularisation method is incorporated into the density construction process to further enforce sparsity. The kernel weights of the selected sparse model are finally updated using the multiplicative nonnegative quadratic programming algorithm, which ensures the nonnegative and unity constraints for the kernel weights and has the desired ability to reduce the model size further. Except for the kernel width, the proposed method has no other parameters that need tuning, and the user is not required to specify any additional criterion to terminate the density construction procedure. Several examples demonstrate the ability of this simple regression-based approach to effectively construct a SKID estimate with comparable accuracy to that of the full-sample optimised PW density estimate. (c) 2007 Elsevier B.V. All rights reserved.
Resumo:
Using the classical Parzen window (PW) estimate as the target function, the sparse kernel density estimator is constructed in a forward-constrained regression (FCR) manner. The proposed algorithm selects significant kernels one at a time, while the leave-one-out (LOO) test score is minimized subject to a simple positivity constraint in each forward stage. The model parameter estimation in each forward stage is simply the solution of jackknife parameter estimator for a single parameter, subject to the same positivity constraint check. For each selected kernels, the associated kernel width is updated via the Gauss-Newton method with the model parameter estimate fixed. The proposed approach is simple to implement and the associated computational cost is very low. Numerical examples are employed to demonstrate the efficacy of the proposed approach.
Resumo:
A generalized or tunable-kernel model is proposed for probability density function estimation based on an orthogonal forward regression procedure. Each stage of the density estimation process determines a tunable kernel, namely, its center vector and diagonal covariance matrix, by minimizing a leave-one-out test criterion. The kernel mixing weights of the constructed sparse density estimate are finally updated using the multiplicative nonnegative quadratic programming algorithm to ensure the nonnegative and unity constraints, and this weight-updating process additionally has the desired ability to further reduce the model size. The proposed tunable-kernel model has advantages, in terms of model generalization capability and model sparsity, over the standard fixed-kernel model that restricts kernel centers to the training data points and employs a single common kernel variance for every kernel. On the other hand, it does not optimize all the model parameters together and thus avoids the problems of high-dimensional ill-conditioned nonlinear optimization associated with the conventional finite mixture model. Several examples are included to demonstrate the ability of the proposed novel tunable-kernel model to effectively construct a very compact density estimate accurately.
Resumo:
This paper derives an efficient algorithm for constructing sparse kernel density (SKD) estimates. The algorithm first selects a very small subset of significant kernels using an orthogonal forward regression (OFR) procedure based on the D-optimality experimental design criterion. The weights of the resulting sparse kernel model are then calculated using a modified multiplicative nonnegative quadratic programming algorithm. Unlike most of the SKD estimators, the proposed D-optimality regression approach is an unsupervised construction algorithm and it does not require an empirical desired response for the kernel selection task. The strength of the D-optimality OFR is owing to the fact that the algorithm automatically selects a small subset of the most significant kernels related to the largest eigenvalues of the kernel design matrix, which counts for the most energy of the kernel training data, and this also guarantees the most accurate kernel weight estimate. The proposed method is also computationally attractive, in comparison with many existing SKD construction algorithms. Extensive numerical investigation demonstrates the ability of this regression-based approach to efficiently construct a very sparse kernel density estimate with excellent test accuracy, and our results show that the proposed method compares favourably with other existing sparse methods, in terms of test accuracy, model sparsity and complexity, for constructing kernel density estimates.
Resumo:
Chatterbox Challenge is an annual web-based contest for artificial conversational systems, ACE. The 2010 instantiation was the tenth consecutive contest held between March and June in the 60th year following the publication of Alan Turing’s influential disquisition ‘computing machinery and intelligence’. Loosely based on Turing’s viva voca interrogator-hidden witness imitation game, a thought experiment to ascertain a machine’s capacity to respond satisfactorily to unrestricted questions, the contest provides a platform for technology comparison and evaluation. This paper provides an insight into emotion content in the entries since the 2005 Chatterbox Challenge. The authors find that synthetic textual systems, none of which are backed by academic or industry funding, are, on the whole and more than half a century since Weizenbaum’s natural language understanding experiment, little further than Eliza in terms of expressing emotion in dialogue. This may be a failure on the part of the academic AI community for ignoring the Turing test as an engineering challenge.
Resumo:
Objectives: Our objective was to test the performance of CA125 in classifying serum samples from a cohort of malignant and benign ovarian cancers and age-matched healthy controls and to assess whether combining information from matrix-assisted laser desorption/ionization (MALDI) time-of-flight profiling could improve diagnostic performance. Materials and Methods: Serum samples from women with ovarian neoplasms and healthy volunteers were subjected to CA125 assay and MALDI time-of-flight mass spectrometry (MS) profiling. Models were built from training data sets using discriminatory MALDI MS peaks in combination with CA125 values and tested their ability to classify blinded test samples. These were compared with models using CA125 threshold levels from 193 patients with ovarian cancer, 290 with benign neoplasm, and 2236 postmenopausal healthy controls. Results: Using a CA125 cutoff of 30 U/mL, an overall sensitivity of 94.8% (96.6% specificity) was obtained when comparing malignancies versus healthy postmenopausal controls, whereas a cutoff of 65 U/mL provided a sensitivity of 83.9% (99.6% specificity). High classification accuracies were obtained for early-stage cancers (93.5% sensitivity). Reasons for high accuracies include recruitment bias, restriction to postmenopausal women, and inclusion of only primary invasive epithelial ovarian cancer cases. The combination of MS profiling information with CA125 did not significantly improve the specificity/accuracy compared with classifications on the basis of CA125 alone. Conclusions: We report unexpectedly good performance of serum CA125 using threshold classification in discriminating healthy controls and women with benign masses from those with invasive ovarian cancer. This highlights the dependence of diagnostic tests on the characteristics of the study population and the crucial need for authors to provide sufficient relevant details to allow comparison. Our study also shows that MS profiling information adds little to diagnostic accuracy. This finding is in contrast with other reports and shows the limitations of serum MS profiling for biomarker discovery and as a diagnostic tool
Resumo:
This study investigated whether children’s fears could be un-learned using Rachman’s indirect pathways for learning fear. We hypothesised that positive information and modelling a non-anxious response are effective methods of un-learning fears acquired through verbal information. One hundred and seven children aged 6–8 years received negative information about one animal and no information about another. Fear beliefs and behavioural avoidance were measured. Children were randomised to receive positive verbal information, modelling, or a control task. Fear beliefs and behavioural avoidance were measured again. Positive information and modelling led to lower fear beliefs and behavioural avoidance than the control condition. Positive information was more effective than modelling in reducing fear beliefs and both methods significantly reduced behavioural avoidance. The results support Rachman’s indirect pathways as viable fear un-learning pathways and supports associative learning theories.
Resumo:
We evaluate a number of real estate sentiment indices to ascertain current and forward-looking information content that may be useful for forecasting demand and supply activities. Analyzing the dynamic relationships within a Vector Auto-Regression (VAR) framework and using the quarterly US data over 1988-2010, we test the efficacy of several sentiment measures by comparing them with other coincident economic indicators. Overall, our analysis suggests that the sentiment in real estate convey valuable information that can help predict changes in real estate returns. These findings have important implications for investment decisions, from consumers' as well as institutional investors' perspectives.
Resumo:
BACKGROUND: Using continuing professional development (CPD) as part of the revalidation of pharmacy professionals has been proposed in the UK but not implemented. We developed a CPD Outcomes Framework (‘the framework’) for scoring CPD records, where the score range was -100 to +150 based on demonstrable relevance and impact of the CPD on practice. OBJECTIVE: This exploratory study aimed to test the outcome of training people to use the framework, through distance-learning material (active intervention), by comparing CPD scores before and after training. SETTING: Pharmacy professionals were recruited in the UK in Reading, Banbury, Southampton, Kingston-upon-Thames and Guildford in 2009. METHOD: We conducted a randomised, double-blinded, parallel-group, before and after study. The control group simply received information on new CPD requirements through the post; the active intervention group also received the framework and associated training. Altogether 48 participants (25 control, 23 active) completed the study. All participants submitted CPD records to the research team before and after receiving the posted resources. The records (n=226) were scored blindly by the researchers using the framework. A subgroup of CPD records (n=96) submitted first (before-stage) and rewritten (after-stage) were analysed separately. MAIN OUTCOME MEASURE: Scores for CPD records received before and after distributing group-dependent material through the post. RESULTS: Using a linear-regression model both analyses found an increase in CPD scores in favour of the active intervention group. For the complete set of records, the effect was a mean difference of 9.9 (95% CI = 0.4 to 19.3), p-value = 0.04. For the subgroup of rewritten records, the effect was a mean difference of 17.3 (95% CI = 5.6 to 28.9), p-value = 0.0048. CONCLUSION: The intervention improved participants’ CPD behaviour. Training pharmacy professionals to use the framework resulted in better CPD activities and CPD records, potentially helpful for revalidation of pharmacy professionals. IMPACT: • Using a bespoke Continuing Professional Development outcomes framework improves the value of pharmacy professionals’ CPD activities and CPD records, with the potential to improve patient care. • The CPD outcomes framework could be helpful to pharmacy professionals internationally who want to improve the quality of their CPD activities and CPD records. • Regulators and officials across Europe and beyond can assess the suitability of the CPD outcomes framework for use in pharmacy CPD and revalidation in their own setting.
Resumo:
Introduction: The inability to distinguish periapical cysts from granulomas before performing root canal treatment leads to uncertainty in treatment outcomes because cysts have lower healing rates. Searching for differential expression of molecules within cysts or granulomas could provide information with regard to the identity of the lesion or suggest mechanistic differences that may form the basis for future therapeutic intervention. Thus, we investigated whether granulomas and cysts exhibit differential expression of extracellular matrix (ECM) molecules. Methods: Human periapical granulomas, periapical cysts, and healthy periodontal ligament tissues were used to investigate the differential expression of ECM molecules by microarray analysis. Because matrix metalloproteinases (MMP) showed the highest differential expression in the microarray analysis, MMPs were further examined by in situ zymography and immunohistochemistry. Data were analyzed by using one-way analysis of variance followed by the Tu-key test. Results: We observed that cysts and granulomas differentially expressed several ECM molecules, especially those from the MMP family. Compared with cysts, granulomas exhibited higher MMP enzymatic activity in areas stained for MMP-9. These areas were composed of polymorphonuclear cells (PMNs) in contrast to cysts. Similarly, MMP-13 was expressed by a greater number of cells in granulomas compared with cysts. Conclusion: Our findings indicate that high enzymatic MIMP activity in PMNs together with MMP-9 and MMP-13 stained cells could be a molecular signature of granulomas unlike periapical cysts. (J Endod 2009;35:1234-1242)
Resumo:
The matrix-tolerance hypothesis suggests that the most abundant species in the inter-habitat matrix would be less vulnerable to their habitat fragmentation. This model was tested with leaf-litter frogs in the Atlantic Forest where the fragmentation process is older and more severe than in the Amazon, where the model was first developed. Frog abundance data from the agricultural matrix, forest fragments and continuous forest localities were used. We found an expected negative correlation between the abundance of frogs in the matrix and their vulnerability to fragmentation, however, results varied with fragment size and species traits. Smaller fragments exhibited stronger matrix-vulnerability correlation than intermediate fragments, while no significant relation was observed for large fragments. Moreover, some species that avoid the matrix were not sensitive to a decrease in the patch size, and the opposite was also true, indicating significant differences with that expected from the model. Most of the species that use the matrix were forest species with aquatic larvae development, but those species do not necessarily respond to fragmentation or fragment size, and thus affect more intensively the strengthen of the expected relationship. Therefore, the main relationship expected by the matrix-tolerance hypothesis was observed in the Atlantic Forest; however we noted that the prediction of this hypothesis can be substantially affected by the size of the fragments, and by species traits. We propose that matrix-tolerance model should be broadened to become a more effective model, including other patch characteristics, particularly fragment size, and individual species traits (e. g., reproductive mode and habitat preference).
Resumo:
We have considered a Bayesian approach for the nonlinear regression model by replacing the normal distribution on the error term by some skewed distributions, which account for both skewness and heavy tails or skewness alone. The type of data considered in this paper concerns repeated measurements taken in time on a set of individuals. Such multiple observations on the same individual generally produce serially correlated outcomes. Thus, additionally, our model does allow for a correlation between observations made from the same individual. We have illustrated the procedure using a data set to study the growth curves of a clinic measurement of a group of pregnant women from an obstetrics clinic in Santiago, Chile. Parameter estimation and prediction were carried out using appropriate posterior simulation schemes based in Markov Chain Monte Carlo methods. Besides the deviance information criterion (DIC) and the conditional predictive ordinate (CPO), we suggest the use of proper scoring rules based on the posterior predictive distribution for comparing models. For our data set, all these criteria chose the skew-t model as the best model for the errors. These DIC and CPO criteria are also validated, for the model proposed here, through a simulation study. As a conclusion of this study, the DIC criterion is not trustful for this kind of complex model.
Resumo:
The Birnbaum-Saunders regression model is commonly used in reliability studies. We derive a simple matrix formula for second-order covariances of maximum-likelihood estimators in this class of models. The formula is quite suitable for computer implementation, since it involves only simple operations on matrices and vectors. Some simulation results show that the second-order covariances can be quite pronounced in small to moderate sample sizes. We also present empirical applications.
Resumo:
This paper provides general matrix formulas for computing the score function, the (expected and observed) Fisher information and the A matrices (required for the assessment of local influence) for a quite general model which includes the one proposed by Russo et al. (2009). Additionally, we also present an expression for the generalized leverage on fixed and random effects. The matrix formulation has notational advantages, since despite the complexity of the postulated model, all general formulas are compact, clear and have nice forms. (C) 2010 Elsevier B.V. All rights reserved.