899 resultados para Models performance


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Several studies have reported high performance of simple decision heuristics multi-attribute decision making. In this paper, we focus on situations where attributes are binary and analyze the performance of Deterministic-Elimination-By-Aspects (DEBA) and similar decision heuristics. We consider non-increasing weights and two probabilistic models for the attribute values: one where attribute values are independent Bernoulli randomvariables; the other one where they are binary random variables with inter-attribute positive correlations. Using these models, we show that good performance of DEBA is explained by the presence of cumulative as opposed to simple dominance. We therefore introduce the concepts of cumulative dominance compliance and fully cumulative dominance compliance and show that DEBA satisfies those properties. We derive a lower bound with which cumulative dominance compliant heuristics will choose a best alternative and show that, even with many attributes, this is not small. We also derive an upper bound for the expected loss of fully cumulative compliance heuristics and show that this is moderateeven when the number of attributes is large. Both bounds are independent of the values ofthe weights.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Most methods for small-area estimation are based on composite estimators derived from design- or model-based methods. A composite estimator is a linear combination of a direct and an indirect estimator with weights that usually depend on unknown parameters which need to be estimated. Although model-based small-area estimators are usually based on random-effects models, the assumption of fixed effects is at face value more appropriate.Model-based estimators are justified by the assumption of random (interchangeable) area effects; in practice, however, areas are not interchangeable. In the present paper we empirically assess the quality of several small-area estimators in the setting in which the area effects are treated as fixed. We consider two settings: one that draws samples from a theoretical population, and another that draws samples from an empirical population of a labor force register maintained by the National Institute of Social Security (NISS) of Catalonia. We distinguish two types of composite estimators: a) those that use weights that involve area specific estimates of bias and variance; and, b) those that use weights that involve a common variance and a common squared bias estimate for all the areas. We assess their precision and discuss alternatives to optimizing composite estimation in applications.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper theoretically and empirically documents a puzzle that arises when an RBC economy with a job matching function is used to model unemployment. The standard model can generate sufficiently large cyclical fluctuations in unemployment, or a sufficiently small response of unemployment to labor market policies, but it cannot do both. Variable search and separation, finite UI benefit duration, efficiency wages, and capital all fail to resolve this puzzle. However, either sticky wages or match-specific productivity shocks can improve the model's performance by making the firm's flow of surplus more procyclical, which makes hiring more procyclical too.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Revenue management (RM) is a complicated business process that can best be described ascontrol of sales (using prices, restrictions, or capacity), usually using software as a tool to aiddecisions. RM software can play a mere informative role, supplying analysts with formatted andsummarized data who use it to make control decisions (setting a price or allocating capacity fora price point), or, play a deeper role, automating the decisions process completely, at the otherextreme. The RM models and algorithms in the academic literature by and large concentrateon the latter, completely automated, level of functionality.A firm considering using a new RM model or RM system needs to evaluate its performance.Academic papers justify the performance of their models using simulations, where customerbooking requests are simulated according to some process and model, and the revenue perfor-mance of the algorithm compared to an alternate set of algorithms. Such simulations, whilean accepted part of the academic literature, and indeed providing research insight, often lackcredibility with management. Even methodologically, they are usually awed, as the simula-tions only test \within-model" performance, and say nothing as to the appropriateness of themodel in the first place. Even simulations that test against alternate models or competition arelimited by their inherent necessity on fixing some model as the universe for their testing. Theseproblems are exacerbated with RM models that attempt to model customer purchase behav-ior or competition, as the right models for competitive actions or customer purchases remainsomewhat of a mystery, or at least with no consensus on their validity.How then to validate a model? Putting it another way, we want to show that a particularmodel or algorithm is the cause of a certain improvement to the RM process compared to theexisting process. We take care to emphasize that we want to prove the said model as the causeof performance, and to compare against a (incumbent) process rather than against an alternatemodel.In this paper we describe a \live" testing experiment that we conducted at Iberia Airlineson a set of flights. A set of competing algorithms control a set of flights during adjacentweeks, and their behavior and results are observed over a relatively long period of time (9months). In parallel, a group of control flights were managed using the traditional mix of manualand algorithmic control (incumbent system). Such \sandbox" testing, while common at manylarge internet search and e-commerce companies is relatively rare in the revenue managementarea. Sandbox testing has an undisputable model of customer behavior but the experimentaldesign and analysis of results is less clear. In this paper we describe the philosophy behind theexperiment, the organizational challenges, the design and setup of the experiment, and outlinethe analysis of the results. This paper is a complement to a (more technical) related paper thatdescribes the econometrics and statistical analysis of the results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper discusses inference in self exciting threshold autoregressive (SETAR)models. Of main interest is inference for the threshold parameter. It iswell-known that the asymptotics of the corresponding estimator depend uponwhether the SETAR model is continuous or not. In the continuous case, thelimiting distribution is normal and standard inference is possible. Inthe discontinuous case, the limiting distribution is non-normal and cannotbe estimated consistently. We show valid inference can be drawn by theuse of the subsampling method. Moreover, the method can even be extendedto situations where the (dis)continuity of the model is unknown. In thiscase, also the inference for the regression parameters of the modelbecomes difficult and subsampling can be used advantageously there aswell. In addition, we consider an hypothesis test for the continuity ofthe SETAR model. A simulation study examines small sample performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Research on judgment and decision making presents a confusing picture of human abilities. For example, much research has emphasized the dysfunctional aspects of judgmental heuristics, and yet, other findings suggest that these can be highly effective. A further line of research has modeled judgment as resulting from as if linear models. This paper illuminates the distinctions in these approaches by providing a common analytical framework based on the central theoretical premise that understanding human performance requires specifying how characteristics of the decision rules people use interact with the demands of the tasks they face. Our work synthesizes the analytical tools of lens model research with novel methodology developed to specify the effectiveness of heuristics in different environments and allows direct comparisons between the different approaches. We illustrate with both theoretical analyses and simulations. We further link our results to the empirical literature by a meta-analysis of lens model studies and estimate both human andheuristic performance in the same tasks. Our results highlight the trade-off betweenlinear models and heuristics. Whereas the former are cognitively demanding, the latterare simple to use. However, they require knowledge and thus maps of when andwhich heuristic to employ.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this paper is to compare the performance of twopredictive radiological models, logistic regression (LR) and neural network (NN), with five different resampling methods. One hundred and sixty-seven patients with proven calvarial lesions as the only known disease were enrolled. Clinical and CT data were used for LR and NN models. Both models were developed with cross validation, leave-one-out and three different bootstrap algorithms. The final results of each model were compared with error rate and the area under receiver operating characteristic curves (Az). The neural network obtained statistically higher Az than LR with cross validation. The remaining resampling validation methods did not reveal statistically significant differences between LR and NN rules. The neural network classifier performs better than the one based on logistic regression. This advantage is well detected by three-fold cross-validation, but remains unnoticed when leave-one-out or bootstrap algorithms are used.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Combined positron emission tomography and computed tomography (PET/CT) scanners play a major role in medicine for in vivo imaging in an increasing number of diseases in oncology, cardiology, neurology, and psychiatry. With the advent of short-lived radioisotopes other than 18F and newer scanners, there is a need to optimize radioisotope activity and acquisition protocols, as well as to compare scanner performances on an objective basis. The Discovery-LS (D-LS) was among the first clinical PET/CT scanners to be developed and has been extensively characterized with older National Electrical Manufacturer Association (NEMA) NU 2-1994 standards. At the time of publication of the latest version of the standards (NU 2-2001) that have been adapted for whole-body imaging under clinical conditions, more recent models from the same manufacturer, i.e., Discovery-ST (D-ST) and Discovery-STE (D-STE), were commercially available. We report on the full characterization both in the two- and three-dimensional acquisition mode of the D-LS according to latest NEMA NU 2-2001 standards (spatial resolution, sensitivity, count rate performance, accuracy of count losses, and random coincidence correction and image quality), as well as a detailed comparison with the newer D-ST widely used and whose characteristics are already published.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We modeled work performance as outcomes of individual-differences mediated by technical performance. Beyond the "usual suspects" (e.g., general mental ability, and personality), we also measured the ethical development of participants (n = 460). We surmised that ethical development - which has not been extensively studied as a predictor of work performance while controlling for established predictors - captures unique variance in both technical and work performance. Results demonstrated incremental validity for ethical development in predicting technical performance, which in turn predicted work performance. The indirect effect of ethical development was significant too. Our results highlight the importance of process models of performance, which include proximal as well as distal individual differences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present study provides a comprehensive view of (a) the time dynamics of the psychophysiological responding in performing music students (n = 66) before, during, and after a private and a public performance and (b) the moderating effect of music performance anxiety (MPA). Heart rate (HR), minute ventilation (VE), and all affective and somatic self-report variables increased in the public session compared to the private session. Furthermore, the activation of all variables was stronger during the performances than before or after. Differences between phases were larger in the public than in the private session for HR, VE, total breath duration, anxiety, and trembling. Furthermore, while higher MPA scores were associated with higher scores and with larger changes between sessions and phases for self-reports, this association was less coherent for physiological variables. Finally, self-reported intra-individual performance improvements or deteriorations were not associated with MPA. This study makes a novel contribution by showing how the presence of an audience influences low- and high-anxious musicians' psychophysiological responding before, during and after performing. Overall, the findings are more consistent with models of anxiety that emphasize the importance of cognitive rather than physiological factors in MPA.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper we analyse, using Monte Carlo simulation, the possible consequences of incorrect assumptions on the true structure of the random effects covariance matrix and the true correlation pattern of residuals, over the performance of an estimation method for nonlinear mixed models. The procedure under study is the well known linearization method due to Lindstrom and Bates (1990), implemented in the nlme library of S-Plus and R. Its performance is studied in terms of bias, mean square error (MSE), and true coverage of the associated asymptotic confidence intervals. Ignoring other criteria like the convenience of avoiding over parameterised models, it seems worst to erroneously assume some structure than do not assume any structure when this would be adequate.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

1. Identifying the boundary of a species' niche from observational and environmental data is a common problem in ecology and conservation biology and a variety of techniques have been developed or applied to model niches and predict distributions. Here, we examine the performance of some pattern-recognition methods as ecological niche models (ENMs). Particularly, one-class pattern recognition is a flexible and seldom used methodology for modelling ecological niches and distributions from presence-only data. The development of one-class methods that perform comparably to two-class methods (for presence/absence data) would remove modelling decisions about sampling pseudo-absences or background data points when absence points are unavailable. 2. We studied nine methods for one-class classification and seven methods for two-class classification (five common to both), all primarily used in pattern recognition and therefore not common in species distribution and ecological niche modelling, across a set of 106 mountain plant species for which presence-absence data was available. We assessed accuracy using standard metrics and compared trade-offs in omission and commission errors between classification groups as well as effects of prevalence and spatial autocorrelation on accuracy. 3. One-class models fit to presence-only data were comparable to two-class models fit to presence-absence data when performance was evaluated with a measure weighting omission and commission errors equally. One-class models were superior for reducing omission errors (i.e. yielding higher sensitivity), and two-classes models were superior for reducing commission errors (i.e. yielding higher specificity). For these methods, spatial autocorrelation was only influential when prevalence was low. 4. These results differ from previous efforts to evaluate alternative modelling approaches to build ENM and are particularly noteworthy because data are from exhaustively sampled populations minimizing false absence records. Accurate, transferable models of species' ecological niches and distributions are needed to advance ecological research and are crucial for effective environmental planning and conservation; the pattern-recognition approaches studied here show good potential for future modelling studies. This study also provides an introduction to promising methods for ecological modelling inherited from the pattern-recognition discipline.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Research presented herein describes an application of a newly developed material called Ultra-High Performance Concrete (UHPC) to a single-span bridge. The two primary objectives of this research were to develop a shear design procedure for possible code adoption and to provide a performance evaluation to ensure the viability of the first UHPC bridge in the United States. Two other secondary objectives included defining of material properties and understanding of flexural behavior of a UHPC bridge girder. In order to obtain information in these areas, several tests were carried out including material testing, large-scale laboratory flexure testing, large-scale laboratory shear testing, large-scale laboratory flexure-shear testing, small-scale laboratory shear testing, and field testing of a UHPC bridge. Experimental and analytical results of the described tests are presented. Analytical models to understand the flexure and shear behavior of UHPC members were developed using iterative computer based procedures. Previous research is referenced explaining a simplified flexural design procedure and a simplified pure shear design procedure. This work describes a shear design procedure based on the Modified Compression Field Theory (MCFT) which can be used in the design of UHPC members. Conclusions are provided regarding the viability of the UHPC bridge and recommendations are made for future research.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Is it possible to build predictive models (PMs) of soil particle-size distribution (psd) in a region with complex geology and a young and unstable land-surface? The main objective of this study was to answer this question. A set of 339 soil samples from a small slope catchment in Southern Brazil was used to build PMs of psd in the surface soil layer. Multiple linear regression models were constructed using terrain attributes (elevation, slope, catchment area, convergence index, and topographic wetness index). The PMs explained more than half of the data variance. This performance is similar to (or even better than) that of the conventional soil mapping approach. For some size fractions, the PM performance can reach 70 %. Largest uncertainties were observed in geologically more complex areas. Therefore, significant improvements in the predictions can only be achieved if accurate geological data is made available. Meanwhile, PMs built on terrain attributes are efficient in predicting the particle-size distribution (psd) of soils in regions of complex geology.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Modern agriculture techniques have a great impact on crops and soil quality, especially by the increased machinery traffic and weight. Several devices have been developed for determining soil properties in the field, aimed at managing compacted areas. Penetrometry is a widely used technique; however, there are several types of penetrometers, which have different action modes that can affect the soil resistance measurement. The objective of this study was to compare the functionality of two penetrometry methods (manual and automated mode) in the field identification of compacted, highly mechanized sugarcane areas, considering the influence of soil water volumetric content (θ) on soil penetration resistance (PR). Three sugarcane fields on a Rhodic Eutrudrox were chosen, under a sequence of harvest systems: one manual harvest (1ManH), one mechanized harvest (1MH) and three mechanized harvests (3MH). The different degrees of mechanization were associated to cumulative compaction processes. An electronic penetrometer was used on PR measurements, so that the rod was introduced into the soil by hand (Manual) and by an electromechanical motor (Auto). The θ was measured in the field with a soil moisture sensor. Results showed an effect of θ on PR measurements and that regression models must be used to correct data before comparing harvesting systems. The rod introduction modes resulted in different mean PR values, where the "Manual" overestimated PR compared to the "Auto" mode at low θ.