63 resultados para Predictive values


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Data mining is the process to identify valid, implicit, previously unknown, potentially useful and understandable information from large databases. It is an important step in the process of knowledge discovery in databases, (Olaru & Wehenkel, 1999). In a data mining process, input data can be structured, seme-structured, or unstructured. Data can be in text, categorical or numerical values. One of the important characteristics of data mining is its ability to deal data with large volume, distributed, time variant, noisy, and high dimensionality. A large number of data mining algorithms have been developed for different applications. For example, association rules mining can be useful for market basket problems, clustering algorithms can be used to discover trends in unsupervised learning problems, classification algorithms can be applied in decision-making problems, and sequential and time series mining algorithms can be used in predicting events, fault detection, and other supervised learning problems (Vapnik, 1999). Classification is among the most important tasks in the data mining, particularly for data mining applications into engineering fields. Together with regression, classification is mainly for predictive modelling. So far, there have been a number of classification algorithms in practice. According to (Sebastiani, 2002), the main classification algorithms can be categorized as: decision tree and rule based approach such as C4.5 (Quinlan, 1996); probability methods such as Bayesian classifier (Lewis, 1998); on-line methods such as Winnow (Littlestone, 1988) and CVFDT (Hulten 2001), neural networks methods (Rumelhart, Hinton & Wiliams, 1986); example-based methods such as k-nearest neighbors (Duda & Hart, 1973), and SVM (Cortes & Vapnik, 1995). Other important techniques for classification tasks include Associative Classification (Liu et al, 1998) and Ensemble Classification (Tumer, 1996).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Bioelectrical impedance analysis (BIA) offers the potential for a simple, portable and relatively inexpensive technique for the in vivo measurement of total body water (TBW). The potential of BIA as a technique of body composition analysis is even greater when one considers that body water can be used as a surrogate measure of lean body mass. However, BIA has not found universal acceptance even with the introduction of multi-frequency BIA (MFBIA) which, potentially, may improve the predictive accuracy of the measurement. There are a number of reasons for this lack of acceptance, although perhaps the major reason is that no single algorithm has been developed which can be applied to all subject groups. This may be due, in part, to the commonly used wrist-to-ankle protocol which is not indicated by the basic theory of bioimpedance, where the body is considered as five interconnecting cylinders. Several workers have suggested the use of segmental BIA measurements to provide a protocol more in keeping with basic theory. However, there are other difficulties associated with the application of BIA, such as effects of hydration and ion status, posture and fluid distribution. A further putative advantage of MFBIA is the independent assessment not only of TBW but also of the extracellular fluid volume (ECW), hence heralding the possibility of,being able to assess the fluid distribution between these compartments. Results of studies in this area have been, to date, mixed. Whereas strong relationships of impedance values at low frequencies with ECW, and at high frequencies with TBW, have been reported, changes in impedance are not always well correlated with changes in the size of the fluid compartments (assessed by alternative and more direct means) in pathological conditions. Furthermore, the theoretical advantages of Cole-Cole modelling over selected frequency prediction have not always been apparent. This review will consider the principles, methodology and applications of BIA. The principles and methodology will,be considered in relation to the basic theory of BIA and difficulties experienced in its application. The relative merits of single and multiple frequency BIA will be addressed, with particular attention to the latter's role in the assessment of compartmental fluid volumes. (C) 1998 Elsevier Science Ltd. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The linear relationship between work accomplished (W-lim) and time to exhaustion (t(lim)) can be described by the equation: W-lim = a + CP.t(lim). Critical power (CP) is the slope of this line and is thought to represent a maximum rate of ATP synthesis without exhaustion, presumably an inherent characteristic of the aerobic energy system. The present investigation determined whether the choice of predictive tests would elicit significant differences in the estimated CP. Ten female physical education students completed, in random order and on consecutive days, five art-out predictive tests at preselected constant-power outputs. Predictive tests were performed on an electrically-braked cycle ergometer and power loadings were individually chosen so as to induce fatigue within approximately 1-10 mins. CP was derived by fitting the linear W-lim-t(lim) regression and calculated three ways: 1) using the first, third and fifth W-lim-t(lim) coordinates (I-135), 2) using coordinates from the three highest power outputs (I-123; mean t(lim) = 68-193 s) and 3) using coordinates from the lowest power outputs (I-345; mean t(lim) = 193-485 s). Repeated measures ANOVA revealed that CPI123 (201.0 +/- 37.9W) > CPI135 (176.1 +/- 27.6W) > CPI345 (164.0 +/- 22.8W) (P < 0.05). When the three sets of data were used to fit the hyperbolic Power-t(lim) regression, statistically significant differences between each CP were also found (P < 0.05). The shorter the predictive trials, the greater the slope of the W-lim-t(lim) regression; possibly because of the greater influence of 'aerobic inertia' on these trials. This may explain why CP has failed to represent a maximal, sustainable work rate. The present findings suggest that if CP is to represent the highest power output that an individual can maintain for a very long time without fatigue then CP should be calculated over a range of predictive tests in which the influence of aerobic inertia is minimised.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This note considers the value of surface response equations which can be used to calculate critical values for a range of unit root and cointegration tests popular in applied economic research.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

1. Although population viability analysis (PVA) is widely employed, forecasts from PVA models are rarely tested. This study in a fragmented forest in southern Australia contrasted field data on patch occupancy and abundance for the arboreal marsupial greater glider Petauroides volans with predictions from a generic spatially explicit PVA model. This work represents one of the first landscape-scale tests of its type. 2. Initially we contrasted field data from a set of eucalypt forest patches totalling 437 ha with a naive null model in which forecasts of patch occupancy were made, assuming no fragmentation effects and based simply on remnant area and measured densities derived from nearby unfragmented forest. The naive null model predicted an average total of approximately 170 greater gliders, considerably greater than the true count (n = 81). 3. Congruence was examined between field data and predictions from PVA under several metapopulation modelling scenarios. The metapopulation models performed better than the naive null model. Logistic regression showed highly significant positive relationships between predicted and actual patch occupancy for the four scenarios (P = 0.001-0.006). When the model-derived probability of patch occupancy was high (0.50-0.75, 0.75-1.00), there was greater congruence between actual patch occupancy and the predicted probability of occupancy. 4. For many patches, probability distribution functions indicated that model predictions for animal abundance in a given patch were not outside those expected by chance. However, for some patches the model either substantially over-predicted or under-predicted actual abundance. Some important processes, such as inter-patch dispersal, that influence the distribution and abundance of the greater glider may not have been adequately modelled. 5. Additional landscape-scale tests of PVA models, on a wider range of species, are required to assess further predictions made using these tools. This will help determine those taxa for which predictions are and are not accurate and give insights for improving models for applied conservation management.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this study is to create a two-tiered assessment combining restoration and conservation, both needed for biodiversity management. The first tier of this approach assesses the condition of a site using a standard bioassessment method, AUSRIVAS, to determine whether significant loss of biodiversity has occurred because of human activity. The second tier assesses the conservation value of sites that were determined to be unimpacted in the first step against a reference database. This ensures maximum complementarity without having to set a priori target areas. Using the reference database, we assign site-specific and comparable coefficients for both restoration (Observed/Expected taxa with > 50% probability of occurrence) and conservation values (O/E taxa with < 50%, rare taxa). In a trial on 75 sites on rivers around Sydney, NSW, Australia we were able to identify three regions: (1) an area that may need restoration; (2) an area that had a high conservation value and; (3) a region that was identified as having significant biodiversity loss but with high potential to respond to rehabilitation and become a biodiversity hotspot. These examples highlight the use of the new framework as a comprehensive system for biodiversity assessment.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The problem of the negative values of the interaction parameter in the equation of Frumkin has been analyzed with respect to the adsorption of nonionic molecules on energetically homogeneous surface. For this purpose, the adsorption states of a homologue series of ethoxylated nonionic surfactants on air/water interface have been determined using four different models and literature data (surface tension isotherms). The results obtained with the Frumkin adsorption isotherm imply repulsion between the adsorbed species (corresponding to negative values of the interaction parameter), while the classical lattice theory for energetically homogeneous surface (e.g., water/air) admits attraction alone. It appears that this serious contradiction can be overcome by assuming heterogeneity in the adsorption layer, that is, effects of partial condensation (formation of aggregates) on the surface. Such a phenomenon is suggested in the Fainerman-Lucassen-Reynders-Miller (FLM) 'Aggregation model'. Despite the limitations of the latter model (e.g., monodispersity of the aggregates), we have been able to estimate the sign and the order of magnitude of Frumkin's interaction parameter and the range of the aggregation numbers of the surface species. (C) 2004 Elsevier B.V All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

PURPOSE: Many guidelines advocate measurement of total or low density lipoprotein cholesterol (LDL), high density lipoprotein cholesterol (HDL), and triglycerides (TG) to determine treatment recommendations for preventing coronary heart disease (CHD) and cardiovascular disease (CVD). This analysis is a comparison of lipid variables as predictors of cardiovascular disease. METHODS: Hazard ratios for coronary and cardiovascular deaths by fourths of total cholesterol (TC), LDL, HDL, TG, non-HDL, TC/HDL, and TG/HDL values, and for a one standard deviation change in these variables, were derived in an individual participant data meta-analysis of 32 cohort studies conducted in the Asia-Pacific region. The predictive value of each lipid variable was assessed using the likelihood ratio statistic. RESULTS: Adjusting for confounders and regression dilution, each lipid variable had a positive (negative for HDL) log-linear association with fatal CHD and CVD. Individuals in the highest fourth of each lipid variable had approximately twice the risk of CHD compared with those with lowest levels. TG and HDL were each better predictors of CHD and CVD risk compared with TC alone, with test statistics similar to TC/HDL and TG/HDL ratios. Calculated LDL was a relatively poor predictor. CONCLUSIONS: While LDL reduction remains the main target of intervention for lipid-lowering, these data support the potential use of TG or lipid ratios for CHD risk prediction. (c) 2005 Elsevier Inc. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We add to current discussions about the interface between ecology, values, and objectivity by reporting on a novel Delphi-based study of the scientific reasoning employed by a group of eight ecologists as they collectively considered current ecological thinking. We rely on contextual empiricism, with its features of multiple ways of relating theory to reality and science as a social activity, to provide a richer understanding of scientific objectivity. This understanding recognizes the place and contributions of values and, in so doing, moves the discussion beyond whether or not science is value neutral.

Relevância:

20.00% 20.00%

Publicador: