1000 resultados para classification


Relevância:

20.00% 20.00%

Publicador:

Resumo:

A major challenge facing freshwater ecologists and managers is the development of models that link stream ecological condition to catchment scale effects, such as land use. Previous attempts to make such models have followed two general approaches. The bottom-up approach employs mechanistic models, which can quickly become too complex to be useful. The top-down approach employs empirical models derived from large data sets, and has often suffered from large amounts of unexplained variation in stream condition.

We believe that the lack of success of both modelling approaches may be at least partly explained by scientists considering too wide a breadth of catchment type. Thus, we believe that by stratifying large sets of catchments into groups of similar types prior to modelling, both types of models may be improved. This paper describes preliminary work using a Bayesian classification software package, ‘Autoclass’ (Cheeseman and Stutz 1996) to create classes of catchments within the Murray Darling Basin based on physiographic data.

Autoclass uses a model-based classification method that employs finite mixture modelling and trades off model fit versus complexity, leading to a parsimonious solution. The software provides information on the posterior probability that the classification is ‘correct’ and also probabilities for alternative classifications. The importance of each attribute in defining the individual classes is calculated and presented, assisting description of the classes. Each case is ‘assigned’ to a class based on membership probability, but the probability of membership of other classes is also provided. This feature deals very well with cases that do not fit neatly into a larger class. Lastly, Autoclass requires the user to specify the measurement error of continuous variables.

Catchments were derived from the Australian digital elevation model. Physiographic data werederived from national spatial data sets. There was very little information on measurement errors for the spatial data, and so a conservative error of 5% of data range was adopted for all continuous attributes. The incorporation of uncertainty into spatial data sets remains a research challenge.

The results of the classification were very encouraging. The software found nine classes of catchments in the Murray Darling Basin. The classes grouped together geographically, and followed altitude and latitude gradients, despite the fact that these variables were not included in the classification. Descriptions of the classes reveal very different physiographic environments, ranging from dry and flat catchments (i.e. lowlands), through to wet and hilly catchments (i.e. mountainous areas). Rainfall and slope were two important discriminators between classes. These two attributes, in particular, will affect the ways in which the stream interacts with the catchment, and can thus be expected to modify the effects of land use change on ecological condition. Thus, realistic models of the effects of land use change on streams would differ between the different types of catchments, and sound management practices will differ.

A small number of catchments were assigned to their primary class with relatively low probability. These catchments lie on the boundaries of groups of catchments, with the second most likely class being an adjacent group. The locations of these ‘uncertain’ catchments show that the Bayesian classification dealt well with cases that do not fit neatly into larger classes.

Although the results are intuitive, we cannot yet assess whether the classifications described in this paper would assist the modelling of catchment scale effects on stream ecological condition. It is most likely that catchment classification and modelling will be an iterative process, where the needs of the model are used to guide classification, and the results of classifications used to suggest further refinements to models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There has been an increasing interest in face recognition in recent years. Many recognition methods have been developed so far, some very encouraging. A key remaining issue is the existence of variations in the input face image. Today, methods exist that can handle specific image variations. But we are yet to see methods that can be used more effectively in unconstrained situations. This paper presents a method that can handle partial translation, rotation, or scale variations in the input face image. The principal is to automatically identify objects within images using their partial self-similarities. The paper presents two recognition methods which can be used to recognise objects within images. A face recognition system is then presented that is insensitive to limited translation, rotation, or scale variations in the input face image. The performance of the system is evaluated through four experiments. The results show that the system achieves higher recognition rates than those of a number of existing approaches.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An automatic road sign recognition system first locates road signs within images captured by an imaging sensor on-board of a vehicle, and then identifies the detected road signs. This paper presents an automatic neural-network-based road sign recognition system. First, a study of the existing road sign recognition research is presented. In this study, the issues associated with automatic road sign recognition are described, the existing methods developed to tackle the road sign recognition problem are reviewed, and a comparison of the features of these methods is given. Second, the developed road sign recognition system is described. The system is capable of analysing live colour road scene images, detecting multiple road signs within each image, and classifying the type of road signs detected. The system consists of two modules: detection and classification. The detection module segments the input image in the hue-saturation-intensity colour space, and then detects road signs using a Multi-layer Perceptron neural-network. The classification module determines the type of detected road signs using a series of one to one architectural Multi-layer Perceptron neural networks. Two sets of classifiers are trained using the Resillient-Backpropagation and Scaled-Conjugate-Gradient algorithms. The two modules of the system are evaluated individually first. Then the system is tested as a whole. The experimental results demonstrate that the system is capable of achieving an average recognition hit-rate of 95.96% using the scaled-conjugate-gradient trained classifiers.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Aim: To determine the time needed to provide clinical pharmacy services to individual patient episodes for medical and surgical patients and the effect of patient presentation and complexity on the clinical pharmacy workload. Method: During a 5-month period in 2006 at two general hospitals, pharmacists recorded a defined range of activities that they provided for patients, including the actual times required for these tasks. A customised database linked to the two hospitals' patient administration systems stored the data according to the specific patient episode number. The influence of patient presentation and complexity on the clinical pharmacy activities provided was also examined. Results: The average time required by pharmacists to undertake a medication history interview and medication reconciliation was 9.6 (SD 4.9) minutes. Interventions required 5.7 (SD 4.6) minutes, clinical review of the medical record 5.5 (SD 4.0) minutes and medication order review 3.5 (SD 2.0) minutes. For all of these activities, the time required for medical patients was greater than for surgical patients and greater for 'complicated' patients. The average time required to perform all clinical pharmacy activities for 1071 completed patient episodes was 14.4 (SD 10.9) minutes and was greater for medical and 'complicated' patients. Conclusion: The time needed to provide clinical pharmacy services was affected by whether the patients were medical or surgical. The existence of comorbidities or complications affected these times. The times required to perform clinical pharmacy activities may not be consistent with recently proposed staff ratios for the provision of a basic clinical pharmacy service.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The proliferation of malware is a serious threat to computer and information systems throughout the world. Antimalware companies are continually challenged to identify and counter new malware as it is released into the wild. In attempts to speed up this identification and response, many researchers have examined ways to efficiently automate classification of malware as it appears in the environment. In this paper, we present a fast, simple and scalable method of classifying Trojans based only on the lengths of their functions. Our results indicate that function length may play a significant role in classifying malware, and, combined with other features, may result in a fast, inexpensive and scalable method of malware classification.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This research examined the corporate branding approaches and strategies adopted by six prominent Australian arts and cultural organisations. The aim of this exploration was to identify patterns in branding across different arts and cultural organisations, and attempt to provide an initial classification for understanding how these organisations approach branding strategy. We found that three factors influenced branding strategy in the surveyed organisations, viz., the focus of branding process, the degree of consistency in branding communication, and the required level of customers’ involvement in the branded products. The organisations studied were then plotted on a continuum that considered each of these factors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Content adaptation is an attractive solution for the ever growing desktop based Web content delivered to the user via heterogeneous devices, in order to provide acceptable experience while surfing the Web. Bridging the mismatch between the rich content and the user device's resources (display, processing, navigation, network bandwidth, media support) without user intervention requires a proactive behavior. While content adaptation poses multitude of benefits, without proper strategies, adaptation will not be truly optimized. There have been many projects focused on content adaptation that have been designed with different goals and approaches. In this paper, we introduce a comprehensive classification for content adaptation system. The classification is used to group the approaches applied in the implementation of existing content adaptation system. Survey on some content adaptation systems also been provided. We also present the research spectrum in content adaptation and discuss the challenges.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

k-nearest neighbors (kNN) is a popular method for function approximation and classification. One drawback of this method is that the nearest neighbors can be all located on one side of the point in question x. An alternative natural neighbors method is expensive for more than three variables. In this paper we propose the use of the discrete Choquet integral for combining the values of the nearest neighbors so that redundant information is canceled out. We design a fuzzy measure based on location of the nearest neighbors, which favors neighbors located all around x.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A method is presented that achieves lung nodule detection by classification of nodule and non-nodule patterns. It is based on random forests which are ensemble learners that grow classification trees. Each tree produces a classification decision, and an integrated output is calculated. The performance of the developed method is compared against that of the support vector machine and the decision tree methods. Three experiments are performed using lung scans of 32 patients including thousands of images within which nodule locations are marked by expert radiologists. The classification errors and execution times are presented and discussed. The lowest classification error (2.4%) has been produced by the developed method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The reliability of an induced classifier can be affected by several factors including the data oriented factors and the algorithm oriented factors. In some cases, the reliability could also be affected by knowledge oriented factors. In this paper, we analyze three special cases to examine the reliability of the discovered knowledge. Our case study results show that (1) in the cases of mining from low quality data, rough classification approach is more reliable than exact approach which in general tolerate to low quality data; (2) Without sufficient large size of the data, the reliability of the discovered knowledge will be decreased accordingly; (3) The reliability of point learning approach could easily be misled by noisy data. It will in most cases generate an unreliable interval and thus affect the reliability of the discovered knowledge. It is also reveals that the inexact field is a good learning strategy that could model the potentials and to improve the discovery reliability.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper introduces a new technique in the investigation of object classification and illustrates the potential use of this technique for the analysis of a range of biological data, using avian morphometric data as an example. The nascent variable precision rough sets (VPRS) model is introduced and compared with the decision tree method ID3 (through a ‘leave n out’ approach), using the same dataset of morphometric measures of European barn swallows (Hirundo rustica) and assessing the accuracy of gender classification based on these measures. The results demonstrate that the VPRS model, allied with the use of a modern method of discretization of data, is comparable with the more traditional non-parametric ID3 decision tree method. We show that, particularly in small samples, the VPRS model can improve classification and to a lesser extent prediction aspects over ID3. Furthermore, through the ‘leave n out’ approach, some indication can be produced of the relative importance of the different morphometric measures used in this problem. In this case we suggest that VPRS has advantages over ID3, as it intelligently uses more of the morphometric data available for the data classification, whilst placing less emphasis on variables with low reliability. In biological terms, the results suggest that the gender of swallows can be determined with reasonable accuracy from morphometric data and highlight the most important variables in this process. We suggest that both analysis techniques are potentially useful for the analysis of a range of different types of biological datasets, and that VPRS in particular has potential for application to a range of biological circumstances.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article investigates the potential of a novel technique for object classification, called Classification and Ranking Belief Simplex (CaRBS), which is based on the Dempster-Shafer theory of evidence. As such, the classification of objects and the evidence from their characteristics have a level of ignorance associated with them. Its potential is exposited in the application of the classification of European barn swallows according to their gender. The classification of biological data in the presence of ignorance about such data sets is a common problem in biology. Comparisons of the results from CaRBS with those from multivariate discriminant analysis and neural networks are made. Also shown throughout the investigation is the interpretability of the results with the utilisation of the simplex plot method of representing data

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Regardless of the technical procedure used in signalling corporate collapse, the bottom line rests on the predictive power of the corresponding statistical model. In that regard, it is imperative to empirically test the model using a data sample of both collapsed and non-collapsed companies. A superior model is one that successfully classifies collapsed and non-collapsed companies in their respective categories with a high degree of accuracy. Empirical studies of this nature have thus far done one of two things. (1) Some have classified companies based on a specific statistical modelling process. (2) Some have classified companies based on two (sometimes – but rarely – more than two) independent statistical modelling processes for the purposes of comparing one with the other. In the latter case, the mindset of the researchers has been – invariably – to pitch one procedure against the other. This paper raises the question, why pitch one statistical process against another; why not make the two procedures work together? As such, this paper puts forward an innovative dual-classification scheme for signalling corporate collapse: dual in the sense that it relies on two statistical procedures concurrently. Using a data sample of Australian publicly listed companies, the proposed scheme is tested against the traditional approach taken thus far in the pertinent literature. The results demonstrate that the proposed dual-classification scheme signals collapse with a higher degree of accuracy.