960 resultados para rough sets


Relevância:

60.00% 60.00%

Publicador:

Resumo:

The year 1968 saw a major shift from univariate to multivariate methodological approaches to ratio-based modelling of corporate collapse. This was facilitated by the introduction of a new statistical tool called Multiple Discriminant Analysis (MDA). However, it did not take long before other statistical tools were developed. The primary objective for developing these tools was to enable deriving models that would at least do as good a job asMDA, but rely on fewer assumptions. With the introduction of new statistical tools, researchers became pre-occupied with testing them in signalling collapse. lLTUong the ratio-based approaches were Logit analysis, Neural Network analysis, Probit analysis, ID3, Recursive Partitioning Algorithm, Rough Sets analysis, Decomposition analysis, Going Concern Advisor, Koundinya and Purl judgmental approach, Tabu Search and Mixed Logit analysis. Regardless of which methodological approach was chosen, most were compared to MDA. This paper reviews these various approaches. Emphasis is placed on how they fared against MDA in signalling corporate collapse.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Clinicians could model the brain injury of a patient through his brain activity. However, how this model is defined and how it changes when the patient is recovering are questions yet unanswered. In this paper, the use of MedVir framework is proposed with the aim of answering these questions. Based on complex data mining techniques, this provides not only the differentiation between TBI patients and control subjects (with a 72% of accuracy using 0.632 Bootstrap validation), but also the ability to detect whether a patient may recover or not, and all of that in a quick and easy way through a visualization technique which allows interaction.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In order to address problems of information overload in digital imagery task domains we have developed an interactive approach to the capture and reuse of image context information. Our framework models different aspects of the relationship between images and domain tasks they support by monitoring the interactive manipulation and annotation of task-relevant imagery. The approach allows us to gauge a measure of a user's intentions as they complete goal-directed image tasks. As users analyze retrieved imagery their interactions are captured and an expert task context is dynamically constructed. This human expertise, proficiency, and knowledge can then be leveraged to support other users in carrying out similar domain tasks. We have applied our techniques to two multimedia retrieval applications for two different image domains, namely the geo-spatial and medical imagery domains. © Springer-Verlag Berlin Heidelberg 2007.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A rough set approach for attribute reduction is an important research subject in data mining and machine learning. However, most attribute reduction methods are performed on a complete decision system table. In this paper, we propose methods for attribute reduction in static incomplete decision systems and dynamic incomplete decision systems with dynamically-increasing and decreasing conditional attributes. Our methods use generalized discernibility matrix and function in tolerance-based rough sets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

R. Jensen and Q. Shen. Semantics-Preserving Dimensionality Reduction: Rough and Fuzzy-Rough Based Approaches. IEEE Transactions on Knowledge and Data Engineering, 16(12): 1457-1471. 2004.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

R. Jensen and Q. Shen, 'Fuzzy-Rough Attribute Reduction with Application to Web Categorization,' Fuzzy Sets and Systems, vol. 141, no. 3, pp. 469-485, 2004.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computational Intelligence and Feature Selection provides a high level audience with both the background and fundamental ideas behind feature selection with an emphasis on those techniques based on rough and fuzzy sets, including their hybridizations. It introduces set theory, fuzzy set theory, rough set theory, and fuzzy-rough set theory, and illustrates the power and efficacy of the feature selections described through the use of real-world applications and worked examples. Program files implementing major algorithms covered, together with the necessary instructions and datasets, are available on the Web.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In previous paper, we introduced a concept of multi-soft sets and used it for finding reducts. However, the comparison of the proposed reduct has not been presented yet, especially with rough-set based reduct. In this paper, we present matrices representation of multi-soft sets. We define AND and OR operations on a collection of such matrices and apply it for finding reducts and core of attributes in a multi-valued information system. Finally, we prove that our proposed technique for reduct is equivalent to Pawlak’s rough reduct.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There have only been a small number of applications of consumer decision set theory to holiday destination choice, and these studies have tended to rely on a single cross sectional snapshot of research participants’ stated preferences. Very little has been reported on the relationship between stated destination preferences and actual travel, or changes in decision set composition over time. The paper presents a rare longitudinal examination of destination decision sets, in the context of short break holidays by car in Queensland, Australia. Two questionnaires were administered, three months apart. The first identified destination preferences while the second examined actual travel and revisited destination preferences. In relation to the conference theme, there was very little change in consumer preferences towards the competitive set of destinations over the three month period. A key implication for the destination of interest, which, in an attempt to change market perceptions, launched a new brand campaign during the period of the project, is that a long term investment in a consistent brand message will be required to change market perceptions. The results go some way to support the proposition that the positioning of a destination into a consumer’s decision set represents a source of competitive advantage.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An information filtering (IF) system monitors an incoming document stream to find the documents that match the information needs specified by the user profiles. To learn to use the user profiles effectively is one of the most challenging tasks when developing an IF system. With the document selection criteria better defined based on the users’ needs, filtering large streams of information can be more efficient and effective. To learn the user profiles, term-based approaches have been widely used in the IF community because of their simplicity and directness. Term-based approaches are relatively well established. However, these approaches have problems when dealing with polysemy and synonymy, which often lead to an information overload problem. Recently, pattern-based approaches (or Pattern Taxonomy Models (PTM) [160]) have been proposed for IF by the data mining community. These approaches are better at capturing sematic information and have shown encouraging results for improving the effectiveness of the IF system. On the other hand, pattern discovery from large data streams is not computationally efficient. Also, these approaches had to deal with low frequency pattern issues. The measures used by the data mining technique (for example, “support” and “confidences”) to learn the profile have turned out to be not suitable for filtering. They can lead to a mismatch problem. This thesis uses the rough set-based reasoning (term-based) and pattern mining approach as a unified framework for information filtering to overcome the aforementioned problems. This system consists of two stages - topic filtering and pattern mining stages. The topic filtering stage is intended to minimize information overloading by filtering out the most likely irrelevant information based on the user profiles. A novel user-profiles learning method and a theoretical model of the threshold setting have been developed by using rough set decision theory. The second stage (pattern mining) aims at solving the problem of the information mismatch. This stage is precision-oriented. A new document-ranking function has been derived by exploiting the patterns in the pattern taxonomy. The most likely relevant documents were assigned higher scores by the ranking function. Because there is a relatively small amount of documents left after the first stage, the computational cost is markedly reduced; at the same time, pattern discoveries yield more accurate results. The overall performance of the system was improved significantly. The new two-stage information filtering model has been evaluated by extensive experiments. Tests were based on the well-known IR bench-marking processes, using the latest version of the Reuters dataset, namely, the Reuters Corpus Volume 1 (RCV1). The performance of the new two-stage model was compared with both the term-based and data mining-based IF models. The results demonstrate that the proposed information filtering system outperforms significantly the other IF systems, such as the traditional Rocchio IF model, the state-of-the-art term-based models, including the BM25, Support Vector Machines (SVM), and Pattern Taxonomy Model (PTM).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This study assesses the recently proposed data-driven background dataset refinement technique for speaker verification using alternate SVM feature sets to the GMM supervector features for which it was originally designed. The performance improvements brought about in each trialled SVM configuration demonstrate the versatility of background dataset refinement. This work also extends on the originally proposed technique to exploit support vector coefficients as an impostor suitability metric in the data-driven selection process. Using support vector coefficients improved the performance of the refined datasets in the evaluation of unseen data. Further, attempts are made to exploit the differences in impostor example suitability measures from varying features spaces to provide added robustness.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Maintenance is a time consuming and expensive task for any golf course or driving range manager. For a golf course the primary tasks are grass mowing and maintenance (fertilizer and herbicide spreading), while for a driving range mowing, maintenance and ball collection are required. All these tasks require an operator to drive a vehicle along paths which are generally predefined. This paper presents some preliminary in-field tsting results for an automated tractor vehicle performing golf ball collection on an actual driving range, and mowing on difficult unstructured terrain.