979 resultados para Classification procedure
Resumo:
Ninety-two strong-motion earthquake records from the California region, U.S.A., have been statistically studied using principal component analysis in terms of twelve important standardized strong-motion characteristics. The first two principal components account for about 57 per cent of the total variance. Based on these two components the earthquake records are classified into nine groups in a two-dimensional principal component plane. Also a unidimensional engineering rating scale is proposed. The procedure can be used as an objective approach for classifying and rating future earthquakes.
Resumo:
The introduction of casemix funding for Australian acute health care services has challenged Social Work to demonstrate clear reporting mechanisms, demonstrate effective practice and to justify interventions provided. The term 'casemix' is used to describe the mix and type of patients treated by a hospital or other health care services. There is wide acknowledgement that the procedure-based system of Diagnosis Related Groupings (DRGs) is grounded in a medical/illness perspective and is unsatisfactory in describing and predicting the activity of Social Work and other allied health professions in health care service delivery. The National Allied Health Casemix Committee was established in 1991 as the peak body to represent allied health professions in matters related to casemix classification. This Committee has pioneered a nationally consistent, patient-centred information system for allied health. This paper describes the classification systems and codes developed for Social Work, which includes a minimum data set, a classification hierarchy, the set of activity (input) codes and 'indicator for intervention' codes. The advantages and limitations of the system are also discussed.
Resumo:
Traumatic brain injury (TBI) affects people of all ages and is a cause of long-term disability. In recent years, the epidemiological patterns of TBI have been changing. TBI is a heterogeneous disorder with different forms of presentation and highly individual outcome regarding functioning and health-related quality of life (HRQoL). The meaning of disability differs from person to person based on the individual s personality, value system, past experience, and the purpose he or she sees in life. Understanding of all these viewpoints is needed in comprehensive rehabilitation. This study examines the epidemiology of TBI in Finland as well as functioning and HRQoL after TBI, and compares the subjective and objective assessments of outcome. The frame of reference is the International Classification of Functioning, Disability and Health (ICF). The subjects of Study I represent the population of Finnish TBI patients who experienced their first TBI between 1991 and 2005. The 55 Finnish subjects of Studies II and IV participated in the first wave of the international Quality of life after brain injury (QOLIBRI) validation study. The 795 subjects from six language areas of Study III formed the second wave of the QOLIBRI validation study. The average annual incidence of Finnish hospitalised TBI patients during the years 1991-2005 was 101:100 000 in patients who had TBI as the primary diagnosis and did not have a previous TBI in their medical history. Males (59.2%) were at considerably higher risk of getting a TBI than females. The most common external cause of the injury was falls in all age groups. The number of TBI patients ≥ 70 years of age increased by 59.4% while the number of inhabitants older than 70 years increased by 30.3% in the population of Finland during the same time period. The functioning of a sample of 55 persons with TBI was assessed by extracting information from the patients medical documents using the ICF checklist. The most common problems were found in the ICF components of Body Functions (b) and Activities and Participation (d). HRQoL was assessed with the QOLIBRI which showed the highest level of satisfaction on the Emotions, Physical Problems and Daily Life and Autonomy scales. The highest scores were obtained by the youngest participants and participants living independently without the help of other people, and by people who were working. The relationship between the functional outcome and HRQoL was not straightforward. The procedure of linking the QOLIBRI and the GOSE to the ICF showed that these two outcome measures cover the relevant domains of TBI patients functioning. The QOLIBRI provides the patients subjective view, while the GOSE summarises the objective elements of functioning. Our study indicates that there are certain domains of functioning that are not traditionally sufficiently documented but are important for the HRQoL of persons with TBI. This was the finding especially in the domains of interpersonal relationships, social and leisure activities, self, and the environment. Rehabilitation aims to optimize functioning and to minimize the experience of disability among people with health conditions, and it needs to be based on a comprehensive understanding of human functioning. As an integrative model, the ICF may serve as a frame of reference in achieving such an understanding.
Resumo:
The presence of a large number of spectral bands in the hyperspectral images increases the capability to distinguish between various physical structures. However, they suffer from the high dimensionality of the data. Hence, the processing of hyperspectral images is applied in two stages: dimensionality reduction and unsupervised classification techniques. The high dimensionality of the data has been reduced with the help of Principal Component Analysis (PCA). The selected dimensions are classified using Niche Hierarchical Artificial Immune System (NHAIS). The NHAIS combines the splitting method to search for the optimal cluster centers using niching procedure and the merging method is used to group the data points based on majority voting. Results are presented for two hyperspectral images namely EO-1 Hyperion image and Indian pines image. A performance comparison of this proposed hierarchical clustering algorithm with the earlier three unsupervised algorithms is presented. From the results obtained, we deduce that the NHAIS is efficient.
Resumo:
We consider the problem of developing privacy-preserving machine learning algorithms in a dis-tributed multiparty setting. Here different parties own different parts of a data set, and the goal is to learn a classifier from the entire data set with-out any party revealing any information about the individual data points it owns. Pathak et al [7]recently proposed a solution to this problem in which each party learns a local classifier from its own data, and a third party then aggregates these classifiers in a privacy-preserving manner using a cryptographic scheme. The generaliza-tion performance of their algorithm is sensitive to the number of parties and the relative frac-tions of data owned by the different parties. In this paper, we describe a new differentially pri-vate algorithm for the multiparty setting that uses a stochastic gradient descent based procedure to directly optimize the overall multiparty ob-jective rather than combining classifiers learned from optimizing local objectives. The algorithm achieves a slightly weaker form of differential privacy than that of [7], but provides improved generalization guarantees that do not depend on the number of parties or the relative sizes of the individual data sets. Experimental results corrob-orate our theoretical findings.
Resumo:
In recent years, the performance of semi-supervised learning has been theoretically investigated. However, most of this theoretical development has focussed on binary classification problems. In this paper, we take it a step further by extending the work of Castelli and Cover [1] [2] to the multi-class paradigm. Particularly, we consider the key problem in semi-supervised learning of classifying an unseen instance x into one of K different classes, using a training dataset sampled from a mixture density distribution and composed of l labelled records and u unlabelled examples. Even under the assumption of identifiability of the mixture and having infinite unlabelled examples, labelled records are needed to determine the K decision regions. Therefore, in this paper, we first investigate the minimum number of labelled examples needed to accomplish that task. Then, we propose an optimal multi-class learning algorithm which is a generalisation of the optimal procedure proposed in the literature for binary problems. Finally, we make use of this generalisation to study the probability of error when the binary class constraint is relaxed.
Resumo:
We introduce and explore an approach to estimating statistical significance of classification accuracy, which is particularly useful in scientific applications of machine learning where high dimensionality of the data and the small number of training examples render most standard convergence bounds too loose to yield a meaningful guarantee of the generalization ability of the classifier. Instead, we estimate statistical significance of the observed classification accuracy, or the likelihood of observing such accuracy by chance due to spurious correlations of the high-dimensional data patterns with the class labels in the given training set. We adopt permutation testing, a non-parametric technique previously developed in classical statistics for hypothesis testing in the generative setting (i.e., comparing two probability distributions). We demonstrate the method on real examples from neuroimaging studies and DNA microarray analysis and suggest a theoretical analysis of the procedure that relates the asymptotic behavior of the test to the existing convergence bounds.
Resumo:
Tephrochronology, a key tool in the correlation of Quaternary sequences, relies on the extraction of tephra shards from sediments for visual identification and high-precision geochemical comparison. A prerequisite for the reliable correlation of tephra layers is that the geochemical composition of glass shards remains unaltered by natural processes (e.g. chemical exchange in the sedimentary environment) and/or by laboratory analytical procedures. However, natural glasses, particularly when in the form of small shards with a high surface to volume ratio, are prone to chemical alteration in both acidic and basic environments. Current techniques for the extraction of distal tephra from sediments involve the ‘cleaning’ of samples in precisely such environments and at elevated temperatures. The acid phase of the ‘cleaning’ process risks alteration of the geochemical signature of the shards, while the basic phase leads to considerable sample loss through dissolution of the silica network. Here, we illustrate the degree of alteration and loss to which distal tephras may be prone, and introduce a less destructive procedure for their extraction. This method is based on stepped heavy liquid flotation and which results in samples of sufficient quality for analysis while preserving their geochemical integrity. In trials, this method out-performed chemical extraction procedures in terms of the number of shards recovered and has resulted in the detection of new tephra layers with low shard concentrations. The implications of this study are highly significant because (i) the current database of distal tephra records and their corresponding geochemical signatures may require refinement and (ii) the record of distal tephras may be incomplete due to sample loss induced by corrosive laboratory procedures. It is therefore vital that less corrosive laboratory procedures are developed to make the detection and classification of distal glass tephra more secure.
Resumo:
Urothelial cancer (UC) is highly recurrent and can progress from non-invasive (NMIUC) to a more aggressive muscle-invasive (MIUC) subtype that invades the muscle tissue layer of the bladder. We present a proof of principle study that network-based features of gene pairs can be used to improve classifier performance and the functional analysis of urothelial cancer gene expression data. In the first step of our procedure each individual sample of a UC gene expression dataset is inflated by gene pair expression ratios that are defined based on a given network structure. In the second step an elastic net feature selection procedure for network-based signatures is applied to discriminate between NMIUC and MIUC samples. We performed a repeated random subsampling cross validation in three independent datasets. The network signatures were characterized by a functional enrichment analysis and studied for the enrichment of known cancer genes. We observed that the network-based gene signatures from meta collections of proteinprotein interaction (PPI) databases such as CPDB and the PPI databases HPRD and BioGrid improved the classification performance compared to single gene based signatures. The network based signatures that were derived from PPI databases showed a prominent enrichment of cancer genes (e.g., TP53, TRIM27 and HNRNPA2Bl). We provide a novel integrative approach for large-scale gene expression analysis for the identification and development of novel diagnostical targets in bladder cancer. Further, our method allowed to link cancer gene associations to network-based expression signatures that are not observed in gene-based expression signatures.
Resumo:
Low noise surfaces have been increasingly considered as a viable and cost-effective alternative to acoustical barriers. However, road planners and administrators frequently lack information on the correlation between the type of road surface and the resulting noise emission profile. To address this problem, a method to identify and classify different types of road pavements was developed, whereby near field road noise is analyzed using statistical learning methods. The vehicle rolling sound signal near the tires and close to the road surface was acquired by two microphones in a special arrangement which implements the Close-Proximity method. A set of features, characterizing the properties of the road pavement, was extracted from the corresponding sound profiles. A feature selection method was used to automatically select those that are most relevant in predicting the type of pavement, while reducing the computational cost. A set of different types of road pavement segments were tested and the performance of the classifier was evaluated. Results of pavement classification performed during a road journey are presented on a map, together with geographical data. This procedure leads to a considerable improvement in the quality of road pavement noise data, thereby increasing the accuracy of road traffic noise prediction models.
Resumo:
Affiliation: Centre Robert-Cedergren de l'Université de Montréal en bio-informatique et génomique & Département de biochimie, Université de Montréal
Resumo:
Objective To determine scoliosis curve types using non invasive surface acquisition, without prior knowledge from X-ray data. Methods Classification of scoliosis deformities according to curve type is used in the clinical management of scoliotic patients. In this work, we propose a robust system that can determine the scoliosis curve type from non invasive acquisition of the 3D back surface of the patients. The 3D image of the surface of the trunk is divided into patches and local geometric descriptors characterizing the back surface are computed from each patch and constitute the features. We reduce the dimensionality by using principal component analysis and retain 53 components using an overlap criterion combined with the total variance in the observed variables. In this work, a multi-class classifier is built with least-squares support vector machines (LS-SVM). The original LS-SVM formulation was modified by weighting the positive and negative samples differently and a new kernel was designed in order to achieve a robust classifier. The proposed system is validated using data from 165 patients with different scoliosis curve types. The results of our non invasive classification were compared with those obtained by an expert using X-ray images. Results The average rate of successful classification was computed using a leave-one-out cross-validation procedure. The overall accuracy of the system was 95%. As for the correct classification rates per class, we obtained 96%, 84% and 97% for the thoracic, double major and lumbar/thoracolumbar curve types, respectively. Conclusion This study shows that it is possible to find a relationship between the internal deformity and the back surface deformity in scoliosis with machine learning methods. The proposed system uses non invasive surface acquisition, which is safe for the patient as it involves no radiation. Also, the design of a specific kernel improved classification performance.
Resumo:
A new procedure for the classification of lower case English language characters is presented in this work . The character image is binarised and the binary image is further grouped into sixteen smaller areas ,called Cells . Each cell is assigned a name depending upon the contour present in the cell and occupancy of the image contour in the cell. A data reduction procedure called Filtering is adopted to eliminate undesirable redundant information for reducing complexity during further processing steps . The filtered data is fed into a primitive extractor where extraction of primitives is done . Syntactic methods are employed for the classification of the character . A decision tree is used for the interaction of the various components in the scheme . 1ike the primitive extraction and character recognition. A character is recognized by the primitive by primitive construction of its description . Openended inventories are used for including variants of the characters and also adding new members to the general class . Computer implementation of the proposal is discussed at the end using handwritten character samples . Results are analyzed and suggestions for future studies are made. The advantages of the proposal are discussed in detail .
Resumo:
The conventional method for the assessment of acute dermal toxicity (OECD Test Guideline 402, 1987) uses death of animals as an endpoint to identify the median lethal dose (LD50). A new OECD Testing Guideline called the dermal fixed dose procedure (dermal FDP) is being prepared to provide an alternative to Test Guideline 402. In contrast to Test Guideline 402, the dermal FDP does not provide a point estimate of the LD50, but aims to identify that dose of the substance under investigation that causes clear signs of nonlethal toxicity. This is then used to assign classification according to the new Globally Harmonised System of Classification and Labelling scheme (GHS). The dermal FDP has been validated using statistical modelling rather than by in vivo testing. The statistical modelling approach enables calculation of the probability of each GHS classification and the expected numbers of deaths and animals used in the test for imaginary substances with a range of LD50 values and dose-response curve slopes. This paper describes the dermal FDP and reports the results from the statistical evaluation. It is shown that the procedure will be completed with considerably less death and suffering than guideline 402, and will classify substances either in the same or a more stringent GHS class than that assigned on the basis of the LD50 value.
Statistical evaluation of the fixed concentration procedure for acute inhalation toxicity assessment
Resumo:
The conventional method for the assessment of acute inhalation toxicity (OECD Test Guideline 403, 1981) uses death of animals as an endpoint to identify the median lethal concentration (LC50). A new OECD Testing Guideline called the Fixed Concentration Procedure (FCP) is being prepared to provide an alternative to Test Guideline 403. Unlike Test Guideline 403, the FCP does not provide a point estimate of the LC50, but aims to identify an airborne exposure level that causes clear signs of nonlethal toxicity. This is then used to assign classification according to the new Globally Harmonized System of Classification and Labelling scheme (GHS). The FCP has been validated using statistical simulation rather than byin vivo testing. The statistical simulation approach predicts the GHS classification outcome and the numbers of deaths and animals used in the test for imaginary substances with a range of LC50 values and dose response curve slopes. This paper describes the FCP and reports the results from the statistical simulation study assessing its properties. It is shown that the procedure will be completed with considerably less death and suffering than Test Guideline 403, and will classify substances either in the same or a more stringent GHS class than that assigned on the basis of the LC50 value.