84 resultados para Turing machines.
Resumo:
Monitoring of posture allocations and activities enables accurate estimation of energy expenditure and may aid in obesity prevention and treatment. At present, accurate devices rely on multiple sensors distributed on the body and thus may be too obtrusive for everyday use. This paper presents a novel wearable sensor, which is capable of very accurate recognition of common postures and activities. The patterns of heel acceleration and plantar pressure uniquely characterize postures and typical activities while requiring minimal preprocessing and no feature extraction. The shoe sensor was tested in nine adults performing sitting and standing postures and while walking, running, stair ascent/descent and cycling. Support vector machines (SVMs) were used for classification. A fourfold validation of a six-class subject-independent group model showed 95.2% average accuracy of posture/activity classification on full sensor set and over 98% on optimized sensor set. Using a combination of acceleration/pressure also enabled a pronounced reduction of the sampling frequency (25 to 1 Hz) without significant loss of accuracy (98% versus 93%). Subjects had shoe sizes (US) M9.5-11 and W7-9 and body mass index from 18.1 to 39.4 kg/m2 and thus suggesting that the device can be used by individuals with varying anthropometric characteristics.
Resumo:
The present study aims to analyze attitudes and beliefs of the French-speaking general Swiss population (n = 2500; female n = 1280; mean age = 43 years) as regards gambling, which are to date almost exclusively studied in the North American and Australian contexts. Beliefs related to gambling include the perception of the effectiveness of preventive measures toward gambling, the comparative risk assessment of different addictive behaviors, the perceived risks of different types of gambling and attitudes are related to the gambler's personality. The general population perceived gambling rather negatively and was conscious of the potential risks of gambling; indeed, 59.0% of the sample identified gambling as an addictive practice. Slot machines were estimated to bear the highest risk. Compared with women and older people, men and young people indicated more positive beliefs about gambling; they perceived gambling as less addictive, supported structural preventive measures less often, and perceived gambling as a less serious problem for society. Gamblers were more likely to put their practices into perspective, perceiving gambling more positively than non-gamblers. General population surveys on such beliefs can deliver insights into preventive actions that should be targeted to young men who showed more favorable views of gambling, which have been shown to be associated with increased risk for problematic gambling.
Resumo:
Recently, kernel-based Machine Learning methods have gained great popularity in many data analysis and data mining fields: pattern recognition, biocomputing, speech and vision, engineering, remote sensing etc. The paper describes the use of kernel methods to approach the processing of large datasets from environmental monitoring networks. Several typical problems of the environmental sciences and their solutions provided by kernel-based methods are considered: classification of categorical data (soil type classification), mapping of environmental and pollution continuous information (pollution of soil by radionuclides), mapping with auxiliary information (climatic data from Aral Sea region). The promising developments, such as automatic emergency hot spot detection and monitoring network optimization are discussed as well.
Resumo:
To be diagnostically useful, structural MRI must reliably distinguish Alzheimer's disease (AD) from normal aging in individual scans. Recent advances in statistical learning theory have led to the application of support vector machines to MRI for detection of a variety of disease states. The aims of this study were to assess how successfully support vector machines assigned individual diagnoses and to determine whether data-sets combined from multiple scanners and different centres could be used to obtain effective classification of scans. We used linear support vector machines to classify the grey matter segment of T1-weighted MR scans from pathologically proven AD patients and cognitively normal elderly individuals obtained from two centres with different scanning equipment. Because the clinical diagnosis of mild AD is difficult we also tested the ability of support vector machines to differentiate control scans from patients without post-mortem confirmation. Finally we sought to use these methods to differentiate scans between patients suffering from AD from those with frontotemporal lobar degeneration. Up to 96% of pathologically verified AD patients were correctly classified using whole brain images. Data from different centres were successfully combined achieving comparable results from the separate analyses. Importantly, data from one centre could be used to train a support vector machine to accurately differentiate AD and normal ageing scans obtained from another centre with different subjects and different scanner equipment. Patients with mild, clinically probable AD and age/sex matched controls were correctly separated in 89% of cases which is compatible with published diagnosis rates in the best clinical centres. This method correctly assigned 89% of patients with post-mortem confirmed diagnosis of either AD or frontotemporal lobar degeneration to their respective group. Our study leads to three conclusions: Firstly, support vector machines successfully separate patients with AD from healthy aging subjects. Secondly, they perform well in the differential diagnosis of two different forms of dementia. Thirdly, the method is robust and can be generalized across different centres. This suggests an important role for computer based diagnostic image analysis for clinical practice.
Resumo:
In recent years there has been an explosive growth in the development of adaptive and data driven methods. One of the efficient and data-driven approaches is based on statistical learning theory (Vapnik 1998). The theory is based on Structural Risk Minimisation (SRM) principle and has a solid statistical background. When applying SRM we are trying not only to reduce training error ? to fit the available data with a model, but also to reduce the complexity of the model and to reduce generalisation error. Many nonlinear learning procedures recently developed in neural networks and statistics can be understood and interpreted in terms of the structural risk minimisation inductive principle. A recent methodology based on SRM is called Support Vector Machines (SVM). At present SLT is still under intensive development and SVM find new areas of application (www.kernel-machines.org). SVM develop robust and non linear data models with excellent generalisation abilities that is very important both for monitoring and forecasting. SVM are extremely good when input space is high dimensional and training data set i not big enough to develop corresponding nonlinear model. Moreover, SVM use only support vectors to derive decision boundaries. It opens a way to sampling optimization, estimation of noise in data, quantification of data redundancy etc. Presentation of SVM for spatially distributed data is given in (Kanevski and Maignan 2004).
Resumo:
In this paper, we propose two active learning algorithms for semiautomatic definition of training samples in remote sensing image classification. Based on predefined heuristics, the classifier ranks the unlabeled pixels and automatically chooses those that are considered the most valuable for its improvement. Once the pixels have been selected, the analyst labels them manually and the process is iterated. Starting with a small and nonoptimal training set, the model itself builds the optimal set of samples which minimizes the classification error. We have applied the proposed algorithms to a variety of remote sensing data, including very high resolution and hyperspectral images, using support vector machines. Experimental results confirm the consistency of the methods. The required number of training samples can be reduced to 10% using the methods proposed, reaching the same level of accuracy as larger data sets. A comparison with a state-of-the-art active learning method, margin sampling, is provided, highlighting advantages of the methods proposed. The effect of spatial resolution and separability of the classes on the quality of the selection of pixels is also discussed.
Resumo:
The development of statistical models for forensic fingerprint identification purposes has been the subject of increasing research attention in recent years. This can be partly seen as a response to a number of commentators who claim that the scientific basis for fingerprint identification has not been adequately demonstrated. In addition, key forensic identification bodies such as ENFSI [1] and IAI [2] have recently endorsed and acknowledged the potential benefits of using statistical models as an important tool in support of the fingerprint identification process within the ACE-V framework. In this paper, we introduce a new Likelihood Ratio (LR) model based on Support Vector Machines (SVMs) trained with features discovered via morphometric and spatial analyses of corresponding minutiae configurations for both match and close non-match populations often found in AFIS candidate lists. Computed LR values are derived from a probabilistic framework based on SVMs that discover the intrinsic spatial differences of match and close non-match populations. Lastly, experimentation performed on a set of over 120,000 publicly available fingerprint images (mostly sourced from the National Institute of Standards and Technology (NIST) datasets) and a distortion set of approximately 40,000 images, is presented, illustrating that the proposed LR model is reliably guiding towards the right proposition in the identification assessment of match and close non-match populations. Results further indicate that the proposed model is a promising tool for fingerprint practitioners to use for analysing the spatial consistency of corresponding minutiae configurations.
Resumo:
Contexte et but de l'étude: Les fractures du triquetrum sont les deuxièmes fractures des os du carpe en fréquence, après celles du scaphoïde. Elles représentent environ 3.5% de toutes les lésions traumatiques du poignet, et résultent le plus souvent d'une chute de sa hauteur avec réception sur le poignet en hyper-extension. Leur mécanisme physiopathologique reste débattu. La première théorie fut celle de l'avulsion ligamentaire d'un fragment osseux dorsal. Puis, Levy et coll. ainsi que Garcia-Elias ont successivement suggéré que ces fractures résultaient plutôt d'une impaction ulno-carpienne. De nombreux ligaments (intrinsèques et extrinsèques du carpe) s'insèrent sur les versants palmaires et dorsaux du triquetrum. Ces ligaments jouent un rôle essentiel dans le maintien de la stabilité du carpe. Bien que l'arthro-IRM du poignet soit l'examen de référence pour évaluer ces ligaments, Shahabpour et coll. ont récemment démontré leur visibilité en IRM tridimensionnelle (volumique) après injection iv. de produit de contraste (Gadolinium). L'atteinte ligamentaire associée aux fractures dorsales du triquetrum n'a jusqu'à présent jamais été évalué. Ces lésions pourraient avoir un impact sur l'évolution et la prise en charge de ces fractures. Les objectifs de l'étude étaient donc les suivants: premièrement, déterminer l'ensemble des caractéristiques des fractures dorsales du triquetrum en IRM, en mettant l'accent sur les lésions ligamentaires extrinsèques associées; secondairement, discuter les différents mécanismes physiopathologiques (i.e. avulsion ligamentaire ou impaction ulno-carpienne) de ces fractures d'après nos résultats en IRM. Patients et méthodes: Ceci est une étude rétrospective multicentrique (CHUV, Lausanne; Hôpital Cochin, AP-HP, Paris) d'examens IRM et radiographies conventionnelles du poignet. A partir de janvier 2008, nous avons recherché dans les bases de données institutionnelles les patients présentant une fracture du triquetrum et ayant bénéficié d'une IRM volumique du poignet dans un délai de six semaines entre le traumatisme et l'IRM. Les examens IRM ont été effectués sur deux machines à haut champ magnétique (3 Tesla) avec une antenne dédiée et un protocole d'acquisition incluant une séquence tridimensionnelle isotropique (« 3D VIBE ») après injection iv. de produit de contraste (Gadolinium). Ces examens ont été analysés par deux radiologues ostéo-articulaires expérimentés. Les mesures ont été effectuées par un troisième radiologue ostéo-articulaire. En ce qui concerne l'analyse qualitative, le type de fracture du triquetrum (selon la classification de Garcia-Elias), la distribution de l'oedème osseux post- traumatique, ainsi que le nombre et la distribution des lésions ligamentaires extrinsèques associées ont été évalués. Pour l'analyse quantitative, l'index du processus de la styloïde ulnaire (selon la formule de Garcia-Elias), le volume du fragment osseux détaché du triquetrum, et la distance séparant ce fragment osseux du triquetrum ont été mesurés.
Batch effect confounding leads to strong bias in performance estimates obtained by cross-validation.
Resumo:
BACKGROUND: With the large amount of biological data that is currently publicly available, many investigators combine multiple data sets to increase the sample size and potentially also the power of their analyses. However, technical differences ("batch effects") as well as differences in sample composition between the data sets may significantly affect the ability to draw generalizable conclusions from such studies. FOCUS: The current study focuses on the construction of classifiers, and the use of cross-validation to estimate their performance. In particular, we investigate the impact of batch effects and differences in sample composition between batches on the accuracy of the classification performance estimate obtained via cross-validation. The focus on estimation bias is a main difference compared to previous studies, which have mostly focused on the predictive performance and how it relates to the presence of batch effects. DATA: We work on simulated data sets. To have realistic intensity distributions, we use real gene expression data as the basis for our simulation. Random samples from this expression matrix are selected and assigned to group 1 (e.g., 'control') or group 2 (e.g., 'treated'). We introduce batch effects and select some features to be differentially expressed between the two groups. We consider several scenarios for our study, most importantly different levels of confounding between groups and batch effects. METHODS: We focus on well-known classifiers: logistic regression, Support Vector Machines (SVM), k-nearest neighbors (kNN) and Random Forests (RF). Feature selection is performed with the Wilcoxon test or the lasso. Parameter tuning and feature selection, as well as the estimation of the prediction performance of each classifier, is performed within a nested cross-validation scheme. The estimated classification performance is then compared to what is obtained when applying the classifier to independent data.
Resumo:
Development and environmental issues of small cities in developing countries have largely been overlooked although these settlements are of global demographic importance and often face a "triple challenge"; that is, they have limited financial and human resources to address growing environmental problems that are related to both development (e.g., pollution) and under-development (e.g., inadequate water supply). Neoliberal policy has arguably aggravated this challenge as public investments in infrastructure generally declined while the focus shifted to the metropolitan "economic growth machines". This paper develops a conceptual framework and agenda for the study of small cities in the global south, their environmental dynamics, governance and politics in the current neoliberal context. While small cities are governed in a neoliberal policy context, they are not central to neoliberalism, and their (environmental) governance therefore seems to differ from that of global cities. Furthermore, "actually existing" neoliberal governance of small cities is shaped by the interplay of regional and local politics and environmental situations. The approach of urban political ecology and the concept of rural-urban linkages are used to consider these socio-ecological processes. The conceptual framework and research agenda are illustrated in the case of India, where the agency of small cities in regard to environmental governance seems to remain limited despite formal political decentralization.
Resumo:
The research considers the problem of spatial data classification using machine learning algorithms: probabilistic neural networks (PNN) and support vector machines (SVM). As a benchmark model simple k-nearest neighbor algorithm is considered. PNN is a neural network reformulation of well known nonparametric principles of probability density modeling using kernel density estimator and Bayesian optimal or maximum a posteriori decision rules. PNN is well suited to problems where not only predictions but also quantification of accuracy and integration of prior information are necessary. An important property of PNN is that they can be easily used in decision support systems dealing with problems of automatic classification. Support vector machine is an implementation of the principles of statistical learning theory for the classification tasks. Recently they were successfully applied for different environmental topics: classification of soil types and hydro-geological units, optimization of monitoring networks, susceptibility mapping of natural hazards. In the present paper both simulated and real data case studies (low and high dimensional) are considered. The main attention is paid to the detection and learning of spatial patterns by the algorithms applied.
Resumo:
Machine learning has been largely applied to analyze data in various domains, but it is still new to personalized medicine, especially dose individualization. In this paper, we focus on the prediction of drug concentrations using Support Vector Machines (S VM) and the analysis of the influence of each feature to the prediction results. Our study shows that SVM-based approaches achieve similar prediction results compared with pharmacokinetic model. The two proposed example-based SVM methods demonstrate that the individual features help to increase the accuracy in the predictions of drug concentration with a reduced library of training data.
Resumo:
This article presents an experimental study about the classification ability of several classifiers for multi-classclassification of cannabis seedlings. As the cultivation of drug type cannabis is forbidden in Switzerland lawenforcement authorities regularly ask forensic laboratories to determinate the chemotype of a seized cannabisplant and then to conclude if the plantation is legal or not. This classification is mainly performed when theplant is mature as required by the EU official protocol and then the classification of cannabis seedlings is a timeconsuming and costly procedure. A previous study made by the authors has investigated this problematic [1]and showed that it is possible to differentiate between drug type (illegal) and fibre type (legal) cannabis at anearly stage of growth using gas chromatography interfaced with mass spectrometry (GC-MS) based on therelative proportions of eight major leaf compounds. The aims of the present work are on one hand to continueformer work and to optimize the methodology for the discrimination of drug- and fibre type cannabisdeveloped in the previous study and on the other hand to investigate the possibility to predict illegal cannabisvarieties. Seven classifiers for differentiating between cannabis seedlings are evaluated in this paper, namelyLinear Discriminant Analysis (LDA), Partial Least Squares Discriminant Analysis (PLS-DA), Nearest NeighbourClassification (NNC), Learning Vector Quantization (LVQ), Radial Basis Function Support Vector Machines(RBF SVMs), Random Forest (RF) and Artificial Neural Networks (ANN). The performance of each method wasassessed using the same analytical dataset that consists of 861 samples split into drug- and fibre type cannabiswith drug type cannabis being made up of 12 varieties (i.e. 12 classes). The results show that linear classifiersare not able to manage the distribution of classes in which some overlap areas exist for both classificationproblems. Unlike linear classifiers, NNC and RBF SVMs best differentiate cannabis samples both for 2-class and12-class classifications with average classification results up to 99% and 98%, respectively. Furthermore, RBFSVMs correctly classified into drug type cannabis the independent validation set, which consists of cannabisplants coming from police seizures. In forensic case work this study shows that the discrimination betweencannabis samples at an early stage of growth is possible with fairly high classification performance fordiscriminating between cannabis chemotypes or between drug type cannabis varieties.