971 resultados para Scale invariant feature transform


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Uncorrelated random scale-free networks are useful null models to check the accuracy and the analytical solutions of dynamical processes defined on complex networks. We propose and analyze a model capable of generating random uncorrelated scale-free networks with no multiple and self-connections. The model is based on the classical configuration model, with an additional restriction on the maximum possible degree of the vertices. We check numerically that the proposed model indeed generates scale-free networks with no two- and three-vertex correlations, as measured by the average degree of the nearest neighbors and the clustering coefficient of the vertices of degree k, respectively.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

AIM: Phylogenetic diversity patterns are increasingly being used to better understand the role of ecological and evolutionary processes in community assembly. Here, we quantify how these patterns are influenced by scale choices in terms of spatial and environmental extent and organismic scales. LOCATION: European Alps. METHODS: We applied 42 sampling strategies differing in their combination of focal scales. For each resulting sub-dataset, we estimated the phylogenetic diversity of the species pools, phylogenetic α-diversities of local communities, and statistics commonly used together with null models in order to infer non-random diversity patterns (i.e. phylogenetic clustering versus over-dispersion). Finally, we studied the effects of scale choices on these measures using regression analyses. RESULTS: Scale choices were decisive for revealing signals in diversity patterns. Notably, changes in focal scales sometimes reversed a pattern of over-dispersion into clustering. Organismic scale had a stronger effect than spatial and environmental extent. However, we did not find general rules for the direction of change from over-dispersion to clustering with changing scales. Importantly, these scale issues had only a weak influence when focusing on regional diversity patterns that change along abiotic gradients. MAIN CONCLUSIONS: Our results call for caution when combining phylogenetic data with distributional data to study how and why communities differ from random expectations of phylogenetic relatedness. These analyses seem to be robust when the focus is on relating community diversity patterns to variation in habitat conditions, such as abiotic gradients. However, if the focus is on identifying relevant assembly rules for local communities, the uncertainty arising from a certain scale choice can be immense. In the latter case, it becomes necessary to test whether emerging patterns are robust to alternative scale choices.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

MOTIVATION: Analysis of millions of pyro-sequences is currently playing a crucial role in the advance of environmental microbiology. Taxonomy-independent, i.e. unsupervised, clustering of these sequences is essential for the definition of Operational Taxonomic Units. For this application, reproducibility and robustness should be the most sought after qualities, but have thus far largely been overlooked. RESULTS: More than 1 million hyper-variable internal transcribed spacer 1 (ITS1) sequences of fungal origin have been analyzed. The ITS1 sequences were first properly extracted from 454 reads using generalized profiles. Then, otupipe, cd-hit-454, ESPRIT-Tree and DBC454, a new algorithm presented here, were used to analyze the sequences. A numerical assay was developed to measure the reproducibility and robustness of these algorithms. DBC454 was the most robust, closely followed by ESPRIT-Tree. DBC454 features density-based hierarchical clustering, which complements the other methods by providing insights into the structure of the data. AVAILABILITY: An executable is freely available for non-commercial users at ftp://ftp.vital-it.ch/tools/dbc454. It is designed to run under MPI on a cluster of 64-bit Linux machines running Red Hat 4.x, or on a multi-core OSX system. CONTACT: dbc454@vital-it.ch or nicolas.guex@isb-sib.ch.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Abstract : This work is concerned with the development and application of novel unsupervised learning methods, having in mind two target applications: the analysis of forensic case data and the classification of remote sensing images. First, a method based on a symbolic optimization of the inter-sample distance measure is proposed to improve the flexibility of spectral clustering algorithms, and applied to the problem of forensic case data. This distance is optimized using a loss function related to the preservation of neighborhood structure between the input space and the space of principal components, and solutions are found using genetic programming. Results are compared to a variety of state-of--the-art clustering algorithms. Subsequently, a new large-scale clustering method based on a joint optimization of feature extraction and classification is proposed and applied to various databases, including two hyperspectral remote sensing images. The algorithm makes uses of a functional model (e.g., a neural network) for clustering which is trained by stochastic gradient descent. Results indicate that such a technique can easily scale to huge databases, can avoid the so-called out-of-sample problem, and can compete with or even outperform existing clustering algorithms on both artificial data and real remote sensing images. This is verified on small databases as well as very large problems. Résumé : Ce travail de recherche porte sur le développement et l'application de méthodes d'apprentissage dites non supervisées. Les applications visées par ces méthodes sont l'analyse de données forensiques et la classification d'images hyperspectrales en télédétection. Dans un premier temps, une méthodologie de classification non supervisée fondée sur l'optimisation symbolique d'une mesure de distance inter-échantillons est proposée. Cette mesure est obtenue en optimisant une fonction de coût reliée à la préservation de la structure de voisinage d'un point entre l'espace des variables initiales et l'espace des composantes principales. Cette méthode est appliquée à l'analyse de données forensiques et comparée à un éventail de méthodes déjà existantes. En second lieu, une méthode fondée sur une optimisation conjointe des tâches de sélection de variables et de classification est implémentée dans un réseau de neurones et appliquée à diverses bases de données, dont deux images hyperspectrales. Le réseau de neurones est entraîné à l'aide d'un algorithme de gradient stochastique, ce qui rend cette technique applicable à des images de très haute résolution. Les résultats de l'application de cette dernière montrent que l'utilisation d'une telle technique permet de classifier de très grandes bases de données sans difficulté et donne des résultats avantageusement comparables aux méthodes existantes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Even 30 years after its first publication the Glasgow Coma Scale (GCS) is still used worldwide to describe and assess coma. The GCS consists of three components, the ocular, motor and verbal response to standardized stimulation, and is used as a severity of illness indicator for coma of various origins. The GCS facilitates information transfer and monitoring changes in coma. In addition, it is used as a triage tool in patients with traumatic brain injury. Its prognostic value regarding the outcome after a traumatic brain injury still lacks evidence. One of the main problems is the evaluation of the GCS in sedated, paralysed and/or intubated patients. A multitude of pseudoscores exists but a universal definition has yet to be defined.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We conduct a large-scale comparative study on linearly combining superparent-one-dependence estimators (SPODEs), a popular family of seminaive Bayesian classifiers. Altogether, 16 model selection and weighing schemes, 58 benchmark data sets, and various statistical tests are employed. This paper's main contributions are threefold. First, it formally presents each scheme's definition, rationale, and time complexity and hence can serve as a comprehensive reference for researchers interested in ensemble learning. Second, it offers bias-variance analysis for each scheme's classification error performance. Third, it identifies effective schemes that meet various needs in practice. This leads to accurate and fast classification algorithms which have an immediate and significant impact on real-world applications. Another important feature of our study is using a variety of statistical tests to evaluate multiple learning methods across multiple data sets.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we propose an innovative methodology for automated profiling of illicit tablets bytheir surface granularity; a feature previously unexamined for this purpose. We make use of the tinyinconsistencies at the tablet surface, referred to as speckles, to generate a quantitative granularity profileof tablets. Euclidian distance is used as a measurement of (dis)similarity between granularity profiles.The frequency of observed distances is then modelled by kernel density estimation in order to generalizethe observations and to calculate likelihood ratios (LRs). The resulting LRs are used to evaluate thepotential of granularity profiles to differentiate between same-batch and different-batches tablets.Furthermore, we use the LRs as a similarity metric to refine database queries. We are able to derivereliable LRs within a scope that represent the true evidential value of the granularity feature. Thesemetrics are used to refine candidate hit-lists form a database containing physical features of illicittablets. We observe improved or identical ranking of candidate tablets in 87.5% of cases when granularityis considered.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The presynaptic plasma membrane (PSPM) of cholinergic nerve terminals was purified from Torpedo electric organ using a large-scale procedure. Up to 500 g of frozen electric organ were fractioned in a single run, leading to the isolation of greater than 100 mg of PSPM proteins. The purity of the fraction is similar to that of the synaptosomal plasma membrane obtained after subfractionation of Torpedo synaptosomes as judged by its membrane-bound acetylcholinesterase activity, the number of Glycera convoluta neurotoxin binding sites, and the binding of two monoclonal antibodies directed against PSPM. The specificity of these antibodies for the PSPM is demonstrated by immunofluorescence microscopy.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background: Network reconstructions at the cell level are a major development in Systems Biology. However, we are far from fully exploiting its potentialities. Often, the incremental complexity of the pursued systems overrides experimental capabilities, or increasingly sophisticated protocols are underutilized to merely refine confidence levels of already established interactions. For metabolic networks, the currently employed confidence scoring system rates reactions discretely according to nested categories of experimental evidence or model-based likelihood. Results: Here, we propose a complementary network-based scoring system that exploits the statistical regularities of a metabolic network as a bipartite graph. As an illustration, we apply it to the metabolism of Escherichia coli. The model is adjusted to the observations to derive connection probabilities between individual metabolite-reaction pairs and, after validation, to assess the reliability of each reaction in probabilistic terms. This network-based scoring system uncovers very specific reactions that could be functionally or evolutionary important, identifies prominent experimental targets, and enables further confirmation of modeling results. Conclusions: We foresee a wide range of potential applications at different sub-cellular or supra-cellular levels of biological interactions given the natural bipartivity of many biological networks.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The nanometer¿scale oxidation of Si(100) surfaces in air is performed with an atomic force microscope working in tapping mode. Applying a positive voltage to the sample with respect to the tip, two kinds of modifications are induced on the sample: grown silicon oxide mounds less than 5 nm high and mounds higher than 10 nm (which are assumed to be gold depositions). The threshold voltage necessary to produce the modification is studied as a function of the average tip¿to¿sample distance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Debris flows are among the most dangerous processes in mountainous areas due to their rapid rate of movement and long runout zone. Sudden and rather unexpected impacts produce not only damages to buildings and infrastructure but also threaten human lives. Medium- to regional-scale susceptibility analyses allow the identification of the most endangered areas and suggest where further detailed studies have to be carried out. Since data availability for larger regions is mostly the key limiting factor, empirical models with low data requirements are suitable for first overviews. In this study a susceptibility analysis was carried out for the Barcelonnette Basin, situated in the southern French Alps. By means of a methodology based on empirical rules for source identification and the empirical angle of reach concept for the 2-D runout computation, a worst-case scenario was first modelled. In a second step, scenarios for high, medium and low frequency events were developed. A comparison with the footprints of a few mapped events indicates reasonable results but suggests a high dependency on the quality of the digital elevation model. This fact emphasises the need for a careful interpretation of the results while remaining conscious of the inherent assumptions of the model used and quality of the input data.