18 resultados para k-Means algorithm

em Université de Lausanne, Switzerland


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract: To cluster textual sequence types (discourse types/modes) in French texts, K-means algorithm with high-dimensional embeddings and fuzzy clustering algorithm were applied on clauses whose POS (part-ofspeech) n-gram profiles were previously extracted. Uni-, bi- and trigrams were used on four 19th century French short stories by Maupassant. For high-dimensional embeddings, power transformations on the chi-squared distances between clauses were explored. Preliminary results show that highdimensional embeddings improve the quality of clustering, contrasting the use of bi and trigrams whose performance is disappointing, possibly because of feature space sparsity.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The in situ hybridization Allen Mouse Brain Atlas was mined for proteases expressed in the somatosensory cerebral cortex. Among the 480 genes coding for protease/peptidases, only four were found enriched in cortical interneurons: Reln coding for reelin; Adamts8 and Adamts15 belonging to the class of metzincin proteases involved in reshaping the perineuronal net (PNN) and Mme encoding for Neprilysin, the enzyme degrading amyloid β-peptides. The pattern of expression of metalloproteases (MPs) was analyzed by single-cell reverse transcriptase multiplex PCR after patch clamp and was compared with the expression of 10 canonical interneurons markers and 12 additional genes from the Allen Atlas. Clustering of these genes by K-means algorithm displays five distinct clusters. Among these five clusters, two fast-spiking interneuron clusters expressing the calcium-binding protein Pvalb were identified, one co-expressing Pvalb with Sst (PV-Sst) and another co-expressing Pvalb with three metallopeptidases Adamts8, Adamts15 and Mme (PV-MP). By using Wisteria floribunda agglutinin, a specific marker for PNN, PV-MP interneurons were found surrounded by PNN, whereas the ones expressing Sst, PV-Sst, were not.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

*This study reconstructs the phylogeography of Aegilops geniculata, an allotetraploid relative of wheat, to discuss the impact of past climate changes and recent human activities (e.g. the early expansion of agriculture) on the genetic diversity of ruderal plant species. *We combined chloroplast DNA (cpDNA) sequencing, analysed using statistical parsimony network, with nonhierarchical K-means clustering of amplified fragment length polymorphism (AFLP) genotyping, to unravel patterns of genetic structure across the native range of Ae. geniculata. The AFLP dataset was further explored by measurement of the regional genetic diversity and the detection of isolation by distance patterns. *Both cpDNA and AFLP suggest an eastern Mediterranean origin of Ae. geniculata. Two lineages have spread independently over northern and southern Mediterranean areas. Northern populations show low genetic diversity but strong phylogeographical structure among the main peninsulas, indicating a major influence of glacial cycles. By contrast, low genetic structuring and a high genetic diversity are detected in southern Mediterranean populations. Finally, we highlight human-mediated dispersal resulting in substantial introgression between resident and migrant populations. *We have shown that the evolutionary trajectories of ruderal plants can be similar to those of wild species, but are interfered by human activities, promoting range expansions through increased long-distance dispersal and the creation of suitable habitats.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Axée dans un premier temps sur le formalisme et les méthodes, cette thèse est construite sur trois concepts formalisés: une table de contingence, une matrice de dissimilarités euclidiennes et une matrice d'échange. À partir de ces derniers, plusieurs méthodes d'Analyse des données ou d'apprentissage automatique sont exprimées et développées: l'analyse factorielle des correspondances (AFC), vue comme un cas particulier du multidimensional scaling; la classification supervisée, ou non, combinée aux transformations de Schoenberg; et les indices d'autocorrélation et d'autocorrélation croisée, adaptés à des analyses multivariées et permettant de considérer diverses familles de voisinages. Ces méthodes débouchent dans un second temps sur une pratique de l'analyse exploratoire de différentes données textuelles et musicales. Pour les données textuelles, on s'intéresse à la classification automatique en types de discours de propositions énoncées, en se basant sur les catégories morphosyntaxiques (CMS) qu'elles contiennent. Bien que le lien statistique entre les CMS et les types de discours soit confirmé, les résultats de la classification obtenus avec la méthode K- means, combinée à une transformation de Schoenberg, ainsi qu'avec une variante floue de l'algorithme K-means, sont plus difficiles à interpréter. On traite aussi de la classification supervisée multi-étiquette en actes de dialogue de tours de parole, en se basant à nouveau sur les CMS qu'ils contiennent, mais aussi sur les lemmes et le sens des verbes. Les résultats obtenus par l'intermédiaire de l'analyse discriminante combinée à une transformation de Schoenberg sont prometteurs. Finalement, on examine l'autocorrélation textuelle, sous l'angle des similarités entre diverses positions d'un texte, pensé comme une séquence d'unités. En particulier, le phénomène d'alternance de la longueur des mots dans un texte est observé pour des voisinages d'empan variable. On étudie aussi les similarités en fonction de l'apparition, ou non, de certaines parties du discours, ainsi que les similarités sémantiques des diverses positions d'un texte. Concernant les données musicales, on propose une représentation d'une partition musicale sous forme d'une table de contingence. On commence par utiliser l'AFC et l'indice d'autocorrélation pour découvrir les structures existant dans chaque partition. Ensuite, on opère le même type d'approche sur les différentes voix d'une partition, grâce à l'analyse des correspondances multiples, dans une variante floue, et à l'indice d'autocorrélation croisée. Qu'il s'agisse de la partition complète ou des différentes voix qu'elle contient, des structures répétées sont effectivement détectées, à condition qu'elles ne soient pas transposées. Finalement, on propose de classer automatiquement vingt partitions de quatre compositeurs différents, chacune représentée par une table de contingence, par l'intermédiaire d'un indice mesurant la similarité de deux configurations. Les résultats ainsi obtenus permettent de regrouper avec succès la plupart des oeuvres selon leur compositeur.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

OBJECTIVE: In contrast to conventional (CONV) neuromuscular electrical stimulation (NMES), the use of "wide-pulse, high-frequencies" (WPHF) can generate higher forces than expected by the direct activation of motor axons alone. We aimed at investigating the occurrence, magnitude, variability and underlying neuromuscular mechanisms of these "Extra Forces" (EF). METHODS: Electrically-evoked isometric plantar flexion force was recorded in 42 healthy subjects. Additionally, twitch potentiation, H-reflex and M-wave responses were assessed in 13 participants. CONV (25Hz, 0.05ms) and WPHF (100Hz, 1ms) NMES consisted of five stimulation trains (20s on-90s off). RESULTS: K-means clustering analysis disclosed a responder rate of almost 60%. Within this group of responders, force significantly increased from 4% to 16% of the maximal voluntary contraction force and H-reflexes were depressed after WPHF NMES. In contrast, non-responders showed neither EF nor H-reflex depression. Twitch potentiation and resting EMG data were similar between groups. Interestingly, a large inter- and intrasubject variability of EF was observed. CONCLUSION: The responder percentage was overestimated in previous studies. SIGNIFICANCE: This study proposes a novel methodological framework for unraveling the neurophysiological mechanisms involved in EF and provides further evidence for a central contribution to EF in responders.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Conventional (CONV) neuromuscular electrical stimulation (NMES) (i.e., short pulse duration, low frequencies) induces a higher energetic response as compared to voluntary contractions (VOL). In contrast, wide-pulse, high-frequency (WPHF) NMES might elicit-at least in some subjects (i.e., responders)-a different motor unit recruitment compared to CONV that resembles the physiological muscle activation pattern of VOL. We therefore hypothesized that for these responder subjects, the metabolic demand of WPHF would be lower than CONV and comparable to VOL. 18 healthy subjects performed isometric plantar flexions at 10% of their maximal voluntary contraction force for CONV (25 Hz, 0.05 ms), WPHF (100 Hz, 1 ms) and VOL protocols. For each protocol, force time integral (FTI) was quantified and subjects were classified as responders and non-responders to WPHF based on k-means clustering analysis. Furthermore, a fatigue index based on FTI loss at the end of each protocol compared with the beginning of the protocol was calculated. Phosphocreatine depletion (ΔPCr) was assessed using 31P magnetic resonance spectroscopy. Responders developed four times higher FTI's during WPHF (99 ± 37 ×103 N.s) than non-responders (26 ± 12 ×103 N.s). For both responders and non-responders, CONV was metabolically more demanding than VOL when ΔPCr was expressed relative to the FTI. Only for the responder group, the ∆PCr/FTI ratio of WPHF (0.74 ± 0.19 M/N.s) was significantly lower compared to CONV (1.48 ± 0.46 M/N.s) but similar to VOL (0.65 ± 0.21 M/N.s). Moreover, the fatigue index was not different between WPHF (-16%) and CONV (-25%) for the responders. WPHF could therefore be considered as the less demanding NMES modality-at least in this subgroup of subjects-by possibly exhibiting a muscle activation pattern similar to VOL contractions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Waveform tomographic imaging of crosshole georadar data is a powerful method to investigate the shallow subsurface because of its ability to provide images of pertinent petrophysical parameters with extremely high spatial resolution. All current crosshole georadar waveform inversion strategies are based on the assumption of frequency-independent electromagnetic constitutive parameters. However, in reality, these parameters are known to be frequency-dependent and complex and thus recorded georadar data may show significant dispersive behavior. In this paper, we evaluate synthetically the reconstruction limits of a recently published crosshole georadar waveform inversion scheme in the presence of varying degrees of dielectric dispersion. Our results indicate that, when combined with a source wavelet estimation procedure that provides a means of partially accounting for the frequency-dependent effects through an "effective" wavelet, the inversion algorithm performs remarkably well in weakly to moderately dispersive environments and has the ability to provide adequate tomographic reconstructions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

High-resolution tomographic imaging of the shallow subsurface is becoming increasingly important for a wide range of environmental, hydrological and engineering applications. Because of their superior resolution power, their sensitivity to pertinent petrophysical parameters, and their far reaching complementarities, both seismic and georadar crosshole imaging are of particular importance. To date, corresponding approaches have largely relied on asymptotic, ray-based approaches, which only account for a very small part of the observed wavefields, inherently suffer from a limited resolution, and in complex environments may prove to be inadequate. These problems can potentially be alleviated through waveform inversion. We have developed an acoustic waveform inversion approach for crosshole seismic data whose kernel is based on a finite-difference time-domain (FDTD) solution of the 2-D acoustic wave equations. This algorithm is tested on and applied to synthetic data from seismic velocity models of increasing complexity and realism and the results are compared to those obtained using state-of-the-art ray-based traveltime tomography. Regardless of the heterogeneity of the underlying models, the waveform inversion approach has the potential of reliably resolving both the geometry and the acoustic properties of features of the size of less than half a dominant wavelength. Our results do, however, also indicate that, within their inherent resolution limits, ray-based approaches provide an effective and efficient means to obtain satisfactory tomographic reconstructions of the seismic velocity structure in the presence of mild to moderate heterogeneity and in absence of strong scattering. Conversely, the excess effort of waveform inversion provides the greatest benefits for the most heterogeneous, and arguably most realistic, environments where multiple scattering effects tend to be prevalent and ray-based methods lose most of their effectiveness.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Genes underlying mutant phenotypes can be isolated by combining marker discovery, genetic mapping and resequencing, but a more straightforward strategy for mapping mutations would be the direct comparison of mutant and wild-type genomes. Applying such an approach, however, is hampered by the need for reference sequences and by mutational loads that confound the unambiguous identification of causal mutations. Here we introduce NIKS (needle in the k-stack), a reference-free algorithm based on comparing k-mers in whole-genome sequencing data for precise discovery of homozygous mutations. We applied NIKS to eight mutants induced in nonreference rice cultivars and to two mutants of the nonmodel species Arabis alpina. In both species, comparing pooled F2 individuals selected for mutant phenotypes revealed small sets of mutations including the causal changes. Moreover, comparing M3 seedlings of two allelic mutants unambiguously identified the causal gene. Thus, for any species amenable to mutagenesis, NIKS enables forward genetics without requiring segregating populations, genetic maps and reference sequences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Positron emission tomography with [18F] fluorodeoxyglucose (FDG-PET) plays a well-established role in assisting early detection of frontotemporal lobar degeneration (FTLD). Here, we examined the impact of intensity normalization to different reference areas on accuracy of FDG-PET to discriminate between patients with mild FTLD and healthy elderly subjects. FDG-PET was conducted at two centers using different acquisition protocols: 41 FTLD patients and 42 controls were studied at center 1, 11 FTLD patients and 13 controls were studied at center 2. All PET images were intensity normalized to the cerebellum, primary sensorimotor cortex (SMC), cerebral global mean (CGM), and a reference cluster with most preserved FDG uptake in the aforementioned patients group of center 1. Metabolic deficits in the patient group at center 1 appeared 1.5, 3.6, and 4.6 times greater in spatial extent, when tracer uptake was normalized to the reference cluster rather than to the cerebellum, SMC, and CGM, respectively. Logistic regression analyses based on normalized values from FTLD-typical regions showed that at center 1, cerebellar, SMC, CGM, and cluster normalizations differentiated patients from controls with accuracies of 86%, 76%, 75% and 90%, respectively. A similar order of effects was found at center 2. Cluster normalization leads to a significant increase of statistical power in detecting early FTLD-associated metabolic deficits. The established FTLD-specific cluster can be used to improve detection of FTLD on a single case basis at independent centers - a decisive step towards early diagnosis and prediction of FTLD syndromes enabling specific therapies in the future.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE: We investigated whether the INTERMED, a generic instrument for assessing biopsychosocial case complexity and direct care, identifies organ transplant patients at risk of unfavourable post-transplant development by comparing it to the Transplant Evaluation Rating Scale (TERS), the established measure for pretransplant psychosocial evaluation. METHOD: One hundred nineteen kidney, liver, and heart transplant candidates were evaluated using the INTERMED, TERS, SF-36, EuroQol, Montgomery-Åsberg Depression Rating Scale (MADRS), and Hospital Anxiety & Depression Scale (HADS). RESULTS: We found significant relationships between the INTERMED and the TERS scores. The INTERMED highly correlated with the HADS,MADRS, and mental and physical health scores of the SF-36 Health Survey. CONCLUSIONS: The results demonstrate the validity and usefulness of the INTERMED instrument for pretransplant evaluation. Furthermore, our findings demonstrate the different qualities of INTERMED and TERS in clinical practice. The advantages of the psychiatric focus of the TERS and the biopsychosocial perspective of the INTERMED are discussed in the context of current literature on integrated care.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Polochic and Motagua faults define the active plate boundary between the North American and Caribbean plates in central Guatemala. A splay of the Polochic Fault traverses the rapidly growing city of San Miguel Uspantan that is periodically affected by destructive earthquakes. This fault splay was located using a 2D electrical resistivity tomography (ERT) survey that also characterized the fault damage zone and evaluated the thickness and nature of recent deposits upon which most of the city is built. ERT images show the fault as a similar to 50 m wide, near-vertical low-resistivity anomaly, bounded within a few meters by high resistivity anomalies. Forward modeling reproduces the key aspects of the observed electrical resistivity data with remarkable fidelity thus defining the overall location, geometry, and internal structure of the fault zone as well as the affected lithologies. Our results indicate that the city is constructed on a similar to 20 m thick surficial layer consisting of poorly consolidated, highly porous, water-logged pumice. This soft layer is likely to amplify seismic waves and to liquefy upon moderate to strong ground shaking. The electrical conductivity as well as the major element chemistry of the groundwater provides evidence to suggest that the local aquifer might, at least in part, be fed by water rising along the fault. Therefore, the potential threat posed by this fault splay may not be limited to its seismic activity per se, but could be compounded its potential propensity to enhance seismic site effects by injecting water into the soft surficial sediments. The results of this study provide the basis for a rigorous analysis of seismic hazard and sustainable development of San Miguel Uspantan and illustrate the potential of ERT surveying for paleoseismic studies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fetal MRI reconstruction aims at finding a high-resolution image given a small set of low-resolution images. It is usually modeled as an inverse problem where the regularization term plays a central role in the reconstruction quality. Literature has considered several regularization terms s.a. Dirichlet/Laplacian energy, Total Variation (TV)- based energies and more recently non-local means. Although TV energies are quite attractive because of their ability in edge preservation, standard explicit steepest gradient techniques have been applied to optimize fetal-based TV energies. The main contribution of this work lies in the introduction of a well-posed TV algorithm from the point of view of convex optimization. Specifically, our proposed TV optimization algorithm or fetal reconstruction is optimal w.r.t. the asymptotic and iterative convergence speeds O(1/n2) and O(1/√ε), while existing techniques are in O(1/n2) and O(1/√ε). We apply our algorithm to (1) clinical newborn data, considered as ground truth, and (2) clinical fetal acquisitions. Our algorithm compares favorably with the literature in terms of speed and accuracy.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Active screening by mobile teams is considered the best method for detecting human African trypanosomiasis (HAT) caused by Trypanosoma brucei gambiense but the current funding context in many post-conflict countries limits this approach. As an alternative, non-specialist health care workers (HCWs) in peripheral health facilities could be trained to identify potential cases who need testing based on their symptoms. We explored the predictive value of syndromic referral algorithms to identify symptomatic cases of HAT among a treatment-seeking population in Nimule, South Sudan. METHODOLOGY/PRINCIPAL FINDINGS: Symptom data from 462 patients (27 cases) presenting for a HAT test via passive screening over a 7 month period were collected to construct and evaluate over 14,000 four item syndromic algorithms considered simple enough to be used by peripheral HCWs. For comparison, algorithms developed in other settings were also tested on our data, and a panel of expert HAT clinicians were asked to make referral decisions based on the symptom dataset. The best performing algorithms consisted of three core symptoms (sleep problems, neurological problems and weight loss), with or without a history of oedema, cervical adenopathy or proximity to livestock. They had a sensitivity of 88.9-92.6%, a negative predictive value of up to 98.8% and a positive predictive value in this context of 8.4-8.7%. In terms of sensitivity, these out-performed more complex algorithms identified in other studies, as well as the expert panel. The best-performing algorithm is predicted to identify about 9/10 treatment-seeking HAT cases, though only 1/10 patients referred would test positive. CONCLUSIONS/SIGNIFICANCE: In the absence of regular active screening, improving referrals of HAT patients through other means is essential. Systematic use of syndromic algorithms by peripheral HCWs has the potential to increase case detection and would increase their participation in HAT programmes. The algorithms proposed here, though promising, should be validated elsewhere.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Simulated-annealing-based conditional simulations provide a flexible means of quantitatively integrating diverse types of subsurface data. Although such techniques are being increasingly used in hydrocarbon reservoir characterization studies, their potential in environmental, engineering and hydrological investigations is still largely unexploited. Here, we introduce a novel simulated annealing (SA) algorithm geared towards the integration of high-resolution geophysical and hydrological data which, compared to more conventional approaches, provides significant advancements in the way that large-scale structural information in the geophysical data is accounted for. Model perturbations in the annealing procedure are made by drawing from a probability distribution for the target parameter conditioned to the geophysical data. This is the only place where geophysical information is utilized in our algorithm, which is in marked contrast to other approaches where model perturbations are made through the swapping of values in the simulation grid and agreement with soft data is enforced through a correlation coefficient constraint. Another major feature of our algorithm is the way in which available geostatistical information is utilized. Instead of constraining realizations to match a parametric target covariance model over a wide range of spatial lags, we constrain the realizations only at smaller lags where the available geophysical data cannot provide enough information. Thus we allow the larger-scale subsurface features resolved by the geophysical data to have much more due control on the output realizations. Further, since the only component of the SA objective function required in our approach is a covariance constraint at small lags, our method has improved convergence and computational efficiency over more traditional methods. Here, we present the results of applying our algorithm to the integration of porosity log and tomographic crosshole georadar data to generate stochastic realizations of the local-scale porosity structure. Our procedure is first tested on a synthetic data set, and then applied to data collected at the Boise Hydrogeophysical Research Site.