822 resultados para discriminant analysis and cluster analysis
Resumo:
Objective To assess the impact of exercise referral schemes on physical activity and health outcomes. Design Systematic review and meta-analysis. Data sources Medline, Embase, PsycINFO, Cochrane Library, ISI Web of Science, SPORTDiscus, and ongoing trial registries up to October 2009. We also checked study references. Study selection - Design: randomised controlled trials or non-randomised controlled (cluster or individual) studies published in peer review journals. - Population: sedentary individuals with or without medical diagnosis. - Exercise referral schemes defined as: clear referrals by primary care professionals to third party service providers to increase physical activity or exercise, physical activity or exercise programmes tailored to individuals, and initial assessment and monitoring throughout programmes. - Comparators: usual care, no intervention, or alternative exercise referral schemes. Results Eight randomised controlled trials met the inclusion criteria, comparing exercise referral schemes with usual care (six trials), alternative physical activity intervention (two), and an exercise referral scheme plus a self determination theory intervention (one). Compared with usual care, follow-up data for exercise referral schemes showed an increased number of participants who achieved 90-150 minutes of physical activity of at least moderate intensity per week (pooled relative risk 1.16, 95% confidence intervals 1.03 to 1.30) and a reduced level of depression (pooled standardised mean difference −0.82, −1.28 to −0.35). Evidence of a between group difference in physical activity of moderate or vigorous intensity or in other health outcomes was inconsistent at follow-up. We did not find any difference in outcomes between exercise referral schemes and the other two comparator groups. None of the included trials separately reported outcomes in individuals with specific medical diagnoses. Substantial heterogeneity in the quality and nature of the exercise referral schemes across studies might have contributed to the inconsistency in outcome findings. Conclusions Considerable uncertainty remains as to the effectiveness of exercise referral schemes for increasing physical activity, fitness, or health indicators, or whether they are an efficient use of resources for sedentary people with or without a medical diagnosis.
Resumo:
Brassica napus is one of the most important oil crops in the world, and stem rot caused by the fungus Sclerotinia sclerotiorum results in major losses in yield and quality. To elucidate resistance genes and pathogenesis-related genes, genome-wide association analysis of 347 accessions was performed using the Illumina 60K Brassica SNP (single nucleotide polymorphism) array. In addition, the detached stem inoculation assay was used to select five highly resistant (R) and susceptible (S) B. napus lines, 48 h postinoculation with S. sclerotiorum for transcriptome sequencing. We identified 17 significant associations for stem resistance on chromosomes A8 and C6, five of which were on A8 and 12 on C6. The SNPs identified on A8 were located in a 409-kb haplotype block, and those on C6 were consistent with previous QTL mapping efforts. Transcriptome analysis suggested that S. sclerotiorum infection activates the immune system, sulphur metabolism, especially glutathione (GSH) and glucosinolates in both R and S genotypes. Genes found to be specific to the R genotype related to the jasmonic acid pathway, lignin biosynthesis, defence response, signal transduction and encoding transcription factors. Twenty-four genes were identified in both the SNP-trait association and transcriptome sequencing analyses, including a tau class glutathione S-transferase (GSTU) gene cluster. This study provides useful insight into the molecular mechanisms underlying the plant's response to S. sclerotiorum.
Resumo:
In meteorology, observations and forecasts of a wide range of phenomena for example, snow, clouds, hail, fog, and tornados can be categorical, that is, they can only have discrete values (e.g., "snow" and "no snow"). Concentrating on satellite-based snow and cloud analyses, this thesis explores methods that have been developed for evaluation of categorical products and analyses. Different algorithms for satellite products generate different results; sometimes the differences are subtle, sometimes all too visible. In addition to differences between algorithms, the satellite products are influenced by physical processes and conditions, such as diurnal and seasonal variation in solar radiation, topography, and land use. The analysis of satellite-based snow cover analyses from NOAA, NASA, and EUMETSAT, and snow analyses for numerical weather prediction models from FMI and ECMWF was complicated by the fact that we did not have the true knowledge of snow extent, and we were forced simply to measure the agreement between different products. The Sammon mapping, a multidimensional scaling method, was then used to visualize the differences between different products. The trustworthiness of the results for cloud analyses [EUMETSAT Meteorological Products Extraction Facility cloud mask (MPEF), together with the Nowcasting Satellite Application Facility (SAFNWC) cloud masks provided by Météo-France (SAFNWC/MSG) and the Swedish Meteorological and Hydrological Institute (SAFNWC/PPS)] compared with ceilometers of the Helsinki Testbed was estimated by constructing confidence intervals (CIs). Bootstrapping, a statistical resampling method, was used to construct CIs, especially in the presence of spatial and temporal correlation. The reference data for validation are constantly in short supply. In general, the needs of a particular project drive the requirements for evaluation, for example, for the accuracy and the timeliness of the particular data and methods. In this vein, we discuss tentatively how data provided by general public, e.g., photos shared on the Internet photo-sharing service Flickr, can be used as a new source for validation. Results show that they are of reasonable quality and their use for case studies can be warmly recommended. Last, the use of cluster analysis on meteorological in-situ measurements was explored. The Autoclass algorithm was used to construct compact representations of synoptic conditions of fog at Finnish airports.
Resumo:
Background. Several types of networks, such as transcriptional, metabolic or protein-protein interaction networks of various organisms have been constructed, that have provided a variety of insights into metabolism and regulation. Here, we seek to exploit the reaction-based networks of three organisms for comparative genomics. We use concepts from spectral graph theory to systematically determine how differences in basic metabolism of organisms are reflected at the systems level and in the overall topological structures of their metabolic networks. Methodology/Principal Findings. Metabolome-based reaction networks of Mycobacterium tuberculosis, Mycobacterium leprae and Escherichia coli have been constructed based on the KEGG LIGAND database, followed by graph spectral analysis of the network to identify hubs as well as the sub-clustering of reactions. The shortest and alternate paths in the reaction networks have also been examined. Sub-cluster profiling demonstrates that reactions of the mycolic acid pathway in mycobacteria form a tightly connected sub-cluster. Identification of hubs reveals reactions involving glutamate to be central to mycobacterial metabolism, and pyruvate to be at the centre of the E. coli metabolome. The analysis of shortest paths between reactions has revealed several paths that are shorter than well established pathways. Conclusions. We conclude that severe downsizing of the leprae genome has not significantly altered the global structure of its reaction network but has reduced the total number of alternate paths between its reactions while keeping the shortest paths between them intact. The hubs in the mycobacterial networks that are absent in the human metabolome can be explored as potential drug targets. This work demonstrates the usefulness of constructing metabolome based networks of organisms and the feasibility of their analyses through graph spectral methods. The insights obtained from such studies provide a broad overview of the similarities and differences between organisms, taking comparative genomics studies to a higher dimension.
Resumo:
Experiments have repeatedly observed both thermodynamic and dynamic anomalies in aqueous binary mixtures, surprisingly at low solute concentration. Examples of such binary mixtures include water-DMSO, water-ethanol, water-tertiary butyl alcohol (TBA), and water-dioxane, to name a few. The anomalies have often been attributed to the onset of a structural transition, whose nature, however, has been left rather unclear. Here we study the origin of such anomalies using large scale computer simulations and theoretical analysis in water-DMSO binary mixture. At very low DMSO concentration (below 10%), small aggregates of DMSO are solvated by water through the formation of DMSO-(H2O)(2) moieties. As the concentration is increased beyond 10-12% of DMSO, spanning clusters comprising the same moieties appear in the system. Those clusters are formed and stabilized not only through H-bonding but also through the association of CH3 groups of DMSO. We attribute the experimentally observed anomalies to a continuum percolation-like transition at DMSO concentration X-DMSO approximate to 12-15%. The largest cluster size of CH3-CH3 aggregation clearly indicates the formation of such percolating clusters. As a result, a significant slowing down is observed in the decay of associated rotational auto time correlation functions (of the S = O bond vector of DMSO and O-H bond vector of water). Markedly unusual behavior in the mean square fluctuation of total dipole moment again suggests a structural transition around the same concentration range. Furthermore, we map our findings to an interacting lattice model which substantiates the continuum percolation model as the reason for low concentration anomalies in binary mixtures where the solutes involved have both hydrophilic and hydrophobic moieties.
Resumo:
In this article, we present a novel application of a quantum clustering (QC) technique to objectively cluster the conformations, sampled by molecular dynamics simulations performed on different ligand bound structures of the protein. We further portray each conformational population in terms of dynamically stable network parameters which beautifully capture the ligand induced variations in the ensemble in atomistic detail. The conformational populations thus identified by the QC method and verified by network parameters are evaluated for different ligand bound states of the protein pyrrolysyl-tRNA synthetase (DhPylRS) from D. hafniense. The ligand/environment induced re-distribution of protein conformational ensembles forms the basis for understanding several important biological phenomena such as allostery and enzyme catalysis. The atomistic level characterization of each population in the conformational ensemble in terms of the re-orchestrated networks of amino acids is a challenging problem, especially when the changes are minimal at the backbone level. Here we demonstrate that the QC method is sensitive to such subtle changes and is able to cluster MD snapshots which are similar at the side-chain interaction level. Although we have applied these methods on simulation trajectories of a modest time scale (20 ns each), we emphasize that our methodology provides a general approach towards an objective clustering of large-scale MD simulation data and may be applied to probe multistate equilibria at higher time scales, and to problems related to protein folding for any protein or protein-protein/RNA/DNA complex of interest with a known structure.
Resumo:
As the beneficial effects of curcumin have often been reported to be limited to its small concentrations, we have undertaken a study to find the aggregation properties of curcumin in water by varying the number of monomers. Our molecular dynamics simulation results show that the equilibrated structure is always an aggregated state with remarkable structural rearrangements as we vary the number of curcumin monomers from 4 to 16 monomers. We find that the curcumin monomers form clusters in a very definite pattern where they tend to aggregate both in parallel and anti-parallel orientation of the phenyl rings, often seen in the formation of beta-sheet in proteins. A considerable enhancement in the population of parallel alignments is observed with increasing the system size from 12 to 16 curcumin monomers. Due to the prevalence of such parallel alignment for large system size, a more closely packed cluster is formed with maximum number of hydrophobic contacts. We also follow the pathway of cluster growth, in particular the transition from the initial segregated to the final aggregated state. We find the existence of a metastable structural intermediate involving a number of intermediate-sized clusters dispersed in the solution. We have constructed a free energy landscape of aggregation where the metatsable state has been identified. The course of aggregation bears similarity to nucleation and growth in highly metastable state. The final aggregated form remains stable with the total exclusion of water from its sequestered hydrophobic core. We also investigate water structure near the cluster surface along with their orientation. We find that water molecules form a distorted tetrahedral geometry in the 1st solvation layer of the cluster, interacting rather strongly with the hydrophilic groups at the surface of the curcumin. The dynamics of such quasi-bound water molecules near the surface of curcumin cluster is considerably slower than the bulk signifying a restricted motion as often found in protein hydration layer. (C) 2014 AIP Publishing LLC.
Resumo:
Identification of homogeneous hydrometeorological regions (HMRs) is necessary for various applications. Such regions are delineated by various approaches considering rainfall and temperature as two key variables. In conventional approaches, formation of regions is based on principal components (PCs)/statistics/indices determined from time series of the key variables at monthly and seasonal scales. An issue with use of PCs for regionalization is that they have to be extracted from contemporaneous records of hydrometeorological variables. Therefore, delineated regions may not be effective when the available records are limited over contemporaneous time period. A drawback associated with the use of statistics/indices is that they do not provide effective representation of the key variables when the records exhibit non-stationarity. Consequently, the resulting regions may not be effective for the desired purpose. To address these issues, a new approach is proposed in this article. The approach considers information extracted from wavelet transformations of the observed multivariate hydrometeorological time series as the basis for regionalization by global fuzzy c-means clustering procedure. The approach can account for dynamic variability in the time series and its non-stationarity (if any). Effectiveness of the proposed approach in forming HMRs is demonstrated by application to India, as there are no prior attempts to form such regions over the country. Drought severity-area-frequency (SAF) curves are constructed corresponding to each of the newly formed regions for the use in regional drought analysis, by considering standardized precipitation evapotranspiration index (SPEI) as the drought indicator.
Resumo:
We propose a highly efficient content-lossless compression scheme for Chinese document images. The scheme combines morphologic analysis with pattern matching to cluster patterns. In order to achieve the error maps with minimal error numbers, the morphologic analysis is applied to decomposing and recomposing the Chinese character patterns. In the pattern matching, the criteria are adapted to the characteristics of Chinese characters. Since small-size components sometimes can be inserted into the blank spaces of large-size components, we can achieve small-size pattern library images. Arithmetic coding is applied to the final compression. Our method achieves much better compression performance than most alternative methods, and assures content-lossless reconstruction. (c) 2006 Society of Photo-Optical Instrumentation Engineers.
Resumo:
We propose a highly efficient content-lossless compression scheme for Chinese document images. The scheme combines morphologic analysis with pattern matching to cluster patterns. In order to achieve the error maps with minimal error numbers, the morphologic analysis is applied to decomposing and recomposing the Chinese character patterns. In the pattern matching, the criteria are adapted to the characteristics of Chinese characters. Since small-size components sometimes can be inserted into the blank spaces of large-size components, we can achieve small-size pattern library images. Arithmetic coding is applied to the final compression. Our method achieves much better compression performance than most alternative methods, and assures content-lossless reconstruction. (c) 2006 Society of Photo-Optical Instrumentation Engineers.
Resumo:
STEEL, the Caltech created nonlinear large displacement analysis software, is currently used by a large number of researchers at Caltech. However, due to its complexity, lack of visualization tools (such as pre- and post-processing capabilities) rapid creation and analysis of models using this software was difficult. SteelConverter was created as a means to facilitate model creation through the use of the industry standard finite element solver ETABS. This software allows users to create models in ETABS and intelligently convert model information such as geometry, loading, releases, fixity, etc., into a format that STEEL understands. Models that would take several days to create and verify now take several hours or less. The productivity of the researcher as well as the level of confidence in the model being analyzed is greatly increased.
It has always been a major goal of Caltech to spread the knowledge created here to other universities. However, due to the complexity of STEEL it was difficult for researchers or engineers from other universities to conduct analyses. While SteelConverter did help researchers at Caltech improve their research, sending SteelConverter and its documentation to other universities was less than ideal. Issues of version control, individual computer requirements, and the difficulty of releasing updates made a more centralized solution preferred. This is where the idea for Caltech VirtualShaker was born. Through the creation of a centralized website where users could log in, submit, analyze, and process models in the cloud, all of the major concerns associated with the utilization of SteelConverter were eliminated. Caltech VirtualShaker allows users to create profiles where defaults associated with their most commonly run models are saved, and allows them to submit multiple jobs to an online virtual server to be analyzed and post-processed. The creation of this website not only allowed for more rapid distribution of this tool, but also created a means for engineers and researchers with no access to powerful computer clusters to run computationally intensive analyses without the excessive cost of building and maintaining a computer cluster.
In order to increase confidence in the use of STEEL as an analysis system, as well as verify the conversion tools, a series of comparisons were done between STEEL and ETABS. Six models of increasing complexity, ranging from a cantilever column to a twenty-story moment frame, were analyzed to determine the ability of STEEL to accurately calculate basic model properties such as elastic stiffness and damping through a free vibration analysis as well as more complex structural properties such as overall structural capacity through a pushover analysis. These analyses showed a very strong agreement between the two softwares on every aspect of each analysis. However, these analyses also showed the ability of the STEEL analysis algorithm to converge at significantly larger drifts than ETABS when using the more computationally expensive and structurally realistic fiber hinges. Following the ETABS analysis, it was decided to repeat the comparisons in a software more capable of conducting highly nonlinear analysis, called Perform. These analyses again showed a very strong agreement between the two softwares in every aspect of each analysis through instability. However, due to some limitations in Perform, free vibration analyses for the three story one bay chevron brace frame, two bay chevron brace frame, and twenty story moment frame could not be conducted. With the current trend towards ultimate capacity analysis, the ability to use fiber based models allows engineers to gain a better understanding of a building’s behavior under these extreme load scenarios.
Following this, a final study was done on Hall’s U20 structure [1] where the structure was analyzed in all three softwares and their results compared. The pushover curves from each software were compared and the differences caused by variations in software implementation explained. From this, conclusions can be drawn on the effectiveness of each analysis tool when attempting to analyze structures through the point of geometric instability. The analyses show that while ETABS was capable of accurately determining the elastic stiffness of the model, following the onset of inelastic behavior the analysis tool failed to converge. However, for the small number of time steps the ETABS analysis was converging, its results exactly matched those of STEEL, leading to the conclusion that ETABS is not an appropriate analysis package for analyzing a structure through the point of collapse when using fiber elements throughout the model. The analyses also showed that while Perform was capable of calculating the response of the structure accurately, restrictions in the material model resulted in a pushover curve that did not match that of STEEL exactly, particularly post collapse. However, such problems could be alleviated by choosing a more simplistic material model.
Resumo:
Early glasses (about 1066 BC-220 AD) unearthed from Xinjiang of China were chemically characterized by using PIXE and ICP-AES. It was found that these glasses were basically attributed to PbO-BaO-SiO2 system, K2O-SiO2 system, Na2O-CaO-SiO2 system and Na2O-CaO-PbO-SiO2 system. The results from the cluster analysis showed that some glasses had basically similar recipe and technology. The PbO-BaO-SiO2 glass and the K2O-SiO2 glass were thought to come from the central area and the south of ancient China, respectively. The part of the Na2O-CaO-SiO2 glass (including the Na2O-CaO-PbO-SiO2 glass) might be imported from Mesopotamia, while the other part might be locally produced. (c) 2005 Elsevier B.V. All rights reserved.
Resumo:
Background: In complex with its cofactor UAF1, the USP1 deubiquitinase plays an important role in cellular processes related to cancer, including the response to DNA damage. The USP1/UAF1 complex is emerging as a novel target in cancer therapy, but several aspects of its function and regulation remain to be further clarified. These include the role of the serine 313 phosphorylation site, the relative contribution of different USP1 sequence motifs to UAF1 binding, and the potential effect of cancer-associated mutations on USP1 regulation by autocleavage. Methods: We have generated a large set of USP1 structural variants, including a catalytically inactive form (C90S), non-phosphorylatable (S313A) and phosphomimetic (S313D) mutants, deletion mutants lacking potential UAF1 binding sites, a mutant (GG/AA) unable to undergo autocleavage at the well-characterized G670/G671 diglycine motif, and four USP1 mutants identified in tumor samples that cluster around this cleavage site (G667A, L669P, K673T and A676T). Using cell-based assays, we have determined the ability of these mutants to bind UAF1, to reverse DNA damage-induced monoubiquitination of PCNA, and to undergo autocleavage. Results: A non-phosphorylatable S313A mutant of USP1 retained the ability to bind UAF1 and to reverse PCNA ubiquitination in cell-based assays. Regardless of the presence of a phosphomimetic S313D mutation, deletion of USP1 fragment 420-520 disrupted UAF1 binding, as determined using a nuclear relocation assay. The UAF1 binding site in a second UAF1-interacting DUB, USP46, was mapped to a region homologous to USP1(420-520). Regarding USP1 autocleavage, co-expression of the C90S and GG/AA mutants did not result in cleavage, while the cancer-associated mutation L669P was found to reduce cleavage efficiency. Conclusions: USP1 phosphorylation at S313 is not critical for PCNA deubiquitination, neither for binding to UAF1 in a cellular environment. In this context, USP1 amino acid motif 420-520 is necessary and sufficient for UAF1 binding. This motif, and a homologous amino acid segment that mediates USP46 binding to UAF1, map to the Fingers sub-domain of these DUBs. On the other hand, our results support the view that USP1 autocleavage may occur in cis, and can be altered by a cancer-associated mutation.
Resumo:
Cluster analysis of ranking data, which occurs in consumer questionnaires, voting forms or other inquiries of preferences, attempts to identify typical groups of rank choices. Empirically measured rankings are often incomplete, i.e. different numbers of filled rank positions cause heterogeneity in the data. We propose a mixture approach for clustering of heterogeneous rank data. Rankings of different lengths can be described and compared by means of a single probabilistic model. A maximum entropy approach avoids hidden assumptions about missing rank positions. Parameter estimators and an efficient EM algorithm for unsupervised inference are derived for the ranking mixture model. Experiments on both synthetic data and real-world data demonstrate significantly improved parameter estimates on heterogeneous data when the incomplete rankings are included in the inference process.