104 resultados para Cluster Counting Algorithm
Resumo:
Colour-magnitude diagrams (CMDs) of the Small Magellanic Cloud (SMC) star cluster NGC 419, derived from Hubble Space Telescope (HST)/Advanced Camera for Surveys (ACS) data, reveal a well-delineated secondary clump located below the classical compact red clump typical of intermediate-age populations. We demonstrate that this feature belongs to the cluster itself, rather than to the underlying SMC field. Then, we use synthetic CMDs to show that it corresponds very well to the secondary clump predicted to appear as a result of He-ignition in stars just massive enough to avoid e(-)-degeneracy settling in their H-exhausted cores. The main red clump instead is made of the slightly less massive stars which passed through e(-) degeneracy and ignited He at the tip of the red giant branch. In other words, NGC 419 is the rare snapshot of a cluster while undergoing the fast transition from classical to degenerate H-exhausted cores. At this particular moment of a cluster`s life, the colour distance between the main-sequence turn-off and the red clump(s) depends sensitively on the amount of convective core overshooting, Lambda(c). By coupling measurements of this colour separation with fits to the red clump morphology, we are able to estimate simultaneously the cluster mean age (1.35(-0.04)(+0.11) Gyr) and overshooting efficiency (Lambda(c) = 0.47(-0.04)(+0.14)). Therefore, clusters like NGC 419 may constitute important marks in the age scale of intermediate-age populations. After eye inspection of other CMDs derived from HST/ACS data, we suggest that the same secondary clump may also be present in the Large Magellanic Cloud clusters NGC 1751, 1783, 1806, 1846, 1852 and 1917.
Resumo:
The bees of the Peponapes genus (Eucerini, Apidae) have a Neotropical distribution with the center of species diversity located in Mexico and are specialized in Cucurbita plants. which have many species of economic importance. such as squashes and pumpkins Peponapis fervens is the only species of the genus known from southern South America The Cucurbita species occurring in the same area as P fervens Include four domesticated species (C ficifolia, C maxima maxima, C moschata and C pepo) and one non-domesticated species (Cucurbita maxima andreana) It was suggested that C. in andreana was the original pollen source to P fervens, and this bee expanded its geographical range due to the domestication of Cucurbita The potential geographical areas of these species were determined and compared using ecological niche modeling that was performed with the computational system openModeller and GARP with best subsets algorithm The climatic variables obtained through modeling were compared using Cluster Analysis Results show that the potential areas of domesticated species practically spread all over South America The potential area of P fervens Includes the areas of C m andreana but reaches a larger area, where the domesticated species of Cucurbita also Occur The Cluster Analysis shows a high climatic similarity between P fervens and C. m. andreana Nevertheless. P fervens presents the ability to occupy areas with wider ranges of climatic variables and to exploit resources provided by domesticated species (C) 2009 Elsevier B V All rights reserved
Resumo:
Background and Objectives: There are some indications that low-level laser therapy (LLLT) may delay the development of skeletal muscle fatigue during high-intensity exercise. There have also been claims that LED cluster probes may be effective for this application however there are differences between LED and laser sources like spot size, spectral width, power output, etc. In this study we wanted to test if light emitting diode therapy (LEDT) can alter muscle performance, fatigue development and biochemical markers for skeletal muscle recovery in an experimental model of biceps humeri muscle contractions. Study Design/Materials and Methods: Ten male professional volleyball players (23.6 [SD +/- 5.6] years old) entered a randomized double-blinded placebo-controlled crossover trial. Active cluster LEDT (69 LEDs with wavelengths 660/850 nm, 10/30 mW, 30 seconds total irradiation time, 41.7J of total energy irradiated) or an identical placebo LEDT was delivered under double-blinded conditions to the middle of biceps humeri muscle immediately before exercise. All subjects performed voluntary biceps humeri contractions with a workload of 75% of their maximal voluntary contraction force (MVC) until exhaustion. Results: Active LEDT increased the number of biceps humeri contractions by 12.9% (38.60 [SD +/- 9.03] vs. 34.20 [SD +/- 8.68], P = 0.021) and extended the elapsed time to perform contractions by 11.6% (P = 0.036) versus placebo. In addition, post-exercise levels of biochemical markers decreased significantly with active LEDT: Blood Lactate (P = 0.042), Creatine Kinase (P = 0.035), and C-Reative Protein levels (P = 0.030), when compared to placebo LEDT. Conclusion: We conclude that this particular procedure and dose of LEDT immediately before exhaustive biceps humeri contractions, causes a slight delay in the development of skeletal muscle fatigue, decreases post-exercise blood lactate levels and inhibits the release of Creatine Kinase and C-Reative Protein. Lasers Surg. Med. 41:572-577, 2009. (C) 2009 Wiley-Liss, Inc.
Resumo:
One of the top ten most influential data mining algorithms, k-means, is known for being simple and scalable. However, it is sensitive to initialization of prototypes and requires that the number of clusters be specified in advance. This paper shows that evolutionary techniques conceived to guide the application of k-means can be more computationally efficient than systematic (i.e., repetitive) approaches that try to get around the above-mentioned drawbacks by repeatedly running the algorithm from different configurations for the number of clusters and initial positions of prototypes. To do so, a modified version of a (k-means based) fast evolutionary algorithm for clustering is employed. Theoretical complexity analyses for the systematic and evolutionary algorithms under interest are provided. Computational experiments and statistical analyses of the results are presented for artificial and text mining data sets. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
A large amount of biological data has been produced in the last years. Important knowledge can be extracted from these data by the use of data analysis techniques. Clustering plays an important role in data analysis, by organizing similar objects from a dataset into meaningful groups. Several clustering algorithms have been proposed in the literature. However, each algorithm has its bias, being more adequate for particular datasets. This paper presents a mathematical formulation to support the creation of consistent clusters for biological data. Moreover. it shows a clustering algorithm to solve this formulation that uses GRASP (Greedy Randomized Adaptive Search Procedure). We compared the proposed algorithm with three known other algorithms. The proposed algorithm presented the best clustering results confirmed statistically. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
The evolution of commodity computing lead to the possibility of efficient usage of interconnected machines to solve computationally-intensive tasks, which were previously solvable only by using expensive supercomputers. This, however, required new methods for process scheduling and distribution, considering the network latency, communication cost, heterogeneous environments and distributed computing constraints. An efficient distribution of processes over such environments requires an adequate scheduling strategy, as the cost of inefficient process allocation is unacceptably high. Therefore, a knowledge and prediction of application behavior is essential to perform effective scheduling. In this paper, we overview the evolution of scheduling approaches, focusing on distributed environments. We also evaluate the current approaches for process behavior extraction and prediction, aiming at selecting an adequate technique for online prediction of application execution. Based on this evaluation, we propose a novel model for application behavior prediction, considering chaotic properties of such behavior and the automatic detection of critical execution points. The proposed model is applied and evaluated for process scheduling in cluster and grid computing environments. The obtained results demonstrate that prediction of the process behavior is essential for efficient scheduling in large-scale and heterogeneous distributed environments, outperforming conventional scheduling policies by a factor of 10, and even more in some cases. Furthermore, the proposed approach proves to be efficient for online predictions due to its low computational cost and good precision. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
In this paper, we present different ofrailtyo models to analyze longitudinal data in the presence of covariates. These models incorporate the extra-Poisson variability and the possible correlation among the repeated counting data for each individual. Assuming a CD4 counting data set in HIV-infected patients, we develop a hierarchical Bayesian analysis considering the different proposed models and using Markov Chain Monte Carlo methods. We also discuss some Bayesian discrimination aspects for the choice of the best model.
Resumo:
In this paper, we present an algorithm for cluster analysis that integrates aspects from cluster ensemble and multi-objective clustering. The algorithm is based on a Pareto-based multi-objective genetic algorithm, with a special crossover operator, which uses clustering validation measures as objective functions. The algorithm proposed can deal with data sets presenting different types of clusters, without the need of expertise in cluster analysis. its result is a concise set of partitions representing alternative trade-offs among the objective functions. We compare the results obtained with our algorithm, in the context of gene expression data sets, to those achieved with multi-objective Clustering with automatic K-determination (MOCK). the algorithm most closely related to ours. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
This paper tackles the problem of showing that evolutionary algorithms for fuzzy clustering can be more efficient than systematic (i.e. repetitive) approaches when the number of clusters in a data set is unknown. To do so, a fuzzy version of an Evolutionary Algorithm for Clustering (EAC) is introduced. A fuzzy cluster validity criterion and a fuzzy local search algorithm are used instead of their hard counterparts employed by EAC. Theoretical complexity analyses for both the systematic and evolutionary algorithms under interest are provided. Examples with computational experiments and statistical analyses are also presented.
Resumo:
In this paper we present a genetic algorithm with new components to tackle capacitated lot sizing and scheduling problems with sequence dependent setups that appear in a wide range of industries, from soft drink bottling to food manufacturing. Finding a feasible solution to highly constrained problems is often a very difficult task. Various strategies have been applied to deal with infeasible solutions throughout the search. We propose a new scheme of classifying individuals based on nested domains to determine the solutions according to the level of infeasibility, which in our case represents bands of additional production hours (overtime). Within each band, individuals are just differentiated by their fitness function. As iterations are conducted, the widths of the bands are dynamically adjusted to improve the convergence of the individuals into the feasible domain. The numerical experiments on highly capacitated instances show the effectiveness of this computational tractable approach to guide the search toward the feasible domain. Our approach outperforms other state-of-the-art approaches and commercial solvers. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
A numerical algorithm for fully dynamical lubrication problems based on the Elrod-Adams formulation of the Reynolds equation with mass-conserving boundary conditions is described. A simple but effective relaxation scheme is used to update the solution maintaining the complementarity conditions on the variables that represent the pressure and fluid fraction. The equations of motion are discretized in time using Newmark`s scheme, and the dynamical variables are updated within the same relaxation process just mentioned. The good behavior of the proposed algorithm is illustrated in two examples: an oscillatory squeeze flow (for which the exact solution is available) and a dynamically loaded journal bearing. This article is accompanied by the ready-to-compile source code with the implementation of the proposed algorithm. [DOI: 10.1115/1.3142903]
Resumo:
The amount of textual information digitally stored is growing every day. However, our capability of processing and analyzing that information is not growing at the same pace. To overcome this limitation, it is important to develop semiautomatic processes to extract relevant knowledge from textual information, such as the text mining process. One of the main and most expensive stages of the text mining process is the text pre-processing stage, where the unstructured text should be transformed to structured format such as an attribute-value table. The stemming process, i.e. linguistics normalization, is usually used to find the attributes of this table. However, the stemming process is strongly dependent on the language in which the original textual information is given. Furthermore, for most languages, the stemming algorithms proposed in the literature are computationally expensive. In this work, several improvements of the well know Porter stemming algorithm for the Portuguese language, which explore the characteristics of this language, are proposed. Experimental results show that the proposed algorithm executes in far less time without affecting the quality of the generated stems.
Resumo:
Conventional procedures employed in the modeling of viscoelastic properties of polymer rely on the determination of the polymer`s discrete relaxation spectrum from experimentally obtained data. In the past decades, several analytical regression techniques have been proposed to determine an explicit equation which describes the measured spectra. With a diverse approach, the procedure herein introduced constitutes a simulation-based computational optimization technique based on non-deterministic search method arisen from the field of evolutionary computation. Instead of comparing numerical results, this purpose of this paper is to highlight some Subtle differences between both strategies and focus on what properties of the exploited technique emerge as new possibilities for the field, In oder to illustrate this, essayed cases show how the employed technique can outperform conventional approaches in terms of fitting quality. Moreover, in some instances, it produces equivalent results With much fewer fitting parameters, which is convenient for computational simulation applications. I-lie problem formulation and the rationale of the highlighted method are herein discussed and constitute the main intended contribution. (C) 2009 Wiley Periodicals, Inc. J Appl Polym Sci 113: 122-135, 2009
Resumo:
In 2006 the Route load balancing algorithm was proposed and compared to other techniques aiming at optimizing the process allocation in grid environments. This algorithm schedules tasks of parallel applications considering computer neighborhoods (where the distance is defined by the network latency). Route presents good results for large environments, although there are cases where neighbors do not have an enough computational capacity nor communication system capable of serving the application. In those situations the Route migrates tasks until they stabilize in a grid area with enough resources. This migration may take long time what reduces the overall performance. In order to improve such stabilization time, this paper proposes RouteGA (Route with Genetic Algorithm support) which considers historical information on parallel application behavior and also the computer capacities and load to optimize the scheduling. This information is extracted by using monitors and summarized in a knowledge base used to quantify the occupation of tasks. Afterwards, such information is used to parameterize a genetic algorithm responsible for optimizing the task allocation. Results confirm that RouteGA outperforms the load balancing carried out by the original Route, which had previously outperformed others scheduling algorithms from literature.
Resumo:
Purpose - The purpose of this paper is to develop a novel unstructured simulation approach for injection molding processes described by the Hele-Shaw model. Design/methodology/approach - The scheme involves dual dynamic meshes with active and inactive cells determined from an initial background pointset. The quasi-static pressure solution in each timestep for this evolving unstructured mesh system is approximated using a control volume finite element method formulation coupled to a corresponding modified volume of fluid method. The flow is considered to be isothermal and non-Newtonian. Findings - Supporting numerical tests and performance studies for polystyrene described by Carreau, Cross, Ellis and Power-law fluid models are conducted. Results for the present method are shown to be comparable to those from other methods for both Newtonian fluid and polystyrene fluid injected in different mold geometries. Research limitations/implications - With respect to the methodology, the background pointset infers a mesh that is dynamically reconstructed here, and there are a number of efficiency issues and improvements that would be relevant to industrial applications. For instance, one can use the pointset to construct special bases and invoke a so-called ""meshless"" scheme using the basis. This would require some interesting strategies to deal with the dynamic point enrichment of the moving front that could benefit from the present front treatment strategy. There are also issues related to mass conservation and fill-time errors that might be addressed by introducing suitable projections. The general question of ""rate of convergence"" of these schemes requires analysis. Numerical results here suggest first-order accuracy and are consistent with the approximations made, but theoretical results are not available yet for these methods. Originality/value - This novel unstructured simulation approach involves dual meshes with active and inactive cells determined from an initial background pointset: local active dual patches are constructed ""on-the-fly"" for each ""active point"" to form a dynamic virtual mesh of active elements that evolves with the moving interface.