6 resultados para many-objective problems
em Repositorio Institucional de la Universidad de Málaga
Resumo:
Phylogenetic inference consist in the search of an evolutionary tree to explain the best way possible genealogical relationships of a set of species. Phylogenetic analysis has a large number of applications in areas such as biology, ecology, paleontology, etc. There are several criterias which has been defined in order to infer phylogenies, among which are the maximum parsimony and maximum likelihood. The first one tries to find the phylogenetic tree that minimizes the number of evolutionary steps needed to describe the evolutionary history among species, while the second tries to find the tree that has the highest probability of produce the observed data according to an evolutionary model. The search of a phylogenetic tree can be formulated as a multi-objective optimization problem, which aims to find trees which satisfy simultaneously (and as much as possible) both criteria of parsimony and likelihood. Due to the fact that these criteria are different there won't be a single optimal solution (a single tree), but a set of compromise solutions. The solutions of this set are called "Pareto Optimal". To find this solutions, evolutionary algorithms are being used with success nowadays.This algorithms are a family of techniques, which aren’t exact, inspired by the process of natural selection. They usually find great quality solutions in order to resolve convoluted optimization problems. The way this algorithms works is based on the handling of a set of trial solutions (trees in the phylogeny case) using operators, some of them exchanges information between solutions, simulating DNA crossing, and others apply aleatory modifications, simulating a mutation. The result of this algorithms is an approximation to the set of the “Pareto Optimal” which can be shown in a graph with in order that the expert in the problem (the biologist when we talk about inference) can choose the solution of the commitment which produces the higher interest. In the case of optimization multi-objective applied to phylogenetic inference, there is open source software tool, called MO-Phylogenetics, which is designed for the purpose of resolving inference problems with classic evolutionary algorithms and last generation algorithms. REFERENCES [1] C.A. Coello Coello, G.B. Lamont, D.A. van Veldhuizen. Evolutionary algorithms for solving multi-objective problems. Spring. Agosto 2007 [2] C. Zambrano-Vega, A.J. Nebro, J.F Aldana-Montes. MO-Phylogenetics: a phylogenetic inference software tool with multi-objective evolutionary metaheuristics. Methods in Ecology and Evolution. En prensa. Febrero 2016.
Resumo:
Ligand-protein docking is an optimization problem based on predicting the position of a ligand with the lowest binding energy in the active site of the receptor. Molecular docking problems are traditionally tackled with single-objective, as well as with multi-objective approaches, to minimize the binding energy. In this paper, we propose a novel multi-objective formulation that considers: the Root Mean Square Deviation (RMSD) difference in the coordinates of ligands and the binding (intermolecular) energy, as two objectives to evaluate the quality of the ligand-protein interactions. To determine the kind of Pareto front approximations that can be obtained, we have selected a set of representative multi-objective algorithms such as NSGA-II, SMPSO, GDE3, and MOEA/D. Their performances have been assessed by applying two main quality indicators intended to measure convergence and diversity of the fronts. In addition, a comparison with LGA, a reference single-objective evolutionary algorithm for molecular docking (AutoDock) is carried out. In general, SMPSO shows the best overall results in terms of energy and RMSD (value lower than 2A for successful docking results). This new multi-objective approach shows an improvement over the ligand-protein docking predictions that could be promising in in silico docking studies to select new anticancer compounds for therapeutic targets that are multidrug resistant.
Resumo:
Background: Complex chronic diseases are a challenge for the current configuration of Health services. Case management is a service frequently provided for people with chronic conditions and despite its effectiveness in many outcomes, such as mortality or readmissions, uncertainty remains about the most effective form of team organization, structures, and the nature of the interventions. Many processes and outcomes of case management for people with complex chronic conditions cannot be addressed with the information provided by electronic clinical records. Registries are frequently used to deal with this weakness. The aim of this study was to generate a registry-based information system of patients receiving case management to identify their clinical characteristics, their context of care, events identified during their follow-up, interventions developed by case managers, and services used. Methods and design: The study was divided into three phases, covering the detection of information needs, the design and its implementation in the healthcare system, using literature review and expert consensus methods to select variables that would be included in the registry. Objective: To describe the essential characteristics of the provision of ca re lo people who receive case management (structure, process and outcomes), with special emphasis on those with complex chronic diseases. Study population: Patients from any District of Primary Care, who initiate the utilization of case management services, to avoid information bias that may occur when including subjects who have already been received the service, and whose outcomes and characteristics could not be properly collected. Results: A total of 102 variables representing structure, processes and outcomes of case management were selected for their inclusion in the registry after the consensus phase. Total sample was composed of 427 patients, of which 211 (49.4%) were women and 216 (50.6%) were men. The average functional level (Barthel lndex) was 36.18 (SD 29.02), cognitive function (Pfeiffer) showed an average of 4.37 {SD 6.57), Chat1son Comorbidity lndex, obtained a mean of 3.03 (SD 2.7) and Social Support (Duke lndex) was 34.2 % (SD 17.57). More than half of patients include in the Registry, correspond lo immobilized or transitional care for patients discharged from hospital (66.5 %). The patient's educational level was low or very low (50.4%). Caregivers overstrain (Caregiver stress index), obtained an average value of 6.09% (SD 3.53). Only 1.2 % of patients had declared their advanced directives, 58.6 had not defined the tutelage and the vast majority lived at home 98.8 %. Regarding the major events recorded at RANGE Registry, 25.8 % of the selected patients died in the first three months, 8.2 % suffered a hospital admission at least once time, 2.3%, two times, and 1.2% three times, 7.5% suffered a fall, 8.7% had pressure ulcer, 4.7% had problems with medication, and 3.3 % were institutionalized. Stroke is the more prevalent health problem recorded (25.1%), followed by hypertension (11.1%) and COPD (11.1%). Patients registered by NCMs had as main processes diabetes (16.8%) and dementia (11.3 %). The most frequent nursing diagnoses referred to the self-care deficit in various activities of daily living. Regarding to nursing interventions, described by the Nursing Intervention Classification (NIC), dementia management is the most used intervention, followed by mutual goal setting, caregiver and emotional support. Conclusions: The patient profile who receive case management services is a chronic complex patient with severe dependence, cognitive impairment, normal social support, low educational level, health problems such as stroke, hypertension or COPD, diabetes or dementia, and has an informal caregiver. At the first follow up, mortality was 19.2%, and a discrete rate of readmissions and falls.
Resumo:
Efficient hill climbers have been recently proposed for single- and multi-objective pseudo-Boolean optimization problems. For $k$-bounded pseudo-Boolean functions where each variable appears in at most a constant number of subfunctions, it has been theoretically proven that the neighborhood of a solution can be explored in constant time. These hill climbers, combined with a high-level exploration strategy, have shown to improve state of the art methods in experimental studies and open the door to the so-called Gray Box Optimization, where part, but not all, of the details of the objective functions are used to better explore the search space. One important limitation of all the previous proposals is that they can only be applied to unconstrained pseudo-Boolean optimization problems. In this work, we address the constrained case for multi-objective $k$-bounded pseudo-Boolean optimization problems. We find that adding constraints to the pseudo-Boolean problem has a linear computational cost in the hill climber.
Resumo:
Technologies for Big Data and Data Science are receiving increasing research interest nowadays. This paper introduces the prototyping architecture of a tool aimed to solve Big Data Optimization problems. Our tool combines the jMetal framework for multi-objective optimization with Apache Spark, a technology that is gaining momentum. In particular, we make use of the streaming facilities of Spark to feed an optimization problem with data from different sources. We demonstrate the use of our tool by solving a dynamic bi-objective instance of the Traveling Salesman Problem (TSP) based on near real-time traffic data from New York City, which is updated several times per minute. Our experiment shows that both jMetal and Spark can be integrated providing a software platform to deal with dynamic multi-optimization problems.
Resumo:
Sequence problems belong to the most challenging interdisciplinary topics of the actuality. They are ubiquitous in science and daily life and occur, for example, in form of DNA sequences encoding all information of an organism, as a text (natural or formal) or in form of a computer program. Therefore, sequence problems occur in many variations in computational biology (drug development), coding theory, data compression, quantitative and computational linguistics (e.g. machine translation). In recent years appeared some proposals to formulate sequence problems like the closest string problem (CSP) and the farthest string problem (FSP) as an Integer Linear Programming Problem (ILPP). In the present talk we present a general novel approach to reduce the size of the ILPP by grouping isomorphous columns of the string matrix together. The approach is of practical use, since the solution of sequence problems is very time consuming, in particular when the sequences are long.