986 resultados para Ordered Gene Problems
Resumo:
Ordered gene problems are a very common classification of optimization problems. Because of their popularity countless algorithms have been developed in an attempt to find high quality solutions to the problems. It is also common to see many different types of problems reduced to ordered gene style problems as there are many popular heuristics and metaheuristics for them due to their popularity. Multiple ordered gene problems are studied, namely, the travelling salesman problem, bin packing problem, and graph colouring problem. In addition, two bioinformatics problems not traditionally seen as ordered gene problems are studied: DNA error correction and DNA fragment assembly. These problems are studied with multiple variations and combinations of heuristics and metaheuristics with two distinct types or representations. The majority of the algorithms are built around the Recentering- Restarting Genetic Algorithm. The algorithm variations were successful on all problems studied, and particularly for the two bioinformatics problems. For DNA Error Correction multiple cases were found with 100% of the codes being corrected. The algorithm variations were also able to beat all other state-of-the-art DNA Fragment Assemblers on 13 out of 16 benchmark problem instances.
Resumo:
The remarkable advances made in recombinant DNA technology over the last two decades have paved way for the use of gene transfer to treat human diseases. Several protocols have been developed for the introduction and expression of genes in humans, but the clinical efficacy has not been conclusively demonstrated in any of them. The eventual success of gene therapy for genetic and acquired disorders depends on the development of better gene transfer vectors for sustained, long term expression of foreign genes as well as a better understanding of the pathophysiology of human diseases, it is heartening to note that some of the gene therapy protocols have found other applications such as the genetic immunization or DNA vaccines, which is being heralded as the third vaccine revolution, Gene therapy is yet to become a dream come true, but the light is seen at the end of the tunnel.
Resumo:
In the past decade, the advent of efficient genome sequencing tools and high-throughput experimental biotechnology has lead to enormous progress in the life science. Among the most important innovations is the microarray tecnology. It allows to quantify the expression for thousands of genes simultaneously by measurin the hybridization from a tissue of interest to probes on a small glass or plastic slide. The characteristics of these data include a fair amount of random noise, a predictor dimension in the thousand, and a sample noise in the dozens. One of the most exciting areas to which microarray technology has been applied is the challenge of deciphering complex disease such as cancer. In these studies, samples are taken from two or more groups of individuals with heterogeneous phenotypes, pathologies, or clinical outcomes. these samples are hybridized to microarrays in an effort to find a small number of genes which are strongly correlated with the group of individuals. Eventhough today methods to analyse the data are welle developed and close to reach a standard organization (through the effort of preposed International project like Microarray Gene Expression Data -MGED- Society [1]) it is not unfrequant to stumble in a clinician's question that do not have a compelling statistical method that could permit to answer it.The contribution of this dissertation in deciphering disease regards the development of new approaches aiming at handle open problems posed by clinicians in handle specific experimental designs. In Chapter 1 starting from a biological necessary introduction, we revise the microarray tecnologies and all the important steps that involve an experiment from the production of the array, to the quality controls ending with preprocessing steps that will be used into the data analysis in the rest of the dissertation. While in Chapter 2 a critical review of standard analysis methods are provided stressing most of problems that In Chapter 3 is introduced a method to adress the issue of unbalanced design of miacroarray experiments. In microarray experiments, experimental design is a crucial starting-point for obtaining reasonable results. In a two-class problem, an equal or similar number of samples it should be collected between the two classes. However in some cases, e.g. rare pathologies, the approach to be taken is less evident. We propose to address this issue by applying a modified version of SAM [2]. MultiSAM consists in a reiterated application of a SAM analysis, comparing the less populated class (LPC) with 1,000 random samplings of the same size from the more populated class (MPC) A list of the differentially expressed genes is generated for each SAM application. After 1,000 reiterations, each single probe given a "score" ranging from 0 to 1,000 based on its recurrence in the 1,000 lists as differentially expressed. The performance of MultiSAM was compared to the performance of SAM and LIMMA [3] over two simulated data sets via beta and exponential distribution. The results of all three algorithms over low- noise data sets seems acceptable However, on a real unbalanced two-channel data set reagardin Chronic Lymphocitic Leukemia, LIMMA finds no significant probe, SAM finds 23 significantly changed probes but cannot separate the two classes, while MultiSAM finds 122 probes with score >300 and separates the data into two clusters by hierarchical clustering. We also report extra-assay validation in terms of differentially expressed genes Although standard algorithms perform well over low-noise simulated data sets, multi-SAM seems to be the only one able to reveal subtle differences in gene expression profiles on real unbalanced data. In Chapter 4 a method to adress similarities evaluation in a three-class prblem by means of Relevance Vector Machine [4] is described. In fact, looking at microarray data in a prognostic and diagnostic clinical framework, not only differences could have a crucial role. In some cases similarities can give useful and, sometimes even more, important information. The goal, given three classes, could be to establish, with a certain level of confidence, if the third one is similar to the first or the second one. In this work we show that Relevance Vector Machine (RVM) [2] could be a possible solutions to the limitation of standard supervised classification. In fact, RVM offers many advantages compared, for example, with his well-known precursor (Support Vector Machine - SVM [3]). Among these advantages, the estimate of posterior probability of class membership represents a key feature to address the similarity issue. This is a highly important, but often overlooked, option of any practical pattern recognition system. We focused on Tumor-Grade-three-class problem, so we have 67 samples of grade I (G1), 54 samples of grade 3 (G3) and 100 samples of grade 2 (G2). The goal is to find a model able to separate G1 from G3, then evaluate the third class G2 as test-set to obtain the probability for samples of G2 to be member of class G1 or class G3. The analysis showed that breast cancer samples of grade II have a molecular profile more similar to breast cancer samples of grade I. Looking at the literature this result have been guessed, but no measure of significance was gived before.
Resumo:
This paper aims to show, analyze and solve the problems related to the translation of the book with meaning-bond alphabetically ordered chapter titles La vita non è in ordine alfabetico, by the Italian writer Andrea Bajani. The procedure is inevitable for a possible translation of the book, and it is necessary to have a preliminary pattern to follow. After translating the whole book, not only is a revision fundamental, but a restructure and reorganization of the collection may be required, hence, what this thesis offers is a scheme to start from, together with an analysis of the possible problems that may arise, and a useful method to find a solution.
Resumo:
It is commonly perceived that variables ‘measuring’ different dimensions of teaching (construed as instructional attributes) used in student evaluation of teaching (SET) questionnaires are so highly correlated that they pose a serious multicollinearity problem for quantitative analysis including regression analysis. Using nearly 12000 individual student responses to SET questionnaires and ten key dimensions of teaching and 25 courses at various undergraduate and postgraduate levels for multiple years at a large Australian university, this paper investigates whether this is indeed the case and if so under what circumstances. This paper tests this proposition first by examining variance inflation factors (VIFs), across courses, levels and over time using individual responses; and secondly by using class averages. In the first instance, the paper finds no sustainable evidence of multicollinearity. While, there were one or two isolated cases of VIFs marginally exceeding the conservative threshold of 5, in no cases did the VIFs for any of the instructional attributes come anywhere close to the high threshold value of 10. In the second instance, however, the paper finds that the attributes are highly correlated as all the VIFs exceed 10. These findings have two implications: (a) given the ordinal nature of the data ordered probit analysis using individual student responses can be employed to quantify the impact of instructional attributes on TEVAL score; (b) Data based on class averages cannot be used for probit analysis. An illustrative exercise using level 2 undergraduate courses data suggests higher TEVAL scores depend first and foremost on improving explanation, presentation, and organization of lecture materials.
Resumo:
Vascular endothelial growth factor (VEGF) is an endothelial cell-specific angiogenic protein suspected to be involved in the pathogenesis of endometriosis by establishing a new blood supply to the human exfoliated endometrium. Several transcription factor-binding sites are found in the VEGF 5'-untranslated region and variation within the region increases the transcriptional activity. Six previous studies which tested between one and three single nucleotide polymorphisms (SNPs) in samples comprising 105-215 cases and 100-219 controls have produced conflicting evidence for association between the SNPs in the VEGF region and endometriosis. To further investigate the reported association between VEGF variants and endometriosis, we tested the four VEGF polymorphisms (-2578 A/C, rs699947; -460 T/C, rs833061; +405 G/C, rs2010963 and +936 C/T, rs3025039) in a large Australian sample of 958 familial endometriosis cases and 959 controls. We also conducted a literature-based review of all relevant association studies of these VEGF SNPs in endometriosis and performed a meta-analysis. There was no evidence for association between endometriosis and the VEGF polymorphisms genotyped in our study. Combined association results from a meta-analysis did not provide any evidence for either genotypic or allelic association with endometriosis. Our detailed review and meta-analysis of the VEGF polymorphisms suggests that genotyping assay problems may underlie the previously reported associations between VEGF variants and endometriosis.
Resumo:
A new language concept for high-level distributed programming is proposed. Programs are organised as a collection of concurrently executing processes. Some of these processes, referred to as liaison processes, have a monitor-like structure and contain ports which may be invoked by other processes for the purposes of synchronisation and communication. Synchronisation is achieved by conditional activation of ports and also through port control constructs which may directly specify the execution ordering of ports. These constructs implement a path-expression-like mechanism for synchronisation and are also equipped with options to provide conditional, non-deterministic and priority ordering of ports. The usefulness and expressive power of the proposed concepts are illustrated through solutions of several representative programming problems. Some implementation issues are also considered.
Resumo:
In this paper, a novel genetic algorithm is developed by generating artificial chromosomes with probability control to solve the machine scheduling problems. Generating artificial chromosomes for Genetic Algorithm (ACGA) is closely related to Evolutionary Algorithms Based on Probabilistic Models (EAPM). The artificial chromosomes are generated by a probability model that extracts the gene information from current population. ACGA is considered as a hybrid algorithm because both the conventional genetic operators and a probability model are integrated. The ACGA proposed in this paper, further employs the ``evaporation concept'' applied in Ant Colony Optimization (ACO) to solve the permutation flowshop problem. The ``evaporation concept'' is used to reduce the effect of past experience and to explore new alternative solutions. In this paper, we propose three different methods for the probability of evaporation. This probability of evaporation is applied as soon as a job is assigned to a position in the permutation flowshop problem. Experimental results show that our ACGA with the evaporation concept gives better performance than some algorithms in the literature.
Resumo:
Synthetic biology combines biological parts from different sources in order to engineer non-native, functional systems. While there is a lot of potential for synthetic biology to revolutionize processes, such as the production of pharmaceuticals, engineering synthetic systems has been challenging. It is oftentimes necessary to explore a large design space to balance the levels of interacting components in the circuit. There are also times where it is desirable to incorporate enzymes that have non-biological functions into a synthetic circuit. Tuning the levels of different components, however, is often restricted to a fixed operating point, and this makes synthetic systems sensitive to changes in the environment. Natural systems are able to respond dynamically to a changing environment by obtaining information relevant to the function of the circuit. This work addresses these problems by establishing frameworks and mechanisms that allow synthetic circuits to communicate with the environment, maintain fixed ratios between components, and potentially add new parts that are outside the realm of current biological function. These frameworks provide a way for synthetic circuits to behave more like natural circuits by enabling a dynamic response, and provide a systematic and rational way to search design space to an experimentally tractable size where likely solutions exist. We hope that the contributions described below will aid in allowing synthetic biology to realize its potential.
Resumo:
The authors developed a time dependent method to study the single molecule dynamics of a simple gene regulatory network: a repressilator with three genes mutually repressing each other. They quantitatively characterize the time evolution dynamics of the repressilator. Furthermore, they study purely dynamical issues such as statistical fluctuations and noise evolution. They illustrated some important features of the biological network such as monostability, spirals, and limit cycle oscillation. Explicit time dependent Fano factors which describe noise evolution and show statistical fluctuations out of equilibrium can be significant and far from the Poisson distribution. They explore the phase space and the interrelationships among fluctuations, order, amplitude, and period of oscillations of the repressilators. The authors found that repressilators follow ordered limit cycle orbits and are more likely to appear in the lower fluctuating regions. The amplitude of the repressilators increases as the suppressing of the genes decreases and production of proteins increases. The oscillation period of the repressilators decreases as the suppressing of the genes decreases and production of proteins increases.
Resumo:
In decision making problems where we need to choose a particular decision or alternative from a set of possible choices, we often have some preferences which determine if we prefer one decision over another. When these preferences give us an ordering on the decisions that is complete, then it is easy to choose the best or one of the best decisions. However it often occurs that the preferences relation is partially ordered, and we have no best decision. In this thesis, we look at what happens when we have such a partial order over a set of decisions, in particular when we have multiple orderings on a set of decisions, and we present a framework for qualitative decision making. We look at the different natural notions of optimal decision that occur in this framework, which gives us different optimality classes, and we examine the relationships between these classes. We then look in particular at a qualitative preference relation called Sorted-Pareto Dominance, which is an extension of Pareto Dominance, and we give a semantics for this relation as one that is compatible with any order-preserving mapping of an ordinal preference scale to a numerical one. We apply Sorted-Pareto dominance to a Soft Constraints setting, where we solve problems in which the soft constraints associate qualitative preferences to decisions in a decision problem. We also examine the Sorted-Pareto dominance relation in the context of our qualitative decision making framework, looking at the relevant optimality classes for the Sorted-Pareto case, which gives us classes of decisions that are necessarily optimal, and optimal for some choice of mapping of an ordinal scale to a quantitative one. We provide some empirical analysis of Sorted-Pareto constraints problems and examine the optimality classes that result.
Resumo:
Although many feature selection methods for classification have been developed, there is a need to identify genes in high-dimensional data with censored survival outcomes. Traditional methods for gene selection in classification problems have several drawbacks. First, the majority of the gene selection approaches for classification are single-gene based. Second, many of the gene selection procedures are not embedded within the algorithm itself. The technique of random forests has been found to perform well in high-dimensional data settings with survival outcomes. It also has an embedded feature to identify variables of importance. Therefore, it is an ideal candidate for gene selection in high-dimensional data with survival outcomes. In this paper, we develop a novel method based on the random forests to identify a set of prognostic genes. We compare our method with several machine learning methods and various node split criteria using several real data sets. Our method performed well in both simulations and real data analysis.Additionally, we have shown the advantages of our approach over single-gene-based approaches. Our method incorporates multivariate correlations in microarray data for survival outcomes. The described method allows us to better utilize the information available from microarray data with survival outcomes.
Resumo:
Polymerase chain reaction (PCR) assessment of clonal immunoglobulin (Ig) and T-cell receptor (TCR) gene rearrangements is an important diagnostic tool in mature B-cell neoplasms. However, lack of standardized PCR protocols resulting in a high level of false negativity has hampered comparability of data in previous clonality studies. In order to address these problems, 22 European laboratories investigated the Ig/TCR rearrangement patterns as well as t(14;18) and t(11;14) translocations of 369 B-cell malignancies belonging to five WHO-defined entities using the standardized BIOMED-2 multiplex PCR tubes accompanied by international pathology panel review. B-cell clonality was detected by combined use of the IGH and IGK multiplex PCR assays in all 260 definitive cases of B-cell chronic lymphocytic leukemia (n¼56), mantle cell lymphoma (n¼54), marginal zone lymphoma (n¼41) and follicular lymphoma (n¼109). Two of 109 cases of diffuse large B-cell lymphoma showed no detectable clonal marker. The use of these techniques to assign cell lineage should be treated with caution as additional clonal TCR gene rearrangements were frequently detected in all disease categories. Our study indicates that the BIOMED-2 multiplex PCR assays provide a powerful strategy for clonality assessment in B-cell malignancies resulting in high Ig clonality detection rates particularly when IGH and IGK strategies are combined.
Resumo:
increasing prevalence of obesity combined with longevity will produce an epidemic of Type 2 (non-insulin-dependent) diabetes in the next 20 years. This. disease is associated with defects in insulin secretion, specifically abnormalities of insulin secretory kinetics and pancreatic beta-cell glucose responsiveness. Mechanisms underlying beta-cell dysfunction include glucose toxicity, lipotoxicity and beta-cell hyperactivity. Defects at various sites in beta-cell signal transduction pathways contribute, but no single lesion can account for the common form of Type 2 diabetes. Recent studies highlight diverse beta-cell actions of GLP-1 (glucagon-like peptide-1) and GIP (glucose-dependent insulinotropic polypeptide). These intestinal hormones target the beta-cell to stimulate glucose-dependent insulin secretion through activation of protein kinase A and associated pathways. Both increase gene expression and proinsulin biosynthesis, protect against apoptosis and stimulate replication/neogenesis of beta-cells. Incretin hormones therefore represent an exciting future multi-action solution to correct beta-cell defect in Type 2 diabetes.
Resumo:
An elegant way to prepare catalytically active microreactors is by applying a coating of zeolite crystals onto a metal microchannel structure. In this study the hydrothermal formation of ZSM-5 zeolitic coatings on AISI 316 stainless steel plates with a microchannel structure has been investigated at different synthesis mixture compositions. The procedures of coating and thermal treatment have also been optimized. Obtaining a uniform thickness of the coating within 0.5 mm wide microchannels requires a careful control of various synthesis variables. The role of these factors and the problems in the synthesis of these zeolitic coatings are discussed. In general, the synthesis is most sensitive to the H2O/Si ratio as well as to the orientation of the plates with respect to the gravity vector. Ratios of H2O/Si=130 and Si/template=13 were found to be optimal for the formation of a zeolitic film with a thickness of one crystal at a temperature of 130 degreesC and a synthesis time of about 35 h. At such conditions, ZSM-5 crystals were formed with a typical size of 1.5 mu mx1.5 mu mx1.0 mum and a very narrow (within 0.2 mum) crystal size distribution. The prepared samples proved to be active in the selective catalytic reduction (SCR) of NO with ammonia. The activity tests have been carried out in a plate-type microreactor. The microreactor shows no mass transfer limitations and a larger SCR reaction rate is observed in comparison with pelletized Ce-ZSM-5 catalysts; (C) 2001 Elsevier Science B.V. All rights reserved.