25 resultados para structure prediction
em Consorci de Serveis Universitaris de Catalunya (CSUC), Spain
Resumo:
Experimentally, Ce2O3 films are used to study cerium oxide in its fully or partially reduced state, as present in many applications. We have explored the space of low energy Ce2O3 nanofilms using structure prediction and density functional calculations, yielding more than 30 distinct nanofilm structures. First, our results help to rationalize the roles of thermodynamics and kinetics in the preparation of reduced ceria nanofilms with different bulk crystalline structures (e.g. A-type or bixbyite) depending on the support used. Second, we predict a novel, as yet experimentally unresolved, nanofilm which has a structure that does not correspond to any previously reported bulk A2B3 phase and which has an energetic stability between that of A-type and bixbyite. To assist identification and fabrication of this new Ce2O3 nanofilm we calculate some observable properties and propose supports for its epitaxial growth.
Resumo:
Background: Recent advances on high-throughput technologies have produced a vast amount of protein sequences, while the number of high-resolution structures has seen a limited increase. This has impelled the production of many strategies to built protein structures from its sequence, generating a considerable amount of alternative models. The selection of the closest model to the native conformation has thus become crucial for structure prediction. Several methods have been developed to score protein models by energies, knowledge-based potentials and combination of both.Results: Here, we present and demonstrate a theory to split the knowledge-based potentials in scoring terms biologically meaningful and to combine them in new scores to predict near-native structures. Our strategy allows circumventing the problem of defining the reference state. In this approach we give the proof for a simple and linear application that can be further improved by optimizing the combination of Zscores. Using the simplest composite score () we obtained predictions similar to state-of-the-art methods. Besides, our approach has the advantage of identifying the most relevant terms involved in the stability of the protein structure. Finally, we also use the composite Zscores to assess the conformation of models and to detect local errors.Conclusion: We have introduced a method to split knowledge-based potentials and to solve the problem of defining a reference state. The new scores have detected near-native structures as accurately as state-of-art methods and have been successful to identify wrongly modeled regions of many near-native conformations.
Resumo:
In a number of programs for gene structure prediction in higher eukaryotic genomic sequences, exon prediction is decoupled from gene assembly: a large pool of candidate exons is predicted and scored from features located in the query DNA sequence, and candidate genes are assembled from such a pool as sequences of nonoverlapping frame-compatible exons. Genes are scored as a function of the scores of the assembled exons, and the highest scoring candidate gene is assumed to be the most likely gene encoded by the query DNA sequence. Considering additive gene scoring functions, currently available algorithms to determine such a highest scoring candidate gene run in time proportional to the square of the number of predicted exons. Here, we present an algorithm whose running time grows only linearly with the size of the set of predicted exons. Polynomial algorithms rely on the fact that, while scanning the set of predicted exons, the highest scoring gene ending in a given exon can be obtained by appending the exon to the highest scoring among the highest scoring genes ending at each compatible preceding exon. The algorithm here relies on the simple fact that such highest scoring gene can be stored and updated. This requires scanning the set of predicted exons simultaneously by increasing acceptor and donor position. On the other hand, the algorithm described here does not assume an underlying gene structure model. Indeed, the definition of valid gene structures is externally defined in the so-called Gene Model. The Gene Model specifies simply which gene features are allowed immediately upstream which other gene features in valid gene structures. This allows for great flexibility in formulating the gene identification problem. In particular it allows for multiple-gene two-strand predictions and for considering gene features other than coding exons (such as promoter elements) in valid gene structures.
Resumo:
Conventional methods of gene prediction rely on the recognition of DNA-sequence signals, the coding potential or the comparison of a genomic sequence with a cDNA, EST, or protein database. Reasons for limited accuracy in many circumstances are species-specific training and the incompleteness of reference databases. Lately, comparative genome analysis has attracted increasing attention. Several analysis tools that are based on human/mouse comparisons are already available. Here, we present a program for the prediction of protein-coding genes, termed SGP-1 (Syntenic Gene Prediction), which is based on the similarity of homologous genomic sequences. In contrast to most existing tools, the accuracy of SGP-1 depends little on species-specific properties such as codon usage or the nucleotide distribution. SGP-1 may therefore be applied to nonstandard model organisms in vertebrates as well as in plants, without the need for extensive parameter training. In addition to predicting genes in large-scale genomic sequences, the program may be useful to validate gene structure annotations from databases. To this end, SGP-1 output also contains comparisons between predicted and annotated gene structures in HTML format. The program can be accessed via a Web server at http://soft.ice.mpg.de/sgp-1. The source code, written in ANSI C, is available on request from the authors.
Resumo:
One of the first useful products from the human genome will be a set of predicted genes. Besides its intrinsic scientific interest, the accuracy and completeness of this data set is of considerable importance for human health and medicine. Though progress has been made on computational gene identification in terms of both methods and accuracy evaluation measures, most of the sequence sets in which the programs are tested are short genomic sequences, and there is concern that these accuracy measures may not extrapolate well to larger, more challenging data sets. Given the absence of experimentally verified large genomic data sets, we constructed a semiartificial test set comprising a number of short single-gene genomic sequences with randomly generated intergenic regions. This test set, which should still present an easier problem than real human genomic sequence, mimics the approximately 200kb long BACs being sequenced. In our experiments with these longer genomic sequences, the accuracy of GENSCAN, one of the most accurate ab initio gene prediction programs, dropped significantly, although its sensitivity remained high. Conversely, the accuracy of similarity-based programs, such as GENEWISE, PROCRUSTES, and BLASTX was not affected significantly by the presence of random intergenic sequence, but depended on the strength of the similarity to the protein homolog. As expected, the accuracy dropped if the models were built using more distant homologs, and we were able to quantitatively estimate this decline. However, the specificities of these techniques are still rather good even when the similarity is weak, which is a desirable characteristic for driving expensive follow-up experiments. Our experiments suggest that though gene prediction will improve with every new protein that is discovered and through improvements in the current set of tools, we still have a long way to go before we can decipher the precise exonic structure of every gene in the human genome using purely computational methodology.
Resumo:
Evaluating other individuals with respect to personality characteristics plays a crucial role in human relations and it is the focus of attention for research in diverse fields such as psychology and interactive computer systems. In psychology, face perception has been recognized as a key component of this evaluation system. Multiple studies suggest that observers use face information to infer personality characteristics. Interactive computer systems are trying to take advantage of these findings and apply them to increase the natural aspect of interaction and to improve the performance of interactive computer systems. Here, we experimentally test whether the automatic prediction of facial trait judgments (e.g. dominance) can be made by using the full appearance information of the face and whether a reduced representation of its structure is sufficient. We evaluate two separate approaches: a holistic representation model using the facial appearance information and a structural model constructed from the relations among facial salient points. State of the art machine learning methods are applied to a) derive a facial trait judgment model from training data and b) predict a facial trait value for any face. Furthermore, we address the issue of whether there are specific structural relations among facial points that predict perception of facial traits. Experimental results over a set of labeled data (9 different trait evaluations) and classification rules (4 rules) suggest that a) prediction of perception of facial traits is learnable by both holistic and structural approaches; b) the most reliable prediction of facial trait judgments is obtained by certain type of holistic descriptions of the face appearance; and c) for some traits such as attractiveness and extroversion, there are relationships between specific structural features and social perceptions.
Resumo:
The linear prediction coding of speech is based in the assumption that the generation model is autoregresive. In this paper we propose a structure to cope with the nonlinear effects presents in the generation of the speech signal. This structure will consist of two stages, the first one will be a classical linear prediction filter, and the second one will model the residual signal by means of two nonlinearities between a linear filter. The coefficients of this filter are computed by means of a gradient search on the score function. This is done in order to deal with the fact that the probability distribution of the residual signal still is not gaussian. This fact is taken into account when the coefficients are computed by a ML estimate. The algorithm based on the minimization of a high-order statistics criterion, uses on-line estimation of the residue statistics and is based on blind deconvolution of Wiener systems [1]. Improvements in the experimental results with speech signals emphasize on the interest of this approach.
Resumo:
During the process of language development, one of the most important tasks that children must face is that of identifying the grammatical category to which words in their language belong. This is essential in order to be able to form grammatically correct utterances. How do children proceed in order to classify words in their language and assign them to their corresponding grammatical category? The present study investigates the usefulness of phonological information for the categorization of nouns in English, given the fact that it is phonology the first source of information that might be available to prelinguistic infants who lack access to semantic information or complex morphosyntactic information. We analyse four different corpora containing linguistic samples of English speaking mothers addressing their children in order to explore the reliability with which words are represented in mothers’ speech based on several phonological criteria. The results of the analysis confirm the prediction that most of the words to which English learning infants are exposed during the first two years of life can be accounted for in terms of their phonological resemblance
Resumo:
The purpose of this article is to introduce a Cartesian product structure into the social choice theoretical framework and to examine if new possibility results to Gibbard's and Sen's paradoxes can be developed thanks to it. We believe that a Cartesian product structure is a pertinent way to describe individual rights in the social choice theory since it discriminates the personal features comprised in each social state. First we define some conceptual and formal tools related to the Cartesian product structure. We then apply these notions to Gibbard's paradox and to Sen's impossibility of a Paretian liberal. Finally we compare the advantages of our approach to other solutions proposed in the literature for both impossibility theorems.
Resumo:
The metropolitan spatial structure displays various patterns, sometimes monocentricity and sometimes multicentricity, which seems much more complicated than the exponential density function used in classic works such as Clark (1961), Muth (1969) or Mills (1973) among others, can effectively represent. It seems that a more flexible density function,such as cubic spline function (Anderson (1982), Zheng (1991), etc.) to describe the density-accessibility relationship is needed. Also, accessibility, the fundamental determinant of density variations, is only partly captured by the inclusion of distance to the city centre as an explanatory variable. Steen (1986) has proposed to correct that miss-especification by including an additional gradient for distance to the nearest transportation axis. In identifying the determinants of urban spatial structure in the context of inter-urban systems, some of the variables proposed by Muth (1969), Mills (1973) and Alperovich (1983) such as city age or population, make no sense in the case of a single urban system. All three criticism to the exponential density function and its determinants apply for the Barcelona Metropolitan Region, a polycentric conurbation structured on well defined transportation axes.
Resumo:
This paper characterizes the equilibria in airline networks and their welfare implications in an unregulated environment. Competing airlines may adopt either fully-connected (FC) or hub-and-spoke (HS) network structures; and passengers exhibiting low brand loyalty to their preferred carrier choose an outside option to travel so that markets are partially served by airlines. In this context, carriers adopt hubbing strategies when costs are sufficiently low, and asymmetric equilibria where one carrier chooses a FC strategy and the other chooses a HS strategy may arise. Quite interestingly, flight frequency can become excessive under HS network configurations.
Resumo:
This paper investigates experimentally how organisational decision processes affect the moral motivations of actors inside a firm that must forego profits to reduce harming a third party. In a "vertical" treatment, one insider unilaterally sets the harm-reduction strategy; the other can only accept or quit. In a "horizontal" treatment, the insiders decide by consensus. Our 2-by-2 design also controls for communication effects. In our data, communication makes vertical firms more ethical; voice appears to mitigate "responsibility-alleviation" in that subordinates with voice feel responsible for what their firms do. Vertical firms are then more ethical than the horizontal firms for which our bargaining data reveal a dynamic form of responsibility-alleviation and our chat data indicate a strong "insider-outsider" effect.
Resumo:
We analyse the implications of optimal taxation for the stochastic behaviour of debt. We show that when a government pursues an optimal fiscal policy under complete markets, the value of debt has the same or less persistence than other variables in the economy and it declines in response to shocks that cause the deficit to increase. By contrast, under incomplete markets debt shows more persistence than other variables and it increases in response to shocks that cause a higher deficit. Data for US government debt reveals diametrically opposite results from those of complete markets and is much more supportive of bond market incompleteness.
Resumo:
Forest fires are a serious threat to humans and nature from an ecological, social and economic point of view. Predicting their behaviour by simulation still delivers unreliable results and remains a challenging task. Latest approaches try to calibrate input variables, often tainted with imprecision, using optimisation techniques like Genetic Algorithms. To converge faster towards fitter solutions, the GA is guided with knowledge obtained from historical or synthetical fires. We developed a robust and efficient knowledge storage and retrieval method. Nearest neighbour search is applied to find the fire configuration from knowledge base most similar to the current configuration. Therefore, a distance measure was elaborated and implemented in several ways. Experiments show the performance of the different implementations regarding occupied storage and retrieval time with overly satisfactory results.
Resumo:
We establish a Quillen model structure on simplicial(symmetric) multicategories. It extends the model structure on simplicial categories due to J. Bergner [2]. We observe that our technique of proof enables us to prove a similar result for (symmetric) multicategories enriched over other monoidal model categories than simplicial sets. Examples include small categories, simplicial abelian groups and compactly generated Hausdorff spaces.