933 resultados para database integration


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Pendant ma thèse de doctorat, j'ai utilisé des espèces modèles, comme la souris et le poisson-zèbre, pour étudier les facteurs qui affectent l'évolution des gènes et leur expression. Plus précisément, j'ai montré que l'anatomie et le développement sont des facteurs clés à prendre en compte, car ils influencent la vitesse d'évolution de la séquence des gènes, l'impact sur eux de mutations (i.e. la délétion du gène est-elle létale ?), et leur tendance à se dupliquer. Où et quand il est exprimé impose à un gène certaines contraintes ou au contraire lui donne des opportunités d'évoluer. J'ai pu comparer ces tendances aux modèles classiques d'évolution de la morphologie, que l'on pensait auparavant refléter directement les contraintes s'appliquant sur le génome. Nous avons montré que les contraintes entre ces deux niveaux d'organisation ne peuvent pas être transférées simplement : il n'y a pas de lien direct entre la conservation du génotype et celle de phénotypes comme la morphologie. Ce travail a été possible grâce au développement d'outils bioinformatiques. Notamment, j'ai travaillé sur le développement de la base de données Bgee, qui a pour but de comparer l'expression des gènes entre différentes espèces de manière automatique et à large échelle. Cela implique une formalisation de l'anatomie, du développement et de concepts liés à l'homologie grâce à l'utilisation d'ontologies. Une intégration cohérente de données d'expression hétérogènes (puces à ADN, marqueurs de séquence exprimée, hybridations in situ) a aussi été nécessaire. Cette base de données est mise à jour régulièrement et disponible librement. Elle devrait contribuer à étendre les possibilités de comparaison de l'expression des gènes entre espèces pour des études d'évo-devo (évolution du développement) et de génomique. During my PhD, I used model species of vertebrates, such as mouse and zebrafish, to study factors affecting the evolution of genes and their expression. More precisely I have shown that anatomy and development are key factors to take into account, influencing the rate of gene sequence evolution, the impact of mutations (i.e. is the deletion of a gene lethal?), and the propensity of a gene to duplicate. Where and when genes are expressed imposes constraints, or on the contrary leaves them some opportunity to evolve. We analyzed these patterns in relation to classical models of morphological evolution in vertebrates, which were previously thought to directly reflect constraints on the genomes. We showed that the patterns of evolution at these two levels of organization do not translate smoothly: there is no direct link between the conservation of genotype and phenotypes such as morphology. This work was made possible by the development of bioinformatics tools. Notably, I worked on the development of the database Bgee, which aims at comparing gene expression between different species in an automated and large-scale way. This involves the formalization of anatomy, development, and concepts related to homology, through the use of ontologies. A coherent integration of heterogeneous expression data (microarray, expressed sequence tags, in situ hybridizations) is also required. This database is regularly updated and freely available. It should contribute to extend the possibilities for comparison of gene expression between species in evo-devo and genomics studies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An essential step of the life cycle of retroviruses is the stable insertion of a copy of their DNA genome into the host cell genome, and lentiviruses are no exception. This integration step, catalyzed by the viral-encoded integrase, ensures long-term expression of the viral genes, thus allowing a productive viral replication and rendering retroviral vectors also attractive for the field of gene therapy. At the same time, this ability to integrate into the host genome raises safety concerns regarding the use of retroviral-based gene therapy vectors, due to the genomic locations of integration sites. The availability of the human genome sequence made possible the analysis of the integration site preferences, which revealed to be nonrandom and retrovirus-specific, i.e. all lentiviruses studied so far favor integration in active transcription units, while other retroviruses have a different integration site distribution. Several mechanisms have been proposed that may influence integration targeting, which include (i) chromatin accessibility, (ii) cell cycle effects, and (iii) tethering proteins. Recent data provide evidence that integration site selection can occur via a tethering mechanism, through the recruitment of the lentiviral integrase by the cellular LEDGF/p75 protein, both proteins being the two major players in lentiviral integration targeting.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Eukaryotic Promoter Database (EPD) is an annotated non-redundant collection of eukaryotic POL II promoters, experimentally defined by a transcription start site (TSS). There may be multiple promoter entries for a single gene. The underlying experimental evidence comes from journal articles and, starting from release 73, from 5' ESTs of full-length cDNA clones used for so-called in silico primer extension. Access to promoter sequences is provided by pointers to TSS positions in nucleotide sequence entries. The annotation part of an EPD entry includes a description of the type and source of the initiation site mapping data, links to other biological databases and bibliographic references. EPD is structured in a way that facilitates dynamic extraction of biologically meaningful promoter subsets for comparative sequence analysis. Web-based interfaces have been developed that enable the user to view EPD entries in different formats, to select and extract promoter sequences according to a variety of criteria and to navigate to related databases exploiting different cross-references. Tools for analysing sequence motifs around TSSs defined in EPD are provided by the signal search analysis server. EPD can be accessed at http://www.epd. isb-sib.ch.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This report looks at the first year of data collection on childhood obesity and the difficulties of this project. This report summarises the data available and provides high level analysis of the prevalence of overweight and obesity among the children measured in 2005-06. The report was produced for the Department of Health by the South East Public Health Observatory on behalf of the Association of Public Health Observatories, and published in December 2006.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A report for the Department of Health by the South East Public Health Observatory on behalf of the Association of Public Health Observatories. The report looks at the first year of data collection on childhood obesity and the difficulties of this project. This report summarises the data available and provides high level analysis of the prevalence of overweight and obesity among the children measured in 2005-06.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This bulletin summarises information on individuals referred to the Northern Ireland Substitute Prescribing Scheme (SPS). It relates to those referred up to and including the 31 March 2014 and focuses on those patients in contact with Substitute Prescribing treatment services during 2013/14.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This bulletin summarises information on individuals referred to the Northern Ireland Substitute Prescribing Scheme (SPS). It relates to those referred up to and including the 31 March 2013 and focuses on those patients in contact with Substitute Prescribing treatment services during 2012/13.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El presente trabajo tiene por objetivo ofrecer una solución para la creación de un índice espacial para la extensión JASPA (Java SPAtial) sobre la base de datos H2. El algoritmo de indexación elegido para la implementación del índice espacial ha sido el Rtree.La implementación se ha realizado con el lenguaje de programación Java lo que ha facilitado su integración con la extensión JASPA y la base de datos H2, dado que ambos proyectos están codificados en el mismo lenguaje. El índice es persistente en memoria secundaria en una tabla de la propia base de datos H2.La solución que se propone está limitada a operaciones espaciales en dos dimensiones y es lo suficientemente flexible como para que no se haya necesitado modificar ni el código fuente de JASPA, ni de H2. Además, se ha previsto que el algoritmo de indexación se pueda mejorar o sustituir fácilmente.Por último, se ha tenido en cuenta que el proceso de creación y manipulación de los índices espaciales sea intuitivo y fácil de usar.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

En una economia basada en el coneixement, la innovació del producte es considera un factor clau a l'hora de determinar la competitivitat, la productivitat i el creixement d'una companyia. No obstant això, l'experiència de les companyies demostra la necessitat d'un nou model de gestió de la innovació del producte: una gestió basada en el màrqueting, en què la cooperació i l'ús intensiu de les tecnologies de la informació i de la comunicació (TIC) són especialment importants. En els darrers anys, la bibliografia sobre màrqueting ha analitzat el paper de la cooperació en l'èxit del procés d'innovació. No obstant això, fins ara pocs treballs han estudiat el paper que té l'ús de les TIC en el màrqueting en l'èxit del desenvolupament de nous productes (NPD, New Product Development en anglès). És una omissió curiosa, tenint en compte que el nou entorn competitiu és definit per una economia i una societat basades principalment en l'ús intensiu de les TIC i del coneixement. L'objectiu d'aquest treball és investigar el paper que l'ús de les TIC en el màrqueting té en el procés de desenvolupament de nous productes, com a element que reforça la integració d'agents al projecte, afavorint l'establiment de relacions dirigides a la cooperació i l'adquisició d'intel·ligència de mercat útil en el procés de desenvolupament de nous productes. L'estudi d'una mostra de 2.038 companyies de tots els sectors de l'activitat econòmica a Catalunya ens permet contrastar hipòtesis inicials i establir un perfil de companyia innovadora basat en les importants relacions que hi ha entre la innovació, l'ús de TIC en el màrqueting i la integració. Sobresurten dues idees en la nostra anàlisi. En primer lloc, l'ús intensiu de les TIC en el màrqueting fa que la companyia sigui més innovadora, ja que percep que el seu ús ajuda a superar barreres a la innovació i accelera els processos, que es tornen més eficients. En segon lloc, incrementant l'ús de les TIC en el màrqueting es fa augmentar la predisposició de la companyia a integrar agents particulars en l'entorn de negoci en el desenvolupament del procés d'innovació i a col·laborar-hi, de manera que es millora el grau d'adaptació del nou producte a les demandes del mercat.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The integration of geophysical data into the subsurface characterization problem has been shown in many cases to significantly improve hydrological knowledge by providing information at spatial scales and locations that is unattainable using conventional hydrological measurement techniques. The investigation of exactly how much benefit can be brought by geophysical data in terms of its effect on hydrological predictions, however, has received considerably less attention in the literature. Here, we examine the potential hydrological benefits brought by a recently introduced simulated annealing (SA) conditional stochastic simulation method designed for the assimilation of diverse hydrogeophysical data sets. We consider the specific case of integrating crosshole ground-penetrating radar (GPR) and borehole porosity log data to characterize the porosity distribution in saturated heterogeneous aquifers. In many cases, porosity is linked to hydraulic conductivity and thus to flow and transport behavior. To perform our evaluation, we first generate a number of synthetic porosity fields exhibiting varying degrees of spatial continuity and structural complexity. Next, we simulate the collection of crosshole GPR data between several boreholes in these fields, and the collection of porosity log data at the borehole locations. The inverted GPR data, together with the porosity logs, are then used to reconstruct the porosity field using the SA-based method, along with a number of other more elementary approaches. Assuming that the grid-cell-scale relationship between porosity and hydraulic conductivity is unique and known, the porosity realizations are then used in groundwater flow and contaminant transport simulations to assess the benefits and limitations of the different approaches.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background The 'database search problem', that is, the strengthening of a case - in terms of probative value - against an individual who is found as a result of a database search, has been approached during the last two decades with substantial mathematical analyses, accompanied by lively debate and centrally opposing conclusions. This represents a challenging obstacle in teaching but also hinders a balanced and coherent discussion of the topic within the wider scientific and legal community. This paper revisits and tracks the associated mathematical analyses in terms of Bayesian networks. Their derivation and discussion for capturing probabilistic arguments that explain the database search problem are outlined in detail. The resulting Bayesian networks offer a distinct view on the main debated issues, along with further clarity. Methods As a general framework for representing and analyzing formal arguments in probabilistic reasoning about uncertain target propositions (that is, whether or not a given individual is the source of a crime stain), this paper relies on graphical probability models, in particular, Bayesian networks. This graphical probability modeling approach is used to capture, within a single model, a series of key variables, such as the number of individuals in a database, the size of the population of potential crime stain sources, and the rarity of the corresponding analytical characteristics in a relevant population. Results This paper demonstrates the feasibility of deriving Bayesian network structures for analyzing, representing, and tracking the database search problem. The output of the proposed models can be shown to agree with existing but exclusively formulaic approaches. Conclusions The proposed Bayesian networks allow one to capture and analyze the currently most well-supported but reputedly counter-intuitive and difficult solution to the database search problem in a way that goes beyond the traditional, purely formulaic expressions. The method's graphical environment, along with its computational and probabilistic architectures, represents a rich package that offers analysts and discussants with additional modes of interaction, concise representation, and coherent communication.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we consider the ATM networks in which the virtual path concept is implemented. The question of how to multiplex two or more diverse traffic classes while providing different quality of service requirements is a very complicated open problem. Two distinct options are available: integration and segregation. In an integration approach all the traffic from different connections are multiplexed onto one VP. This implies that the most restrictive QOS requirements must be applied to all services. Therefore, link utilization will be decreased because unnecessarily stringent QOS is provided to all connections. With the segregation approach the problem can be much simplified if different types of traffic are separated by assigning a VP with dedicated resources (buffers and links). Therefore, resources may not be efficiently utilized because no sharing of bandwidth can take place across the VP. The probability that the bandwidth required by the accepted connections exceeds the capacity of the link is evaluated with the probability of congestion (PC). Since the PC can be expressed as the CLP, we shall simply carry out bandwidth allocation using the PC. We first focus on the influence of some parameters (CLP, bit rate and burstiness) on the capacity required by a VP supporting a single traffic class using the new convolution approach. Numerical results are presented both to compare the required capacity and to observe which conditions under each approach are preferred