183 resultados para Large Estate


Relevância:

20.00% 20.00%

Publicador:

Resumo:

A key element for the development of suitable anti-cancer drugs is the identification of cancer-specific enzymatic activities that can be therapeutically targeted. Mucosa-associated lymphoid tissue transformation protein 1 (MALT1) is a proto-oncogene that contributes to tumorigenesis in diffuse large B-cell lymphoma (DLBCL) of the activated B-cell (ABC) subtype, the least curable subtype of DLBCL. Recent data suggest that MALT1 has proteolytic activity, but it is unknown whether this activity is relevant for tumor growth. Here we report that MALT1 is constitutively active in DLBCL lines of the ABC but not the GCB subtype. Inhibition of the MALT1 proteolytic activity led to reduced expression of growth factors and apoptosis inhibitors, and specifically affected the growth and survival of ABC DLBCL lines. These results demonstrate a key role for the proteolytic activity of MALT1 in DLBCL of the ABC subtype, and provide a rationale for the development of pharmacological inhibitors of MALT1 in DLBCL therapy.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Large animal models are an important resource for the understanding of human disease and for evaluating the applicability of new therapies to human patients. For many diseases, such as cone dystrophy, research effort is hampered by the lack of such models. Lentiviral transgenesis is a methodology broadly applicable to animals from many different species. When conjugated to the expression of a dominant mutant protein, this technology offers an attractive approach to generate new large animal models in a heterogeneous background. We adopted this strategy to mimic the phenotype diversity encounter in humans and generate a cohort of pigs for cone dystrophy by expressing a dominant mutant allele of the guanylate cyclase 2D (GUCY2D) gene. Sixty percent of the piglets were transgenic, with mutant GUCY2D mRNA detected in the retina of all animals tested. Functional impairment of vision was observed among the transgenic pigs at 3 months of age, with a follow-up at 1 year indicating a subsequent slower progression of phenotype. Abnormal retina morphology, notably among the cone photoreceptor cell population, was observed exclusively amongst the transgenic animals. Of particular note, these transgenic animals were characterized by a range in the severity of the phenotype, reflecting the human clinical situation. We demonstrate that a transgenic approach using lentiviral vectors offers a powerful tool for large animal model development. Not only is the efficiency of transgenesis higher than conventional transgenic methodology but this technique also produces a heterogeneous cohort of transgenic animals that mimics the genetic variation encountered in human patients.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

MOTIVATION: Analysis of millions of pyro-sequences is currently playing a crucial role in the advance of environmental microbiology. Taxonomy-independent, i.e. unsupervised, clustering of these sequences is essential for the definition of Operational Taxonomic Units. For this application, reproducibility and robustness should be the most sought after qualities, but have thus far largely been overlooked. RESULTS: More than 1 million hyper-variable internal transcribed spacer 1 (ITS1) sequences of fungal origin have been analyzed. The ITS1 sequences were first properly extracted from 454 reads using generalized profiles. Then, otupipe, cd-hit-454, ESPRIT-Tree and DBC454, a new algorithm presented here, were used to analyze the sequences. A numerical assay was developed to measure the reproducibility and robustness of these algorithms. DBC454 was the most robust, closely followed by ESPRIT-Tree. DBC454 features density-based hierarchical clustering, which complements the other methods by providing insights into the structure of the data. AVAILABILITY: An executable is freely available for non-commercial users at ftp://ftp.vital-it.ch/tools/dbc454. It is designed to run under MPI on a cluster of 64-bit Linux machines running Red Hat 4.x, or on a multi-core OSX system. CONTACT: dbc454@vital-it.ch or nicolas.guex@isb-sib.ch.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

For the last 2 decades, supertree reconstruction has been an active field of research and has seen the development of a large number of major algorithms. Because of the growing popularity of the supertree methods, it has become necessary to evaluate the performance of these algorithms to determine which are the best options (especially with regard to the supermatrix approach that is widely used). In this study, seven of the most commonly used supertree methods are investigated by using a large empirical data set (in terms of number of taxa and molecular markers) from the worldwide flowering plant family Sapindaceae. Supertree methods were evaluated using several criteria: similarity of the supertrees with the input trees, similarity between the supertrees and the total evidence tree, level of resolution of the supertree and computational time required by the algorithm. Additional analyses were also conducted on a reduced data set to test if the performance levels were affected by the heuristic searches rather than the algorithms themselves. Based on our results, two main groups of supertree methods were identified: on one hand, the matrix representation with parsimony (MRP), MinFlip, and MinCut methods performed well according to our criteria, whereas the average consensus, split fit, and most similar supertree methods showed a poorer performance or at least did not behave the same way as the total evidence tree. Results for the super distance matrix, that is, the most recent approach tested here, were promising with at least one derived method performing as well as MRP, MinFlip, and MinCut. The output of each method was only slightly improved when applied to the reduced data set, suggesting a correct behavior of the heuristic searches and a relatively low sensitivity of the algorithms to data set sizes and missing data. Results also showed that the MRP analyses could reach a high level of quality even when using a simple heuristic search strategy, with the exception of MRP with Purvis coding scheme and reversible parsimony. The future of supertrees lies in the implementation of a standardized heuristic search for all methods and the increase in computing power to handle large data sets. The latter would prove to be particularly useful for promising approaches such as the maximum quartet fit method that yet requires substantial computing power.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dans certaines portions des agglomérations (poches de pauvreté de centre-ville, couronnes suburbaines dégradées, espaces périurbains sans aménité), un cumul entre des inégalités sociales (pauvreté, chômage, etc.) et environnementales (exposition au bruit, aux risques industriels, etc.) peut être observé. La persistance de ces inégalités croisées dans le temps indique une tendance de fond : la capacité d'accéder à un cadre de vie de qualité n'est pas équitablement partagée parmi les individus. Ce constat interroge : comment se créent ces inégalités ? Comment infléchir cette tendance et faire la ville plus juste ?¦Apporter des réponses à cette problématique nécessite d'identifier les facteurs de causalités qui entrent en jeu dans le système de (re)production des inégalités urbaines. Le fonctionnement des marchés foncier et immobilier, la « tyrannie des petites décisions » et les politiques publiques à incidence spatiale sont principalement impliqués. Ces dernières, agissant sur tous les éléments du système, sont placées au coeur de ce travail. On va ainsi s'intéresser précisément à la manière dont les collectivités publiques pilotent la production de la ville contemporaine, en portant l'attention sur la maîtrise publique d'ouvrage (MPO) des grands projets urbains.¦Poser la question de la justice dans la fabrique de la ville implique également de questionner les référentiels normatifs de l'action publique : à quelle conception de la justice celle-ci doit- elle obéir? Quatre perspectives (radicale, substantialiste, procédurale et intégrative) sont caractérisées, chacune se traduisant par des principes d'action différenciés. Une méthodologie hybride - empruntant à la sociologie des organisations et à l'analyse des politiques publiques - vient clore le volet théorique, proposant par un détour métaphorique d'appréhender le projet urbain comme une pièce de théâtre dont le déroulement dépend du jeu d'acteurs.¦Cette méthodologie est utilisée dans le volet empirique de la recherche, qui consiste en une analyse de la MPO d'un projet urbain en cours dans la première couronne de l'agglomération lyonnaise : le Carré de Soie. Trois grands objectifs sont poursuivis : descriptif (reconstruire le scénario), analytique (évaluer la nature de la pièce : conte de fée, tragédie ou match d'improvisation ?) et prescriptif (tirer la morale de l'histoire). La description de la MPO montre le déploiement successif de quatre stratégies de pilotage, dont les implications sur les temporalités, le contenu du projet (programmes, morphologies) et les financements publics vont être déterminantes. Sur la base de l'analyse, plusieurs recommandations peuvent être formulées - importance de l'anticipation et de l'articulation entre planification et stratégie foncière notamment - pour permettre à la sphère publique de dominer le jeu et d'assurer la production de justice par le projet urbain (réalisation puis entretien des équipements et espaces publics, financement de logements de qualité à destination d'un large éventail de populations, etc.). Plus généralement, un décalage problématique peut être souligné entre les territoires stratégiques pour le développement de l'agglomération et les capacités de portage limitées des communes concernées. Ce déficit plaide pour le renforcement des capacités d'investissement de la structure intercommunale.¦La seule logique du marché (foncier, immobilier) mène à la polarisation sociale et à la production d'inégalités urbaines. Faire la ville juste nécessite une forte volonté des collectivités publiques, laquelle doit se traduire aussi bien dans l'ambition affichée - une juste hiérarchisation des priorités dans le développement urbain - que dans son opérationnalisation - une juste maîtrise publique d'ouvrage des projets urbains.¦Inner-city neighborhoods, poor outskirts, and peri-urban spaces with no amenities usually suffer from social and environmental inequalities, such as poverty, unemployment, and exposure to noise and industrial hazards. The observed persistence of these inequalities over time points to an underlying trend - namely, that access to proper living conditions is fundamentally unequal, thus eliciting the question of how such inequalities are effected and how this trend can be reversed so as to build a more equitable city.¦Providing answers to such questions requires that the causal factors at play within the system of (re)production of urban inequalities be identified. Real estate markets, "micromotives and macrobehavior", and public policies that bear on space are mostly involved. The latter are central in that they act on all the elements of the system. This thesis therefore focuses on the way public authorities shape the production of contemporary cities, by studying the public project ownership of major urban projects.¦The study of justice within the urban fabric also implies that the normative frames of reference of public action be questioned: what conception of justice should public action refer to? This thesis examines four perspectives (radical, substantialist, procedural, and integrative) each of which results in different principles of action. This theoretical part is concluded by a hybrid methodology that draws from sociology of organizations and public policy analysis and that suggests that the urban project may be understood as a play, whose outcome hinges on the actors' acting.¦This methodology is applied to the empirical analysis of the public project ownership of an ongoing urban project in the Lyon first-ring suburbs: the Carré de Soie. Three main objectives are pursued: descriptive (reconstructing the scenario), analytical (assessing the nature of the play - fairy tale, tragedy or improvisation match), and prescriptive (drawing the moral of the story). The description of the public project ownership shows the successive deployment of four control strategies, whose implications on deadlines, project content (programs, morphologies), and public funding are significant. Building on the analysis, several recommendations can be made to allow the public sphere to control the process and ensure the urban project produces equity (most notably, anticipation and articulation of planning and real- estate strategy, as well as provision and maintenance of equipment and public spaces, funding of quality housing for a wide range of populations, etc.). More generally, a gap can be highlighted between those territories that are strategic to the development of the agglomeration and the limited resources of the municipalities involved. This deficit calls for strengthening the investment abilities of the intermunicipal structure.¦By itself, the real-estate market logic brings about social polarization and urban inequalities. Building an equitable city requires a strong will on the part of public authorities, a will that must be reflected both in the stated ambition - setting priorities of urban development equitably - and in its implementation managing urban public projects fairly.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

BACKGROUND: Genotypes obtained with commercial SNP arrays have been extensively used in many large case-control or population-based cohorts for SNP-based genome-wide association studies for a multitude of traits. Yet, these genotypes capture only a small fraction of the variance of the studied traits. Genomic structural variants (GSV) such as Copy Number Variation (CNV) may account for part of the missing heritability, but their comprehensive detection requires either next-generation arrays or sequencing. Sophisticated algorithms that infer CNVs by combining the intensities from SNP-probes for the two alleles can already be used to extract a partial view of such GSV from existing data sets. RESULTS: Here we present several advances to facilitate the latter approach. First, we introduce a novel CNV detection method based on a Gaussian Mixture Model. Second, we propose a new algorithm, PCA merge, for combining copy-number profiles from many individuals into consensus regions. We applied both our new methods as well as existing ones to data from 5612 individuals from the CoLaus study who were genotyped on Affymetrix 500K arrays. We developed a number of procedures in order to evaluate the performance of the different methods. This includes comparison with previously published CNVs as well as using a replication sample of 239 individuals, genotyped with Illumina 550K arrays. We also established a new evaluation procedure that employs the fact that related individuals are expected to share their CNVs more frequently than randomly selected individuals. The ability to detect both rare and common CNVs provides a valuable resource that will facilitate association studies exploring potential phenotypic associations with CNVs. CONCLUSION: Our new methodologies for CNV detection and their evaluation will help in extracting additional information from the large amount of SNP-genotyping data on various cohorts and use this to explore structural variants and their impact on complex traits.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Volumes of data used in science and industry are growing rapidly. When researchers face the challenge of analyzing them, their format is often the first obstacle. Lack of standardized ways of exploring different data layouts requires an effort each time to solve the problem from scratch. Possibility to access data in a rich, uniform manner, e.g. using Structured Query Language (SQL) would offer expressiveness and user-friendliness. Comma-separated values (CSV) are one of the most common data storage formats. Despite its simplicity, with growing file size handling it becomes non-trivial. Importing CSVs into existing databases is time-consuming and troublesome, or even impossible if its horizontal dimension reaches thousands of columns. Most databases are optimized for handling large number of rows rather than columns, therefore, performance for datasets with non-typical layouts is often unacceptable. Other challenges include schema creation, updates and repeated data imports. To address the above-mentioned problems, I present a system for accessing very large CSV-based datasets by means of SQL. It's characterized by: "no copy" approach - data stay mostly in the CSV files; "zero configuration" - no need to specify database schema; written in C++, with boost [1], SQLite [2] and Qt [3], doesn't require installation and has very small size; query rewriting, dynamic creation of indices for appropriate columns and static data retrieval directly from CSV files ensure efficient plan execution; effortless support for millions of columns; due to per-value typing, using mixed text/numbers data is easy; very simple network protocol provides efficient interface for MATLAB and reduces implementation time for other languages. The software is available as freeware along with educational videos on its website [4]. It doesn't need any prerequisites to run, as all of the libraries are included in the distribution package. I test it against existing database solutions using a battery of benchmarks and discuss the results.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the vast majority of bottom-up proteomics studies, protein digestion is performed using only mammalian trypsin. Although it is clearly the best enzyme available, the sole use of trypsin rarely leads to complete sequence coverage, even for abundant proteins. It is commonly assumed that this is because many tryptic peptides are either too short or too long to be identified by RPLC-MS/MS. We show through in silico analysis that 20-30% of the total sequence of three proteomes (Schizosaccharomyces pombe, Saccharomyces cerevisiae, and Homo sapiens) is expected to be covered by Large post-Trypsin Peptides (LpTPs) with M(r) above 3000 Da. We then established size exclusion chromatography to fractionate complex yeast tryptic digests into pools of peptides based on size. We found that secondary digestion of LpTPs followed by LC-MS/MS analysis leads to a significant increase in identified proteins and a 32-50% relative increase in average sequence coverage compared to trypsin digestion alone. Application of the developed strategy to analyze the phosphoproteomes of S. pombe and of a human cell line identified a significant fraction of novel phosphosites. Overall our data indicate that specific targeting of LpTPs can complement standard bottom-up workflows to reveal a largely neglected portion of the proteome.