152 resultados para Datasets
Resumo:
Immunotherapy is a promising strategy for the treatment of various types of cancer. An antibody that targets programmed death ligand-1 (PD-L1) pathway has been shown to be active towards various types of cancer, including melanoma and lung cancer. MPDL3280A, an anti‑PD-L1 antibody, has shown clear clinical activity in PD-L1-overexpressing bladder cancer with an objective response rate of 40-50%, resulting in a breakthrough therapy designation granted by FDA. These events pronounce the importance of targeting the PD-L1 pathway in the treatment of bladder cancer. In the present study, we investigated the prognostic significance of the expression of three genes in the PD-L1 pathway, including PD-L1, B7.1 and PD-1, in three independent bladder cancer datasets in the Gene Expression Omnibus database. PD-L1, B7.1 and PD-1 were significantly associated with clinicopathological parameters indicative of a more aggressive phenotype of bladder cancer, such as a more advanced stage and a higher tumor grade. In addition, a high level expression of PD-L1 was associated with reduced patient survival. Of note, the combination of PD-L1 and B7.1 expression, but not other combinations of the three genes, were also able to predict patient survival. Our findings support the development of anti-PD-L1, which blocks PD-L1-PD-1 and B7.1-PD-L1 interactions, in treatment of bladder cancer. The observations were consistent in the three independent bladder cancer datasets consisting of a total of 695 human bladder specimens. The datasets were then assessed and it was found that the expression levels of the chemokine CC-motif ligand (CCL), CCL3, CCL8 and CCL18, were correlated with the PD-L1 expression level, while ADAMTS13 was differentially expressed in patients with a different survival status (alive or deceased). Additional investigations are required to elucidate the role of these genes in the PD-L1-mediated immune system suppression and bladder cancer progression. In conclusion, findings of this study suggested that PD-L1 is an important prognostic marker and a therapeutic target for bladder cancer.
Resumo:
Targeting angiogenesis through inhibition of the vascular endothelial growth factor (VEGF) pathway has been successful in the treatment of late stage colorectal cancer. However, not all patients benefit from inhibition of VEGF. Ras status is a powerful biomarker for response to anti-epidermal growth factor receptor therapy; however, an appropriate biomarker for response to anti-VEGF therapy is yet to be identified. VEGF and its receptors, FLT1 and KDR, play a crucial role in colon cancer progression; individually, these factors have been shown to be prognostic in colon cancer; however, expression of none of these factors alone was predictive of tumor response to anti-VEGF therapy. In the present study, we analyzed the expression levels of VEGFA, FLT1, and KDR in two independent colon cancer datasets and found that high expression levels of all three factors afforded a very poor prognosis. The observation was further confirmed in another independent colon cancer dataset, wherein high levels of expression of this three-gene signature was predictive of poor prognosis in patients with proficient mismatch repair a wild-type KRas status, or mutant p53 status. Most importantly, this signature also predicted tumor response to bevacizumab, an antibody targeting VEGFA, in a cohort of bevacizumab-treated patients. Since bevacizumab has been proven to be an important drug in the treatment of advanced stage colon cancer, our results suggest that the three-gene signature approach is valuable in terms of its prognostic value, and that it should be further evaluated in a prospective clinical trial to investigate its predictive value to anti-VEGF treatment.
Resumo:
Modern cancer research on prognostic and predictive biomarkers demands the integration of established and emerging high-throughput technologies. However, these data are meaningless unless carefully integrated with patient clinical outcome and epidemiological information. Integrated datasets hold the key to discovering new biomarkers and therapeutic targets in cancer. We have developed a novel approach and set of methods for integrating and interrogating phenomic, genomic and clinical data sets to facilitate cancer biomarker discovery and patient stratification. Applied to a known paradigm, the biological and clinical relevance of TP53, PICan was able to recapitulate the known biomarker status and prognostic significance at a DNA, RNA and protein levels.
Resumo:
In this paper we explore ways to address the issue of dataset bias in person re-identification by using data augmentation to increase the variability of the available datasets, and we introduce a novel data augmentation method for re-identification based on changing the image background. We show that use of data augmentation can improve the cross-dataset generalisation of convolutional network based re-identification systems, and that changing the image background yields further improvements.
Resumo:
A brief, historical overview of 10 apparently different, although in some cases, upon inspection, closely related, popular proposed reaction mechanisms and their associated rate equations, is given and in which the rate expression for each mechanism is derived from basic principles, Appendix A. In Appendix B, each of the 5 main mechanisms are tested using datasets, comprising initial reaction rate vs. organic pollutant concentration, [P] and incident irradiance, ρ, data, reported previously for TiO2, where P is phenol, 4-chlorophenol and formic acid. The best of those tested, in terms of overall fit, simplicity, usefulness and versatility is the disrupted adsorption kinetic model proposed by Ollis. The usual basic assumptions made in constructing these mechanisms are reported and the main underlying concerns explored.
Resumo:
Slow release drugs must be manufactured to meet target specifications with respect to dissolution curve profiles. In this paper we consider the problem of identifying the drivers of dissolution curve variability of a drug from historical manufacturing data. Several data sources are considered: raw material parameters, coating data, loss on drying and pellet size statistics. The methodology employed is to develop predictive models using LASSO, a powerful machine learning algorithm for regression with high-dimensional datasets. LASSO provides sparse solutions facilitating the identification of the most important causes of variability in the drug fabrication process. The proposed methodology is illustrated using manufacturing data for a slow release drug.
Resumo:
Objectives: To determine whether adjusting the denominator of the common hospital antibiotic use measurement unit (defined daily doses/100 bed-days) by including age-adjusted comorbidity score (100 bed-days/age-adjusted comorbidity score) would result in more accurate and meaningful assessment of hospital antibiotic use.
Methods: The association between the monthly sum of age-adjusted comorbidity and monthly antibiotic use was measured using time-series analysis (January 2008 to June 2012). For the purposes of conducting internal benchmarking, two antibiotic usage datasets were constructed, i.e. 2004-07 (first study period) and 2008-11 (second study period). Monthly antibiotic use was normalized per 100 bed-days and per 100 bed-days/age-adjusted comorbidity score.
Results: Results showed that antibiotic use had significant positive relationships with the sum of age-adjusted comorbidity score (P = 0.0004). The results also showed that there was a negative relationship between antibiotic use and (i) alcohol-based hand rub use (P = 0.0370) and (ii) clinical pharmacist activity (P = 0.0031). Normalizing antibiotic use per 100 bed-days contributed to a comparative usage rate of 1.31, i.e. the average antibiotic use during the second period was 31% higher than during the first period. However, normalizing antibiotic use per 100 bed-days per age-adjusted comorbidity score resulted in a comparative usage rate of 0.98, i.e. the average antibiotic use was 2% lower in the second study period. Importantly, the latter comparative usage rate is independent of differences in patient density and case mix characteristics between the two studied populations.
Conclusions: The proposed modified antibiotic measure provides an innovative approach to compare variations in antibiotic prescribing while taking account of patient case mix effects.
Resumo:
Drastic biodiversity declines have raised concerns about the deterioration of ecosystem functions and have motivated much recent research on the relationship between species diversity and ecosystem functioning. A functional trait framework has been proposed to improve the mechanistic understanding of this relationship, but this has rarely been tested for organisms other than plants. We analysed eight datasets, including five animal groups, to examine how well a trait-based approach, compared with a more traditional taxonomic approach, predicts seven ecosystem functions below- and above-ground. Trait-based indices consistently provided greater explanatory power than species richness or abundance. The frequency distributions of single or multiple traits in the community were the best predictors of ecosystem functioning. This implies that the ecosystem functions we investigated were underpinned by the combination of trait identities (i.e. single-trait indices) and trait complementarity (i.e. multi-trait indices) in the communities. Our study provides new insights into the general mechanisms that link biodiversity to ecosystem functioning in natural animal communities and suggests that the observed responses were due to the identity and dominance patterns of the trait composition rather than the number or abundance of species per se.
Resumo:
Many governments world-wide are promoting longer working life due to the social and economic repercussions of demographic change. However, not all workers are equally able to extend their employment careers. Thus, while national policies raise the overall level of labour market participation, they might create new social and labour market inequalities. This paper explores how institutional differences in the United Kingdom, Germany and Japan affect individual retirement decisions on the aggregate level, and variations in individuals’ degree of choice within and across countries. We investigate which groups of workers are disproportionately at risk of being ‘pushed’ out of employment, and how such inequalities have changed over time. We use comparable national longitudinal survey datasets focusing on the older population in England, Germany and Japan. Results point to cross-national differences in retirement transitions. Retirement transitions in Germany have occurred at an earlier age than in England and Japan. In Japan, the incidence of involuntary retirement is the lowest, reflecting an institutional context prescribing that employers provide employment until pension age, while Germany and England display substantial proportions of involuntary exits triggered by organisational-level redundancies, persistent early retirement plans or individual ill-health.
Resumo:
As an important type of spatial keyword query, the m-closest keywords (mCK) query finds a group of objects such that they cover all query keywords and have the smallest diameter, which is defined as the largest distance between any pair of objects in the group. The query is useful in many applications such as detecting locations of web resources. However, the existing work does not study the intractability of this problem and only provides exact algorithms, which are computationally expensive.
In this paper, we prove that the problem of answering mCK queries is NP-hard. We first devise a greedy algorithm that has an approximation ratio of 2. Then, we observe that an mCK query can be approximately answered by finding the circle with the smallest diameter that encloses a group of objects together covering all query keywords. We prove that the group enclosed in the circle can answer the mCK query with an approximation ratio of 2 over 3. Based on this, we develop an algorithm for finding such a circle exactly, which has a high time complexity. To improve efficiency, we propose another two algorithms that find such a circle approximately, with a ratio of 2 over √3 + ε. Finally, we propose an exact algorithm that utilizes the group found by the 2 over √3 + ε)-approximation algorithm to obtain the optimal group. We conduct extensive experiments using real-life datasets. The experimental results offer insights into both efficiency and accuracy of the proposed approximation algorithms, and the results also demonstrate that our exact algorithm outperforms the best known algorithm by an order of magnitude.
Resumo:
Massive amount of data that are geo-tagged and associated with text information are being generated at an unprecedented scale. These geo-textual data cover a wide range of topics. Users are interested in receiving up-to-date tweets such that their locations are close to a user specified location and their texts are interesting to users. For example, a user may want to be updated with tweets near her home on the topic “food poisoning vomiting.” We consider the Temporal Spatial-Keyword Top-k Subscription (TaSK) query. Given a TaSK query, we continuously maintain up-to-date top-k most relevant results over a stream of geo-textual objects (e.g., geo-tagged Tweets) for the query. The TaSK query takes into account text relevance, spatial proximity, and recency of geo-textual objects in evaluating its relevance with a geo-textual object. We propose a novel solution to efficiently process a large number of TaSK queries over a stream of geotextual objects. We evaluate the efficiency of our approach on two real-world datasets and the experimental results show that our solution is able to achieve a reduction of the processing time by 70-80% compared with two baselines.
Resumo:
The preferences of users are important in route search and planning. For example, when a user plans a trip within a city, their preferences can be expressed as keywords shopping mall, restaurant, and museum, with weights 0.5, 0.4, and 0.1, respectively. The resulting route should best satisfy their weighted preferences. In this paper, we take into account the weighted user preferences in route search, and present a keyword coverage problem, which finds an optimal route from a source location to a target location such that the keyword coverage is optimized and that the budget score satisfies a specified constraint. We prove that this problem is NP-hard. To solve this complex problem, we pro- pose an optimal route search based on an A* variant for which we have defined an admissible heuristic function. The experiments conducted on real-world datasets demonstrate both the efficiency and accu- racy of our proposed algorithms.
Resumo:
Although visual surveillance has emerged as an effective technolody for public security, privacy has become an issue of great concern in the transmission and distribution of surveillance videos. For example, personal facial images should not be browsed without permission. To cope with this issue, face image scrambling has emerged as a simple solution for privacyrelated applications. Consequently, online facial biometric verification needs to be carried out in the scrambled domain thus bringing a new challenge to face classification. In this paper, we investigate face verification issues in the scrambled domain and propose a novel scheme to handle this challenge. In our proposed method, to make feature extraction from scrambled face images robust, a biased random subspace sampling scheme is applied to construct fuzzy decision trees from randomly selected features, and fuzzy forest decision using fuzzy memberships is then obtained from combining all fuzzy tree decisions. In our experiment, we first estimated the optimal parameters for the construction of the random forest, and then applied the optimized model to the benchmark tests using three publically available face datasets. The experimental results validated that our proposed scheme can robustly cope with the challenging tests in the scrambled domain, and achieved an improved accuracy over all tests, making our method a promising candidate for the emerging privacy-related facial biometric applications.
Resumo:
The increasing adoption of cloud computing, social networking, mobile and big data technologies provide challenges and opportunities for both research and practice. Researchers face a deluge of data generated by social network platforms which is further exacerbated by the co-mingling of social network platforms and the emerging Internet of Everything. While the topicality of big data and social media increases, there is a lack of conceptual tools in the literature to help researchers approach, structure and codify knowledge from social media big data in diverse subject matter domains, many of whom are from nontechnical disciplines. Researchers do not have a general-purpose scaffold to make sense of the data and the complex web of relationships between entities, social networks, social platforms and other third party databases, systems and objects. This is further complicated when spatio-temporal data is introduced. Based on practical experience of working with social media datasets and existing literature, we propose a general research framework for social media research using big data. Such a framework assists researchers in placing their contributions in an overall context, focusing their research efforts and building the body of knowledge in a given discipline area using social media data in a consistent and coherent manner.
Resumo:
One of the major challenges in systems biology is to understand the complex responses of a biological system to external perturbations or internal signalling depending on its biological conditions. Genome-wide transcriptomic profiling of cellular systems under various chemical perturbations allows the manifestation of certain features of the chemicals through their transcriptomic expression profiles. The insights obtained may help to establish the connections between human diseases, associated genes and therapeutic drugs. The main objective of this study was to systematically analyse cellular gene expression data under various drug treatments to elucidate drug-feature specific transcriptomic signatures. We first extracted drug-related information (drug features) from the collected textual description of DrugBank entries using text-mining techniques. A novel statistical method employing orthogonal least square learning was proposed to obtain drug-feature-specific signatures by integrating gene expression with DrugBank data. To obtain robust signatures from noisy input datasets, a stringent ensemble approach was applied with the combination of three techniques: resampling, leave-one-out cross validation, and aggregation. The validation experiments showed that the proposed method has the capacity of extracting biologically meaningful drug-feature-specific gene expression signatures. It was also shown that most of signature genes are connected with common hub genes by regulatory network analysis. The common hub genes were further shown to be related to general drug metabolism by Gene Ontology analysis. Each set of genes has relatively few interactions with other sets, indicating the modular nature of each signature and its drug-feature-specificity. Based on Gene Ontology analysis, we also found that each set of drug feature (DF)-specific genes were indeed enriched in biological processes related to the drug feature. The results of these experiments demonstrated the pot- ntial of the method for predicting certain features of new drugs using their transcriptomic profiles, providing a useful methodological framework and a valuable resource for drug development and characterization.