846 resultados para resistant data


Relevância:

20.00% 20.00%

Publicador:

Resumo:

This report is an update of an earlier one produced in January 2010 (see Carrington et al. 2010) which remains as an ePrint through the project’s home page. The report considers extant data which have been sourced with respect to some of the consequences of violent acts, incidents, harms and risky behaviour involving males living in regional and remote Australia and which were available in public data bases at production.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Various time-memory tradeoffs attacks for stream ciphers have been proposed over the years. However, the claimed success of these attacks assumes the initialisation process of the stream cipher is one-to-one. Some stream cipher proposals do not have a one-to-one initialisation process. In this paper, we examine the impact of this on the success of time-memory-data tradeoff attacks. Under the circumstances, some attacks are more successful than previously claimed while others are less. The conditions for both cases are established.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

With the increasing number of XML documents in varied domains, it has become essential to identify ways of finding interesting information from these documents. Data mining techniques were used to derive this interesting information. Mining on XML documents is impacted by its model due to the semi-structured nature of these documents. Hence, in this chapter we present an overview of the various models of XML documents, how these models were used for mining and some of the issues and challenges in these models. In addition, this chapter also provides some insights into the future models of XML documents for effectively capturing the two important features namely structure and content of XML documents for mining.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This special issue of the Journal of Urban Technology brings together five articles that are based on presentations given at the Street Computing workshop held on 24 November 2009 in Melbourne in conjunction with the Australian Computer-Human Interaction conference (OZCHI 2009). Our own article introduces the Street Computing vision and explores the potential, challenges and foundations of this research vision. In order to do so, we first look at the currently available sources of information and discuss their link to existing research efforts. Section 2 then introduces the notion of Street Computing and our research approach in more detail. Section 3 looks beyond the core concept itself and summarises related work in this field of interest.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we present a sequential Monte Carlo algorithm for Bayesian sequential experimental design applied to generalised non-linear models for discrete data. The approach is computationally convenient in that the information of newly observed data can be incorporated through a simple re-weighting step. We also consider a flexible parametric model for the stimulus-response relationship together with a newly developed hybrid design utility that can produce more robust estimates of the target stimulus in the presence of substantial model and parameter uncertainty. The algorithm is applied to hypothetical clinical trial or bioassay scenarios. In the discussion, potential generalisations of the algorithm are suggested to possibly extend its applicability to a wide variety of scenarios

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Several authors stress the importance of data’s crucial foundation for operational, tactical and strategic decisions (e.g., Redman 1998, Tee et al. 2007). Data provides the basis for decision making as data collection and processing is typically associated with reducing uncertainty in order to make more effective decisions (Daft and Lengel 1986). While the first series of investments of Information Systems/Information Technology (IS/IT) into organizations improved data collection, restricted computational capacity and limited processing power created challenges (Simon 1960). Fifty years on, capacity and processing problems are increasingly less relevant; in fact, the opposite exists. Determining data relevance and usefulness is complicated by increased data capture and storage capacity, as well as continual improvements in information processing capability. As the IT landscape changes, businesses are inundated with ever-increasing volumes of data from both internal and external sources available on both an ad-hoc and real-time basis. More data, however, does not necessarily translate into more effective and efficient organizations, nor does it increase the likelihood of better or timelier decisions. This raises questions about what data managers require to assist their decision making processes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Mixture models are a flexible tool for unsupervised clustering that have found popularity in a vast array of research areas. In studies of medicine, the use of mixtures holds the potential to greatly enhance our understanding of patient responses through the identification of clinically meaningful clusters that, given the complexity of many data sources, may otherwise by intangible. Furthermore, when developed in the Bayesian framework, mixture models provide a natural means for capturing and propagating uncertainty in different aspects of a clustering solution, arguably resulting in richer analyses of the population under study. This thesis aims to investigate the use of Bayesian mixture models in analysing varied and detailed sources of patient information collected in the study of complex disease. The first aim of this thesis is to showcase the flexibility of mixture models in modelling markedly different types of data. In particular, we examine three common variants on the mixture model, namely, finite mixtures, Dirichlet Process mixtures and hidden Markov models. Beyond the development and application of these models to different sources of data, this thesis also focuses on modelling different aspects relating to uncertainty in clustering. Examples of clustering uncertainty considered are uncertainty in a patient’s true cluster membership and accounting for uncertainty in the true number of clusters present. Finally, this thesis aims to address and propose solutions to the task of comparing clustering solutions, whether this be comparing patients or observations assigned to different subgroups or comparing clustering solutions over multiple datasets. To address these aims, we consider a case study in Parkinson’s disease (PD), a complex and commonly diagnosed neurodegenerative disorder. In particular, two commonly collected sources of patient information are considered. The first source of data are on symptoms associated with PD, recorded using the Unified Parkinson’s Disease Rating Scale (UPDRS) and constitutes the first half of this thesis. The second half of this thesis is dedicated to the analysis of microelectrode recordings collected during Deep Brain Stimulation (DBS), a popular palliative treatment for advanced PD. Analysis of this second source of data centers on the problems of unsupervised detection and sorting of action potentials or "spikes" in recordings of multiple cell activity, providing valuable information on real time neural activity in the brain.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the last few years we have observed a proliferation of approaches for clustering XML docu- ments and schemas based on their structure and content. The presence of such a huge amount of approaches is due to the different applications requiring the XML data to be clustered. These applications need data in the form of similar contents, tags, paths, structures and semantics. In this paper, we first outline the application contexts in which clustering is useful, then we survey approaches so far proposed relying on the abstract representation of data (instances or schema), on the identified similarity measure, and on the clustering algorithm. This presentation leads to draw a taxonomy in which the current approaches can be classified and compared. We aim at introducing an integrated view that is useful when comparing XML data clustering approaches, when developing a new clustering algorithm, and when implementing an XML clustering compo- nent. Finally, the paper moves into the description of future trends and research issues that still need to be faced.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper argues for a renewed focus on statistical reasoning in the beginning school years, with opportunities for children to engage in data modelling. Results are reported from the first year of a 3-year longitudinal study in which three classes of first-grade children (6-year-olds) and their teachers engaged in data modelling activities. The theme of Looking after our Environment, part of the children’s science curriculum, provided the task context. The goals for the two activities addressed here included engaging children in core components of data modelling, namely, selecting attributes, structuring and representing data, identifying variation in data, and making predictions from given data. Results include the various ways in which children represented and re represented collected data, including attribute selection, and the metarepresentational competence they displayed in doing so. The “data lenses” through which the children dealt with informal inference (variation and prediction) are also reported.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In response to the need to leverage private finance and the lack of competition in some parts of the Australian public sector infrastructure market, especially in the very large economic infrastructure sector procured using Pubic Private Partnerships, the Australian Federal government has demonstrated its desire to attract new sources of in-bound foreign direct investment (FDI). This paper aims to report on progress towards an investigation into the determinants of multinational contractors’ willingness to bid for Australian public sector major infrastructure projects. This research deploys Dunning’s eclectic theory for the first time in terms of in-bound FDI by multinational contractors into Australia. Elsewhere, the authors have developed Dunning’s principal hypothesis to suit the context of this research and to address a weakness arising in this hypothesis that is based on a nominal approach to the factors in Dunning's eclectic framework and which fails to speak to the relative explanatory power of these factors. In this paper, a first stage test of the authors' development of Dunning's hypothesis is presented by way of an initial review of secondary data vis-à-vis the selected sector (roads and bridges) in Australia (as the host location) and with respect to four selected home countries (China; Japan; Spain; and US). In doing so, the next stage in the research method concerning sampling and case studies is also further developed and described in this paper. In conclusion, the extent to which the initial review of secondary data suggests the relative importance of the factors in the eclectic framework is considered. It is noted that more robust conclusions are expected following the future planned stages of the research including primary data from the case studies and a global survey of the world’s largest contractors and which is briefly previewed. Finally, and beyond theoretical contributions expected from the overall approach taken to developing and testing Dunning’s framework, other expected contributions concerning research method and practical implications are mentioned.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A rule-based approach for classifying previously identified medical concepts in the clinical free text into an assertion category is presented. There are six different categories of assertions for the task: Present, Absent, Possible, Conditional, Hypothetical and Not associated with the patient. The assertion classification algorithms were largely based on extending the popular NegEx and Context algorithms. In addition, a health based clinical terminology called SNOMED CT and other publicly available dictionaries were used to classify assertions, which did not fit the NegEx/Context model. The data for this task includes discharge summaries from Partners HealthCare and from Beth Israel Deaconess Medical Centre, as well as discharge summaries and progress notes from University of Pittsburgh Medical Centre. The set consists of 349 discharge reports, each with pairs of ground truth concept and assertion files for system development, and 477 reports for evaluation. The system’s performance on the evaluation data set was 0.83, 0.83 and 0.83 for recall, precision and F1-measure, respectively. Although the rule-based system shows promise, further improvements can be made by incorporating machine learning approaches.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Projects funded by the Australian National Data Service(ANDS). The specific projects that were funded included: a) Greenhouse Gas Emissions Project (N2O) with Prof. Peter Grace from QUT’s Institute of Sustainable Resources. b) Q150 Project for the management of multimedia data collected at Festival events with Prof. Phil Graham from QUT’s Institute of Creative Industries. c) Bio-diversity environmental sensing with Prof. Paul Roe from the QUT Microsoft eResearch Centre. For the purposes of these projects the Eclipse Rich Client Platform (Eclipse RCP) was chosen as an appropriate software development framework within which to develop the respective software. This poster will present a brief overview of the requirements of the projects, an overview of the experiences of the project team in using Eclipse RCP, report on the advantages and disadvantages of using Eclipse and it’s perspective on Eclipse as an integrated tool for supporting future data management requirements.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Prostate cancer (CaP) is the most commonly diagnosed cancer in males in Australia, North America, and Europe. If found early and locally confined, CaP can be treated with radical prostatectomy or radiation therapy; however, 25-40% patients will relapse and go on to advanced disease. The most common therapy in these cases is androgen deprivation therapy (ADT), which suppresses androgen production from the testis. Lack of the testicular androgen supply causes cells of the prostate to undergo apoptosis. However, in some cases the regression initially seen with ADT eventually gives way to a growth of a population of cancerous cells that no longer require testicular androgens. This phenotype is essentially fatal and is termed castrate resistant prostate cancer (CRPC). In addition to eventual regression, there are many undesirable side effects which accompany ADT, including development of a metabolic syndrome, which is defined by the U.S. National Library of Medicine as “a combination of medical disorders that increase the risk of developing cardiovascular disease and diabetes.” This project will focus on the effect of ADT induced hyperinsulinemia, as mimicked by treating androgen receptor positive CaP cells with insulin in a serum (hormone) deprived environment. While this side effect is not widely explored, in this thesis it is demonstrated for the first time that insulin upregulates pathways important to CaP progression. Our group has previously shown that during CaP progression, the enzymes necessary for de novo steroidogenesis are upregulated in the LNCaP xenograft model, total steroid levels are increased in tumours compared to pre castrate levels, and de novo steroidogenesis from radio-labelled acetate has been demonstrated. Because of the CaP dependence on AR for survival, we and other groups believe that CaP cells carry out de novo steroidogenesis to survive in androgen deprived conditions. Because (a) men on ADT often develop metabolic syndrome, and (b) men with lifestyle-induced obesity and hyperinsulinemia have worse prognosis and faster disease progression, and because (c) insulin causes steroidogenesis in other cell lines, the hypothesis that insulin may contribute to CaP progression through upregulation of steroidogenesis was explored. Insulin upregulates steroidogenesis enzymes at the mRNA level in three AR positive cell lines, as well as upregulating these enzymes at the protein level in two cell lines. It has also been demonstrated that insulin increases mitochondrial (functional) levels of steroid acute regulatory protein (StAR). Furthermore, insulin causes increased levels of total steroids in and induction of de novo steroid synthesis by insulin has been demonstrated at levels induced sufficient to activate AR. The effect of insulin analogs on CaP steroidogenesis in LNCaP and VCaP cells has also been investigated because epidemiological studies suggest that some of the analogs developed may have more cancer stimulatory effects than normal insulin. In this project, despite the signalling differences between glargine, X10, and insulin, these analogs did not appear to induce steroidogenesis any more potently that normal insulin. The effect of insulin of MCF7breast cancer cells was also investigated with results suggesting that breast cancer cells may be capable of de novo steroidogenesis, and that increase in estradiol production may be exacerbated by insulin. Insulin has also been long known to stimulate lipogenesis in the liver and adipocytes, and has been demonstrated to increase lipogenesis in breast cancer cells; therefore, investigation of the effect of insulin on lipogenesis, which is a hallmark of aggressive cancers, was investigated. In CaP progression sterol regulatory element binding protein (SREBP) is dysregulated and upregulates fatty acid synthase (FASN), acetyl CoA-carboxylase, and other lipogenesis genes. SREBP is important for steroidogenesis and in this project has been shown to be upregulated by insulin in CaP cells. Fatty acid synthesis provides building blocks of membrane growth, provides substrates for acid oxidation, the main energy source for CaP cells, provides building blocks for anti-apoptotic and proinflammatory molecules, and provides molecules that stimulate steroidogenesis. In this project it has been shown that insulin upregulates FASN and ACC, which synthesize fatty acids, as well as upregulating hormone sensitive lipase (HSL), diazepam-binding inhibitor (DBI), and long-chain acyl-CoA synthetase 3 (ACSL3), which contribute to lipid activation of steroidogenesis. Insulin also upregulates total lipid levels and de novo lipogenesis, which can be suppressed by inhibition of the insulin receptor (INSR). The fatty acids synthesized after insulin treatment are those that have been associated with CaP; furthermore, microarray data suggests insulin may upregulate fatty acid biosynthesis, metabolism and arachidonic acid metabolism pathways, which have been implicated in CaP growth and survival. Pharmacological agents used to treat patients with hyperinsulinemia/ hyperlipidemia have gained much interest in regards to CaP risk and treatment; however, the scientific rationale behind these clinical applications has not been examined. This thesis explores whether the use of metformin or simvastatin would decrease either lipogenesis or steroidogenesis or both in CaP cells. Simvastatin is a 3-hydroxy-3-methylglutaryl-CoA reductase (HMGR) inhibitor, which blocks synthesis of cholesterol, the building block of steroids/ androgens. It has also been postulated to down regulate SREBP in other metabolic disorders. It has been shown in this thesis, in LNCaP cells, that simvastatin inhibited and decreased insulin induced steroidogenesis and lipogenesis, respectively, but increased these pathways in the absence of insulin. Conversely, metformin, which activates AMP-activated protein kinase (AMPK) to shut down lipogenesis, cholesterol synthesis, and protein synthesis, highly suppresses both steroidogenesis and lipogenesis in the presence and absence of insulin. Lastly, because it has been demonstrated to increase steroidogenesis in other cell lines, and because the elucidation of any factors affecting steroidogenesis is important to understanding CaP, the effect of IGF2 on steroidogenesis in CaP cells was investigated. In patient samples, as men progress to CRPC, IGF2 mRNA and the protein levels of the receptors it may signal through are upregulated. It has also been demonstrated that IGF2 upregulates steroidogenic enzymes at both the mRNA and protein levels in LNCaP cells, increases intracellular and secreted steroid/androgen levels in LNCaPs to levels sufficient to stimulate the AR, and upregulated de novo steroidogenesis in LNCaPs and VCaPs. As well, inhibition of INSR and insulin-like growth factor 1 receptor (IGF1R), which IGF2 signals through, suggests that induction of steroidogenesis may be occurring predominantly through IGF1R. In summary, this project has illuminated for the first time that insulin is likely to play a large role in cancer progression, through upregulation of the steroidogenesis and lipogenesis pathways at the mRNA and protein levels, and production levels, and demonstrates a novel role for IGF-II in CaP progression through stimulation of steroidogenesis. It has also been demonstrated that metformin and simvastatin drugs may be useful in suppressing the insulin induction of these pathways. This project affirms the pathways by which ADT- induced metabolic syndrome may exacerbate CaP progression and strongly suggests that the monitoring and modulation of the metabolic state of CaP patients could have a strong impact on their therapeutic outcomes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

It is a big challenge to acquire correct user profiles for personalized text classification since users may be unsure in providing their interests. Traditional approaches to user profiling adopt machine learning (ML) to automatically discover classification knowledge from explicit user feedback in describing personal interests. However, the accuracy of ML-based methods cannot be significantly improved in many cases due to the term independence assumption and uncertainties associated with them. This paper presents a novel relevance feedback approach for personalized text classification. It basically applies data mining to discover knowledge from relevant and non-relevant text and constraints specific knowledge by reasoning rules to eliminate some conflicting information. We also developed a Dempster-Shafer (DS) approach as the means to utilise the specific knowledge to build high-quality data models for classification. The experimental results conducted on Reuters Corpus Volume 1 and TREC topics support that the proposed technique achieves encouraging performance in comparing with the state-of-the-art relevance feedback models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper studies the missing covariate problem which is often encountered in survival analysis. Three covariate imputation methods are employed in the study, and the effectiveness of each method is evaluated within the hazard prediction framework. Data from a typical engineering asset is used in the case study. Covariate values in some time steps are deliberately discarded to generate an incomplete covariate set. It is found that although the mean imputation method is simpler than others for solving missing covariate problems, the results calculated by it can differ largely from the real values of the missing covariates. This study also shows that in general, results obtained from the regression method are more accurate than those of the mean imputation method but at the cost of a higher computational expensive. Gaussian Mixture Model (GMM) method is found to be the most effective method within these three in terms of both computation efficiency and predication accuracy.