117 resultados para 650200 Mining and Extraction


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This chapter discusses reference modelling languages for business systems analysis and design. In particular, it reports on reference models in the context of the design-for/by-reuse paradigm, explains how traditional modelling techniques fail to provide adequate conceptual expressiveness to allow for easy model reuse by configuration or adaptation and elaborates on the need for reference modelling languages to be configurable. We discuss requirements for and the development of reference modelling languages that reflect the need for configurability. Exemplarily, we report on the development, definition and configuration of configurable event-driven process chains. We further outline how configurable reference modelling languages and the corresponding design principles can be used in future scenarios such as process mining and data modelling.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Providing precise positioning services in regional areas to support agriculture, mining, and construction sectors depends on the availability of ground continuously operating GNSS reference stations and communications linking these stations to central computers and users. With the support of CRC for Spatial Information, a more comprehensive review has been completed recently to examine various wired and wireless communication links available for precise positioning services, in particular in the Queensland regional areas. The study covers a wide range of communication technologies that are currently available, including fixed, mobile wireless, and Geo-stationary and or low earth orbiting satellites. These technologies are compared in terms of bandwidth, typical latency, reliability, coverage, and costs. Additionally, some tests were also conducted to determine the performances of different systems in the real environment. Finally, based on user application requirements, the paper discusses the suitability of different communication links.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Guardian reportage of the United Kingdom Member of Parliament (MP) expenses scandal of 2009 used crowdsourcing and computational journalism techniques. Computational journalism can be broadly defined as the application of computer science techniques to the activities of journalism. Its foundation lies in computer assisted reporting techniques and its importance is increasing due to the: (a) increasing availability of large scale government datasets for scrutiny; (b) declining cost, increasing power and ease of use of data mining and filtering software; and Web 2.0; and (c) explosion of online public engagement and opinion.. This paper provides a case study of the Guardian MP expenses scandal reportage and reveals some key challenges and opportunities for digital journalism. It finds journalists may increasingly take an active role in understanding, interpreting, verifying and reporting clues or conclusions that arise from the interrogations of datasets (computational journalism). Secondly a distinction should be made between information reportage and computational journalism in the digital realm, just as a distinction might be made between citizen reporting and citizen journalism. Thirdly, an opportunity exists for online news providers to take a ‘curatorial’ role, selecting and making easily available the best data sources for readers to use (information reportage). These activities have always been fundamental to journalism, however the way in which they are undertaken may change. Findings from this paper may suggest opportunities and challenges for the implementation of computational journalism techniques in practice by digital Australian media providers, and further areas of research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Guardian reportage of the United Kingdom Member of Parliament (MP) expenses scandal of 2009 used crowdsourcing and computational journalism techniques. Computational journalism can be broadly defined as the application of computer science techniques to the activities of journalism. Its foundation lies in computer assisted reporting techniques and its importance is increasing due to the: (a) increasing availability of large scale government datasets for scrutiny; (b) declining cost, increasing power and ease of use of data mining and filtering software; and Web 2.0; and (c) explosion of online public engagement and opinion.. This paper provides a case study of the Guardian MP expenses scandal reportage and reveals some key challenges and opportunities for digital journalism. It finds journalists may increasingly take an active role in understanding, interpreting, verifying and reporting clues or conclusions that arise from the interrogations of datasets (computational journalism). Secondly a distinction should be made between information reportage and computational journalism in the digital realm, just as a distinction might be made between citizen reporting and citizen journalism. Thirdly, an opportunity exists for online news providers to take a ‘curatorial’ role, selecting and making easily available the best data sources for readers to use (information reportage). These activities have always been fundamental to journalism, however the way in which they are undertaken may change. Findings from this paper may suggest opportunities and challenges for the implementation of computational journalism techniques in practice by digital Australian media providers, and further areas of research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

With regard to the long-standing problem of the semantic gap between low-level image features and high-level human knowledge, the image retrieval community has recently shifted its emphasis from low-level features analysis to high-level image semantics extrac- tion. User studies reveal that users tend to seek information using high-level semantics. Therefore, image semantics extraction is of great importance to content-based image retrieval because it allows the users to freely express what images they want. Semantic content annotation is the basis for semantic content retrieval. The aim of image anno- tation is to automatically obtain keywords that can be used to represent the content of images. The major research challenges in image semantic annotation are: what is the basic unit of semantic representation? how can the semantic unit be linked to high-level image knowledge? how can the contextual information be stored and utilized for image annotation? In this thesis, the Semantic Web technology (i.e. ontology) is introduced to the image semantic annotation problem. Semantic Web, the next generation web, aims at mak- ing the content of whatever type of media not only understandable to humans but also to machines. Due to the large amounts of multimedia data prevalent on the Web, re- searchers and industries are beginning to pay more attention to the Multimedia Semantic Web. The Semantic Web technology provides a new opportunity for multimedia-based applications, but the research in this area is still in its infancy. Whether ontology can be used to improve image annotation and how to best use ontology in semantic repre- sentation and extraction is still a worth-while investigation. This thesis deals with the problem of image semantic annotation using ontology and machine learning techniques in four phases as below. 1) Salient object extraction. A salient object servers as the basic unit in image semantic extraction as it captures the common visual property of the objects. Image segmen- tation is often used as the �rst step for detecting salient objects, but most segmenta- tion algorithms often fail to generate meaningful regions due to over-segmentation and under-segmentation. We develop a new salient object detection algorithm by combining multiple homogeneity criteria in a region merging framework. 2) Ontology construction. Since real-world objects tend to exist in a context within their environment, contextual information has been increasingly used for improving object recognition. In the ontology construction phase, visual-contextual ontologies are built from a large set of fully segmented and annotated images. The ontologies are composed of several types of concepts (i.e. mid-level and high-level concepts), and domain contextual knowledge. The visual-contextual ontologies stand as a user-friendly interface between low-level features and high-level concepts. 3) Image objects annotation. In this phase, each object is labelled with a mid-level concept in ontologies. First, a set of candidate labels are obtained by training Support Vectors Machines with features extracted from salient objects. After that, contextual knowledge contained in ontologies is used to obtain the �nal labels by removing the ambiguity concepts. 4) Scene semantic annotation. The scene semantic extraction phase is to get the scene type by using both mid-level concepts and domain contextual knowledge in ontologies. Domain contextual knowledge is used to create scene con�guration that describes which objects co-exist with which scene type more frequently. The scene con�guration is represented in a probabilistic graph model, and probabilistic inference is employed to calculate the scene type given an annotated image. To evaluate the proposed methods, a series of experiments have been conducted in a large set of fully annotated outdoor scene images. These include a subset of the Corel database, a subset of the LabelMe dataset, the evaluation dataset of localized semantics in images, the spatial context evaluation dataset, and the segmented and annotated IAPR TC-12 benchmark.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Following the completion of the draft Human Genome in 2001, genomic sequence data is becoming available at an accelerating rate, fueled by advances in sequencing and computational technology. Meanwhile, large collections of astronomical and geospatial data have allowed the creation of virtual observatories, accessible throughout the world and requiring only commodity hardware. Through a combination of advances in data management, data mining and visualization, this infrastructure enables the development of new scientific and educational applications as diverse as galaxy classification and real-time tracking of earthquakes and volcanic plumes. In the present paper, we describe steps taken along a similar path towards a virtual observatory for genomes – an immersive three-dimensional visual navigation and query system for comparative genomic data.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This review collects and summarises the biological applications of the element cobalt. Small amounts of the ferromagnetic metal can be found in rock, soil, plants and animals, but is mainly obtained as a by-product of nickel and copper mining, and is separated from the ores (mainly cobaltite, erythrite, glaucodot and skutterudite) using a variety of methods. Compounds of cobalt include several oxides, including: green cobalt(II) (CoO), blue cobalt(II,III) (Co3O4), and black cobalt(III) (Co2O3); four halides including pink cobalt(II) fluoride (CoF2), blue cobalt(II) chloride (CoCl2), green cobalt(II) bromide (CoBr2), and blue-black cobalt(II) iodide (CoI2). The main application of cobalt is in its metal form in cobalt-based super alloys, though other uses include lithium cobalt oxide batteries, chemical reaction catalyst, pigments and colouring, and radioisotopes in medicine. It is known to mimic hypoxia on the cellular level by stabilizing the α subunit of hypoxia inducing factor (HIF), when chemically applied as cobalt chloride (CoCl2). This is seen in many biological research applications, where it has shown to promote angiogenesis, erythropoiesis and anaerobic metabolism through the transcriptional activation of genes such as vascular endothelial growth factor (VEGF) and erythropoietin (EPO), contributing significantly to the pathophysiology of major categories of disease, such as myocardial, renal and cerebral ischaemia, high altitude related maladies and bone defects. As a necessary constituent for the formation of vitamin B12, it is essential to all animals, including humans, however excessive exposure can lead to tissue and cellular toxicity. Cobalt has been shown to provide promising potential in clinical applications, however further studies are necessary to clarify its role in hypoxia-responsive genes and the applications of cobalt-chloride treated tissues.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background: Heat-related mortality is a matter of great public health concern, especially in the light of climate change. Although many studies have found associations between high temperatures and mortality, more research is needed to project the future impacts of climate change on heat-related mortality. Objectives: We conducted a systematic review of research and methods for projecting future heat-related mortality under climate change scenarios. Data sources and extraction: A literature search was conducted in August 2010, using the electronic databases PubMed, Scopus, ScienceDirect, ProQuest, and Web of Science. The search was limited to peer-reviewed journal articles published in English up to 2010. Data synthesis: The review included 14 studies that fulfilled the inclusion criteria. Most projections showed that climate change would result in a substantial increase in heat-related mortality. Projecting heat-related mortality requires understanding of the historical temperature-mortality relationships, and consideration of the future changes in climate, population and acclimatization. Further research is needed to provide a stronger theoretical framework for projections, including a better understanding of socio-economic development, adaptation strategies, land-use patterns, air pollution and mortality displacement. Conclusions: Scenario-based projection research will meaningfully contribute to assessing and managing the potential impacts of climate change on heat-related mortality.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

It is a big challenge to guarantee the quality of discovered relevance features in text documents for describing user preferences because of the large number of terms, patterns, and noise. Most existing popular text mining and classification methods have adopted term-based approaches. However, they have all suffered from the problems of polysemy and synonymy. Over the years, people have often held the hypothesis that pattern-based methods should perform better than term- based ones in describing user preferences, but many experiments do not support this hypothesis. This research presents a promising method, Relevance Feature Discovery (RFD), for solving this challenging issue. It discovers both positive and negative patterns in text documents as high-level features in order to accurately weight low-level features (terms) based on their specificity and their distributions in the high-level features. The thesis also introduces an adaptive model (called ARFD) to enhance the exibility of using RFD in adaptive environment. ARFD automatically updates the system's knowledge based on a sliding window over new incoming feedback documents. It can efficiently decide which incoming documents can bring in new knowledge into the system. Substantial experiments using the proposed models on Reuters Corpus Volume 1 and TREC topics show that the proposed models significantly outperform both the state-of-the-art term-based methods underpinned by Okapi BM25, Rocchio or Support Vector Machine and other pattern-based methods.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In Australia, the spread and dominance of non-native plant species has been identified as a serious threat to rangeland biodiversity and ecosystem functioning. Rangelands extend over 70% of Australia’s land mass or more than 6 million km2. These rangelands consist of a diverse set of ecosystems including grasslands, shrub-lands, and woodlands spanning numerous climatic zones, ranging from arid to mesic. Because of the high economic, social, and environmental values, sustainable management of these vast landscapes is critical for Australia’s future. More than 2 million people live in these areas and major industries are ranching, mining, and tourism. In terms of biodiversity values, 53 of 85 of Australia’s biogeographical regions and 5 of 15 identified biodiversity hotspots are found in rangelands.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Assessing and prioritising cost-effective strategies to mitigate the impacts of traffic incidents and accidents on non-recurrent congestion on major roads represents a significant challenge for road network managers. This research examines the influence of numerous factors associated with incidents of various types on their duration. It presents a comprehensive traffic incident data mining and analysis by developing an incident duration model based on twelve months of incident data obtained from the Australian freeway network. Parametric accelerated failure time (AFT) survival models of incident duration were developed, including log-logistic, lognormal, and Weibul-considering both fixed and random parameters, as well as a Weibull model with gamma heterogeneity. The Weibull AFT models with random parameters were appropriate for modelling incident duration arising from crashes and hazards. A Weibull model with gamma heterogeneity was most suitable for modelling incident duration of stationary vehicles. Significant variables affecting incident duration include characteristics of the incidents (severity, type, towing requirements, etc.), and location, time of day, and traffic characteristics of the incident. Moreover, the findings reveal no significant effects of infrastructure and weather on incident duration. A significant and unique contribution of this paper is that the durations of each type of incident are uniquely different and respond to different factors. The results of this study are useful for traffic incident management agencies to implement strategies to reduce incident duration, leading to reduced congestion, secondary incidents, and the associated human and economic losses.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Topic modeling has been widely utilized in the fields of information retrieval, text mining, text classification etc. Most existing statistical topic modeling methods such as LDA and pLSA generate a term based representation to represent a topic by selecting single words from multinomial word distribution over this topic. There are two main shortcomings: firstly, popular or common words occur very often across different topics that bring ambiguity to understand topics; secondly, single words lack coherent semantic meaning to accurately represent topics. In order to overcome these problems, in this paper, we propose a two-stage model that combines text mining and pattern mining with statistical modeling to generate more discriminative and semantic rich topic representations. Experiments show that the optimized topic representations generated by the proposed methods outperform the typical statistical topic modeling method LDA in terms of accuracy and certainty.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Rural communities across Australia are increasingly being asked to shoulder the environmental and social impacts of intensive mining and gas projects. Escalating demand for coal seam gas (CSG) is raising significant environmental justice issues for rural communities. Chief amongst environmental concerns are risks of contamination or depletion of vital underground aquifers as well as treatment and disposal of high-saline water close to high quality agricultural soils. Associated infrastructure such as pipelines, electricity lines, gas processing and port facilities can also adversely affect communities and ecosystems great distances from where the gas is originally extracted. Whilst community submission (and appeal) rights do exist, accessing expert independent information is challenging, legal terminology is complex and submission periods are short, leading ultimately to a lack of procedural justice for landholders and their communities. Since August 2012, Queensland University of Technology (QUT) has worked in partnership with not-for-profit legal centre - Queensland’s Environmental Defenders Office (EDO) - to help better educate communities about mining and CSG assessment processes. The project, now entering its third semester, aims to empower communities to access relevant information and actively engage in legal processes on their own behalf. Students involved in the project so far have helped to research chapters of a comprehensive community guide to mining and CSG law as well as organising multidisciplinary community forums and preparing information on land access and compensation rights for landholders. While environmental justice issues still exist without significant law reform, the project has led to greater awareness amongst the community of the laws relating the CSG. At the same time, it has led to a greater understanding by students and academics of real life environmental justice issues currently faced by rural communities.