994 resultados para Mining industries
Resumo:
Road safety is a major concern worldwide. Road safety will improve as road conditions and their effects on crashes are continually investigated. This paper proposes to use the capability of data mining to include the greater set of road variables for all available crashes with skid resistance values across the Queensland state main road network in order to understand the relationships among crash, traffic and road variables. This paper presents a data mining based methodology for the road asset management data to find out the various road properties that contribute unduly to crashes. The models demonstrate high levels of accuracy in predicting crashes in roads when various road properties are included. This paper presents the findings of these models to show the relationships among skid resistance, crashes, crash characteristics and other road characteristics such as seal type, seal age, road type, texture depth, lane count, pavement width, rutting, speed limit, traffic rates intersections, traffic signage and road design and so on.
Resumo:
Developing safe and sustainable road systems is a common goal in all countries. Applications to assist with road asset management and crash minimization are sought universally. This paper presents a data mining methodology using decision trees for modeling the crash proneness of road segments using available road and crash attributes. The models quantify the concept of crash proneness and demonstrate that road segments with only a few crashes have more in common with non-crash roads than roads with higher crash counts. This paper also examines ways of dealing with highly unbalanced data sets encountered in the study.
Resumo:
It is commonly accepted that wet roads have higher risk of crash than dry roads; however, providing evidence to support this assumption presents some difficulty. This paper presents a data mining case study in which predictive data mining is applied to model the skid resistance and crash relationship to search for discernable differences in the probability of wet and dry road segments having crashes based on skid resistance. The models identify an increased probability of wet road segments having crashes for mid-range skid resistance values.
Resumo:
Association rule mining has contributed to many advances in the area of knowledge discovery. However, the quality of the discovered association rules is a big concern and has drawn more and more attention recently. One problem with the quality of the discovered association rules is the huge size of the extracted rule set. Often for a dataset, a huge number of rules can be extracted, but many of them can be redundant to other rules and thus useless in practice. Mining non-redundant rules is a promising approach to solve this problem. In this paper, we first propose a definition for redundancy, then propose a concise representation, called a Reliable basis, for representing non-redundant association rules. The Reliable basis contains a set of non-redundant rules which are derived using frequent closed itemsets and their generators instead of using frequent itemsets that are usually used by traditional association rule mining approaches. An important contribution of this paper is that we propose to use the certainty factor as the criterion to measure the strength of the discovered association rules. Using this criterion, we can ensure the elimination of as many redundant rules as possible without reducing the inference capacity of the remaining extracted non-redundant rules. We prove that the redundancy elimination, based on the proposed Reliable basis, does not reduce the strength of belief in the extracted rules. We also prove that all association rules, their supports and confidences, can be retrieved from the Reliable basis without accessing the dataset. Therefore the Reliable basis is a lossless representation of association rules. Experimental results show that the proposed Reliable basis can significantly reduce the number of extracted rules. We also conduct experiments on the application of association rules to the area of product recommendation. The experimental results show that the non-redundant association rules extracted using the proposed method retain the same inference capacity as the entire rule set. This result indicates that using non-redundant rules only is sufficient to solve real problems needless using the entire rule set.
Resumo:
Road crashes cost world and Australian society a significant proportion of GDP, affecting productivity and causing significant suffering for communities and individuals. This paper presents a case study that generates data mining models that contribute to understanding of road crashes by allowing examination of the role of skid resistance (F60) and other road attributes in road crashes. Predictive data mining algorithms, primarily regression trees, were used to produce road segment crash count models from the road and traffic attributes of crash scenarios. The rules derived from the regression trees provide evidence of the significance of road attributes in contributing to crash, with a focus on the evaluation of skid resistance.
Resumo:
The proliferation of media services enabled by digital technologies poses a serious challenge to public service broadcasting rationales based on media scarcity. Looking to the past and future, we articulate an important role that the Australian Broadcasting Corporation (ABC) might play in the digital age. We argue that historically the ABC has acted beyond its institutional broadcasting remit to facilitate cultural development and, drawing on the example of Pool (an online community of creative practitioners established and maintained by the ABC), point to a key role it might play in fostering network innovation in what are now conceptualised as the creative industries.
Resumo:
Creative Industries was adopted as a platform in the 90s by the Blair government in the UK to describe the convergence of the arts, media, communication and information technologies as a newly formed cluster, providing economic and cultural capital for the knowledge economy. The philosophy and rhetoric which has grown around this concept (Leadbeater 2000, Castells 2000, Florida 2000, Caves 2000, Hartley 2000) has been influential in re-contextualising culture and the arts in the 21st century. Where governments and educational institutions have embraced the context of the creative industries, it is having a profound effect on the way arts are being positioned, originally as ‘creative content’ for the new economy. Countries and regions which have actively targeted the Creative Industries as an important economic growth factor in a post-industrial environment are numerous, but it is interesting to note that North and South East Asia and Australia have been at the forefront of developing the Creative Industries in its various guises. It could be argued that the initial phase of Creative Industries concentrated on media and communication technologies to provide commercial outcomes in small incubator business models; developing, for example, products for the games industry. Creative Industries is now entering a second phase of development; one in which the broader palette of the arts, though still not at the forefront of debate, is being re-examined. Both phases of Creative Industries have emphasised creativity and innovation as key drivers in the success and effectiveness of this sector, and although the arts by no means has a monopoly on these drivers, it is where they have an important part to play in the creative industries context. Arguably, the second wave of the creative industries acknowledges to a greater extent that commercialisation works in tandem with government and other support in a complex mixed economic model. In relation to the performing arts, the global market has seen an increase in large-scale cultural events such as festivals which are providing employment for the arts industry and multiplier effects in other parts of the economy. Differentiated product is important in this competitive arena and the use of mediated and digitised environments has been able to increase the amount of arts product available to an international market. This changed environment requires the development of new skills for our artists and producers and has given rise to a reappraisal of approaches to arts training and research in the Higher Degree Education sector (Brown 2007, Cunningham 2006). This paper examines pedagogical changes which took place in the first Creative Industries Faculty in the world at Queensland University of Technology as well as the increased opportunities for leading research initiatives. It concludes with the example of an interdisciplinary artwork produced in a creative industries precinct, exemplifying the convergence of arts and communication technologies and that of artistic practice and research.
Resumo:
Item folksonomy or tag information is a kind of typical and prevalent web 2.0 information. Item folksonmy contains rich opinion information of users on item classifications and descriptions. It can be used as another important information source to conduct opinion mining. On the other hand, each item is associated with taxonomy information that reflects the viewpoints of experts. In this paper, we propose to mine for users’ opinions on items based on item taxonomy developed by experts and folksonomy contributed by users. In addition, we explore how to make personalized item recommendations based on users’ opinions. The experiments conducted on real word datasets collected from Amazon.com and CiteULike demonstrated the effectiveness of the proposed approaches.
Resumo:
In asset intensive industries such as mining, oil & gas, utilities etc. most of the capital expenditure happens on acquiring engineering assets. Process of acquiring assets is called as “Procurement” or “Acquisition”. An asset procurement decision should be taken in consideration with the installation, commissioning, operational, maintenance and disposal needs of an asset or spare. However, such cross-functional collaboration and communication does not appear to happen between engineering, maintenance, warehousing and procurement functions in many asset intensive industries. Acquisition planning and execution are two distinct parts of asset acquisition process. Acquisition planning or procurement planning is responsible for determining exactly what is required to be purchased. It is important that an asset acquisition decision is the result of cross-functional decision making process. An acquisition decision leads to a formal purchase order. Most costly asset decisions occur even before they are acquired. Therefore, acquisition decision should be an outcome of an integrated planning & decision making process. Asset intensive organizations both, Government and non Government in Australia spent AUD 102.5 Billion on asset acquisition in year 2008-09. There is widespread evidence of many assets and spare not being used or utilized and in the end are written off. This clearly shows that many organizations end up buying assets or spares which were not required or non-conforming to the needs of user functions. It is due the fact that strategic and software driven procurement process do not consider all the requirements from various functions within the organization which contribute to the operation and maintenance of the asset over its life cycle. There is a lot of research done on how to implement an effective procurement process. There are numerous software solutions available for executing a procurement process. However, not much research is done on how to arrive at a cross functional procurement planning process. It is also important to link procurement planning process to procurement execution process. This research will discuss ““Acquisition Engineering Model” (AEM) framework, which aims at assisting acquisition decision making based on various criteria to satisfy cross-functional organizational requirements. Acquisition Engineering Model (AEM) will consider inputs from corporate asset management strategy, production management, maintenance management, warehousing, finance and HSE. Therefore, it is essential that the multi-criteria driven acquisition planning process is carried out and its output is fed to the asset acquisition (procurement execution) process. An effective procurement decision making framework to perform acquisition planning which considers various functional criteria will be discussed in this paper.
Resumo:
This is the final report from a study into the social impact of mining in Queensland.
Resumo:
It is a big challenge to clearly identify the boundary between positive and negative streams for information filtering systems. Several attempts have used negative feedback to solve this challenge; however, there are two issues for using negative relevance feedback to improve the effectiveness of information filtering. The first one is how to select constructive negative samples in order to reduce the space of negative documents. The second issue is how to decide noisy extracted features that should be updated based on the selected negative samples. This paper proposes a pattern mining based approach to select some offenders from the negative documents, where an offender can be used to reduce the side effects of noisy features. It also classifies extracted features (i.e., terms) into three categories: positive specific terms, general terms, and negative specific terms. In this way, multiple revising strategies can be used to update extracted features. An iterative learning algorithm is also proposed to implement this approach on the RCV1 data collection, and substantial experiments show that the proposed approach achieves encouraging performance and the performance is also consistent for adaptive filtering as well.
Resumo:
Automated analysis of the sentiments presented in online consumer feedbacks can facilitate both organizations’ business strategy development and individual consumers’ comparison shopping. Nevertheless, existing opinion mining methods either adopt a context-free sentiment classification approach or rely on a large number of manually annotated training examples to perform context sensitive sentiment classification. Guided by the design science research methodology, we illustrate the design, development, and evaluation of a novel fuzzy domain ontology based contextsensitive opinion mining system. Our novel ontology extraction mechanism underpinned by a variant of Kullback-Leibler divergence can automatically acquire contextual sentiment knowledge across various product domains to improve the sentiment analysis processes. Evaluated based on a benchmark dataset and real consumer reviews collected from Amazon.com, our system shows remarkable performance improvement over the context-free baseline.
Resumo:
Many data mining techniques have been proposed for mining useful patterns in text documents. However, how to effectively use and update discovered patterns is still an open research issue, especially in the domain of text mining. Since most existing text mining methods adopted term-based approaches, they all suffer from the problems of polysemy and synonymy. Over the years, people have often held the hypothesis that pattern (or phrase) based approaches should perform better than the term-based ones, but many experiments did not support this hypothesis. This paper presents an innovative technique, effective pattern discovery which includes the processes of pattern deploying and pattern evolving, to improve the effectiveness of using and updating discovered patterns for finding relevant and interesting information. Substantial experiments on RCV1 data collection and TREC topics demonstrate that the proposed solution achieves encouraging performance.
Resumo:
It is a big challenge to guarantee the quality of discovered relevance features in text documents for describing user preferences because of the large number of terms, patterns, and noise. Most existing popular text mining and classification methods have adopted term-based approaches. However, they have all suffered from the problems of polysemy and synonymy. Over the years, people have often held the hypothesis that pattern-based methods should perform better than term-based ones in describing user preferences, but many experiments do not support this hypothesis. The innovative technique presented in paper makes a breakthrough for this difficulty. This technique discovers both positive and negative patterns in text documents as higher level features in order to accurately weight low-level features (terms) based on their specificity and their distributions in the higher level features. Substantial experiments using this technique on Reuters Corpus Volume 1 and TREC topics show that the proposed approach significantly outperforms both the state-of-the-art term-based methods underpinned by Okapi BM25, Rocchio or Support Vector Machine and pattern based methods on precision, recall and F measures.