130 resultados para data gathering algorithm

em Deakin Research Online - Australia


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Differential privacy is a strong definition for protecting individual privacy in data releasing and mining. However, it is a rigid definition introducing a large amount of noise to the original dataset, which significantly decreases the quality of data mining results. Recently, how to design a suitable data releasing algorithm for data mining purpose is a hot research area. In this paper, we propose a differential private data releasing algorithm for decision tree construction. The proposed algorithm provides a non-interactive data releasing method through which miner can obtain the complete dataset for data mining purpose. With a given privacy budget, the proposed algorithm generalizes the original dataset, and then specializes it in a differential privacy constrain to construct decision trees. As the designed novel scheme selection operation can fully utilize the allocated privacy budget, the data set released by the proposed algorithm can yield better decision tree models than other method. Experimental results demonstrate that the proposed algorithm outperforms existing methods for private decision tree construction.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In data gathering wireless sensor networks, data loss often happens due to external faults such as random link faults and hazard node faults, since sensor nodes have constrained resources and are often deployed in inhospitable environments. However, already known fault tolerance mechanisms often bring new internal faults (e.g. out-of-power faults and collisions on wireless bandwidth) to the original network and dissipate lots of extra energy and time to reduce data loss. Therefore, we propose a novel Dual Cluster Heads Cooperation (CoDuch) scheme to tolerate external faults while introducing less internal faults and dissipating less extra energy and time. In CoDuch scheme, dual cluster heads cooperate with each other to reduce extra costs by sending only one copy of sensed data to the Base Station; also, dual cluster heads check errors with each other during the collecting data process. Two algorithms are developed based on the CoDuch scheme: CoDuch-l for tolerating link faults and CoDuch-b for tolerating both link faults and node faults; theory and experimental study validate their effectiveness and efficiency. © 2010 The Author Published by Oxford University Press on behalf of The British Computer Society. All rights reserved.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper describes a new method of monotone interpolation and smoothing of multivariate scattered data. It is based on the assumption that the function to be approximated is Lipschitz continuous. The method provides the optimal approximation in the worst case scenario and tight error bounds. Smoothing of noisy data subject to monotonicity constraints is converted into a quadratic programming problem. Estimation of the unknown Lipschitz constant from the data by sample splitting and cross-validation is described. Extension of the method for locally Lipschitz functions is presented.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper, we propose a data based neural network leader-follower control for multi-agent networks where each agent is described by a class of high-order uncertain nonlinear systems with input perturbation. The control laws are developed using multiple-surface sliding control technique. In particular, novel set of sliding variables are proposed to guarantee leader-follower consensus on the sliding surfaces. Novel switching is proposed to overcome the unavailability of instantaneous control output from the neighbor. By utilizing RBF neural network and Fourier series to approximate the unknown functions, leader-follower consensus can be reached, under the condition that the dynamic equations of all agents are unknown. An O(n) data based algorithm is developed, using only the network’s measurable input/output data to generate the distributed virtual control laws. Simulation results demonstrate the effectiveness of the approach.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The reliability of an induced classifier can be affected by several factors including the data oriented factors and the algorithm oriented factors [3]. In some cases, the reliability could also be affected by knowledge oriented factors. In this chapter, we analyze three special cases to examine the reliability of the discovered knowledge. Our case study results show that (1) in the cases of mining from low quality data, rough classification approach is more reliable than exact approach which in general tolerate to low quality data; (2) Without sufficient large size of the data, the reliability of the discovered knowledge will be decreased accordingly; (3) The reliability of point learning approach could easily be misled by noisy data. It will in most cases generate an unreliable interval and thus affect the reliability of the discovered knowledge. It is also reveals that the inexact field is a good learning strategy that could model the potentials and to improve the discovery reliability.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Failures are normal rather than exceptional in the cloud computing environments. To improve system avai-lability, replicating the popular data to multiple suitable locations is an advisable choice, as users can access the data from a nearby site. This is, however, not the case for replicas which must have a fixed number of copies on several locations. How to decide a reasonable number and right locations for replicas has become a challenge in the cloud computing. In this paper, a dynamic data replication strategy is put forward with a brief survey of replication strategy suitable for distributed computing environments. It includes: 1) analyzing and modeling the relationship between system availability and the number of replicas; 2) evaluating and identifying the popular data and triggering a replication operation when the popularity data passes a dynamic threshold; 3) calculating a suitable number of copies to meet a reasonable system byte effective rate requirement and placing replicas among data nodes in a balanced way; 4) designing the dynamic data replication algorithm in a cloud. Experimental results demonstrate the efficiency and effectiveness of the improved system brought by the proposed strategy in a cloud.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Stochastic search techniques such as evolutionary algorithms (EA) are known to be better explorer of search space as compared to conventional techniques including deterministic methods. However, in the era of big data like most other search methods and learning algorithms, suitability of evolutionary algorithms is naturally questioned. Big data pose new computational challenges including very high dimensionality and sparseness of data. Evolutionary algorithms' superior exploration skills should make them promising candidates for handling optimization problems involving big data. High dimensional problems introduce added complexity to the search space. However, EAs need to be enhanced to ensure that majority of the potential winner solutions gets the chance to survive and mature. In this paper we present an evolutionary algorithm with enhanced ability to deal with the problems of high dimensionality and sparseness of data. In addition to an informed exploration of the solution space, this technique balances exploration and exploitation using a hierarchical multi-population approach. The proposed model uses informed genetic operators to introduce diversity by expanding the scope of search process at the expense of redundant less promising members of the population. Next phase of the algorithm attempts to deal with the problem of high dimensionality by ensuring broader and more exhaustive search and preventing premature death of potential solutions. To achieve this, in addition to the above exploration controlling mechanism, a multi-tier hierarchical architecture is employed, where, in separate layers, the less fit isolated individuals evolve in dynamic sub-populations that coexist alongside the original or main population. Evaluation of the proposed technique on well known benchmark problems ascertains its superior performance. The algorithm has also been successfully applied to a real world problem of financial portfolio management. Although the proposed method cannot be considered big data-ready, it is certainly a move in the right direction.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper, we study two tightly coupled issues: space-crossing community detection and its influence on data forwarding in Mobile Social Networks (MSNs) by taking the hybrid underlying networks with infrastructure support into consideration. The hybrid underlying network is composed of large numbers of mobile users and a small portion of Access Points (APs). Because APs can facilitate the communication among long-distance nodes, the concept of physical proximity community can be extended to be one across the geographical space. In this work, we first investigate a space-crossing community detection method for MSNs. Based on the detection results, we design a novel data forwarding algorithm SAAS (Social Attraction and AP Spreading), and show how to exploit the space-crossing communities to improve the data forwarding efficiency. We evaluate our SAAS algorithm on real-life data from MIT Reality Mining and University of Illinois Movement (UIM). Results show that space-crossing community plays a positive role in data forwarding in MSNs in terms of delivery ratio and delay. Based on this new type of community, SAAS achieves a better performance than existing social community-based data forwarding algorithms in practice, including Bubble Rap and Nguyen's Routing algorithms. © 2014 IEEE.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper, we study two tightly coupled issues, space-crossing community detection and its influence on data forwarding in mobile social networks (MSNs). We propose a communication framework containing the hybrid underlying network with access point (AP) support for data forwarding and the base stations for managing most of control traffic. The concept of physical proximity community can be extended to be one across the geographical space, because APs can facilitate the communication among long-distance nodes. Space-crossing communities are obtained by merging some pairs of physical proximity communities. Based on the space-crossing community, we define two cases of node local activity and use them as the input of inner product similarity measurement. We design a novel data forwarding algorithm Social Attraction and Infrastructure Support (SAIS), which applies similarity attraction to route to neighbor more similar to destination, and infrastructure support phase to route the message to other APs within common connected components. We evaluate our SAIS algorithm on real-life datasets from MIT Reality Mining and University of Illinois Movement (UIM). Results show that space-crossing community plays a positive role in data forwarding in MSNs. Based on this new type of community, SAIS achieves a better performance than existing popular social community-based data forwarding algorithms in practice, including Simbet, Bubble Rap and Nguyen's Routing algorithms.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Presents a response of the author on the reliability and validity of Freud's method of free association and interpretation. Data gathering in clinical setting of psychoanalysis; Claims for reliability and validity; View of Freud on determinism.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

If you are a journalist of any kind, you now realize that you need to know how to find the information you need online. This book shows you how to find declassified governmental files and statistics of all kinds, outlines the use of simple and complex search engines for small and large data gathering, and provides directories of subject experts. This book is for the many journalists around the world who didn't attend a formal journalism school before going to work, those who were educated before online research became mainstream, and for any student studying journalism today. It will teach you how to use the Internet wisely, efficiently, and comprehensively so that you will always have your facts straight and fast.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Wireless sensor networks (WSN) are attractive for information gathering in large-scale data rich environments. In order to fully exploit the data gathering and dissemination capabilities of these networks, energy-efficient and scalable solutions for data storage and information discovery are essential. In this paper, we formulate the information discovery problem as a load-balancing problem, with the combined aim being to maximize network lifetime and minimize query processing delay resulting in QoS improvements. We propose a novel information storage and distribution mechanism that takes into account the residual energy levels in individual sensors. Further, we propose a hybrid push-pull strategy that enables fast response to information discovery queries.

Simulations results prove the proposed method(s) of information discovery offer significant QoS benefits for global as well as individual queries in comparison to previous approaches.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This research is the exploration of the lived experience of tertiary students in Australia with the medical condition usually known as ME/CFS (Myalgic Encephalomyelitis /Chronic Fatigue Syndrome) seeking to explore issues of equity and human rights from the perspective of the Disability Discrimination Act 1992. Students feel that their difficulties are not caused just by the illness itself but by the failure of the tertiary institutions to understand the effects of this illness on them, the student, especially within the areas of accommodations and assessments. Their lived experiences are studied to ascertain if their experiences differ from those of other tertiary students. Forty participants came from every state and territory of Australia and twenty -four of Australia's universities as well as eight Technical and Further Education/Open Training Education Network (TAFE/OTEN) colleges are represented. The selection of the chosen methodology, Critical Ethnography from a Habermasian perspective, has been circumscribed by the medical condition which placed limitations on methodology and also data gathering methods. Non-structured stories, in which the participants wrote of their lived experience as students, were considered the most appropriate source of data. These were transmitted by electronic mail (with some by postal mail) to the researcher. A short questionnaire provided a participant background to the stories and was also collated for a composite overview of the participants. The stories are analysed in a number of ways: six selected stories are retold and the issues arising from these stories have been weighed against the remainder of the stories. Four intertwined themes were constructed from the issues raised in each story. Apparent infringements of the Disability Discrimination Act (1992) which impact on quality of life, human rights and equity are found. No accommodations are being made by the academic institutions for the cognitive dysfunctions and learning difficulties. Students are stigmatised and lack credibility to negotiate appropriate academic accommodations. A possible means of improving the ability of students to negotiate appropriate accommodations is explored. Finally the researcher reflects on her own involvement in the research as an 'insider' researcher.