46 resultados para Load Balancing in Wireless LAN
Resumo:
Structured data represented in the form of graphs arises in several fields of the science and the growing amount of available data makes distributed graph mining techniques particularly relevant. In this paper, we present a distributed approach to the frequent subgraph mining problem to discover interesting patterns in molecular compounds. The problem is characterized by a highly irregular search tree, whereby no reliable workload prediction is available. We describe the three main aspects of the proposed distributed algorithm, namely a dynamic partitioning of the search space, a distribution process based on a peer-to-peer communication framework, and a novel receiver-initiated, load balancing algorithm. The effectiveness of the distributed method has been evaluated on the well-known National Cancer Institute’s HIV-screening dataset, where the approach attains close-to linear speedup in a network of workstations.
Resumo:
As a vital factor affecting system cost and lifetime, energy consumption in wireless sensor networks (WSNs) has been paid much attention to. This article reviews existing energy harvesting technology applied in WSNs, and analyzes advantages of harvesting radio frequency (RF) energy in WSNs.
Resumo:
The shallow water equations are solved using a mesh of polygons on the sphere, which adapts infrequently to the predicted future solution. Infrequent mesh adaptation reduces the cost of adaptation and load-balancing and will thus allow for more accurate mapping on adaptation. We simulate the growth of a barotropically unstable jet adapting the mesh every 12 h. Using an adaptation criterion based largely on the gradient of the vorticity leads to a mesh with around 20 per cent of the cells of a uniform mesh that gives equivalent results. This is a similar proportion to previous studies of the same test case with mesh adaptation every 1–20 min. The prediction of the mesh density involves solving the shallow water equations on a coarse mesh in advance of the locally refined mesh in order to estimate where features requiring higher resolution will grow, decay or move to. The adaptation criterion consists of two parts: that resolved on the coarse mesh, and that which is not resolved and so is passively advected on the coarse mesh. This combination leads to a balance between resolving features controlled by the large-scale dynamics and maintaining fine-scale features.
Resumo:
We present a general Multi-Agent System framework for distributed data mining based on a Peer-to-Peer model. Agent protocols are implemented through message-based asynchronous communication. The framework adopts a dynamic load balancing policy that is particularly suitable for irregular search algorithms. A modular design allows a separation of the general-purpose system protocols and software components from the specific data mining algorithm. The experimental evaluation has been carried out on a parallel frequent subgraph mining algorithm, which has shown good scalability performances.
Resumo:
Clustering is defined as the grouping of similar items in a set, and is an important process within the field of data mining. As the amount of data for various applications continues to increase, in terms of its size and dimensionality, it is necessary to have efficient clustering methods. A popular clustering algorithm is K-Means, which adopts a greedy approach to produce a set of K-clusters with associated centres of mass, and uses a squared error distortion measure to determine convergence. Methods for improving the efficiency of K-Means have been largely explored in two main directions. The amount of computation can be significantly reduced by adopting a more efficient data structure, notably a multi-dimensional binary search tree (KD-Tree) to store either centroids or data points. A second direction is parallel processing, where data and computation loads are distributed over many processing nodes. However, little work has been done to provide a parallel formulation of the efficient sequential techniques based on KD-Trees. Such approaches are expected to have an irregular distribution of computation load and can suffer from load imbalance. This issue has so far limited the adoption of these efficient K-Means techniques in parallel computational environments. In this work, we provide a parallel formulation for the KD-Tree based K-Means algorithm and address its load balancing issues.
Resumo:
LIght Detection And Ranging (LIDAR) data for terrain and land surveying has contributed to many environmental, engineering and civil applications. However, the analysis of Digital Surface Models (DSMs) from complex LIDAR data is still challenging. Commonly, the first task to investigate LIDAR data point clouds is to separate ground and object points as a preparatory step for further object classification. In this paper, the authors present a novel unsupervised segmentation algorithm-skewness balancing to separate object and ground points efficiently from high resolution LIDAR point clouds by exploiting statistical moments. The results presented in this paper have shown its robustness and its potential for commercial applications.
Resumo:
Recently, two approaches have been introduced that distribute the molecular fragment mining problem. The first approach applies a master/worker topology, the second approach, a completely distributed peer-to-peer system, solves the scalability problem due to the bottleneck at the master node. However, in many real world scenarios the participating computing nodes cannot communicate directly due to administrative policies such as security restrictions. Thus, potential computing power is not accessible to accelerate the mining run. To solve this shortcoming, this work introduces a hierarchical topology of computing resources, which distributes the management over several levels and adapts to the natural structure of those multi-domain architectures. The most important aspect is the load balancing scheme, which has been designed and optimized for the hierarchical structure. The approach allows dynamic aggregation of heterogenous computing resources and is applied to wide area network scenarios.
Resumo:
The creation of OFDM based Wireless Personal Area Networks (WPANs) has allowed the development of high bit-rate wireless communication devices suitable for streaming High Definition video between consumer products, as demonstrated in Wireless-USB and Wireless-HDMI. However, these devices need high frequency clock rates, particularly for the OFDM, FFT and symbol processing sections resulting in high silicon cost and high electrical power. The high clock rates make hardware prototyping difficult and verification is therefore very important but costly. Acknowledging that electrical power in wireless consumer devices is more critical than the number of implemented logic gates, this paper presents a Double Data Rate (DDR) architecture for implementation inside a OFDM baseband codec in order to reduce the high frequency clock rates by a complete factor of 2. The presented architecture has been implemented and tested for ECMA-368 (Wireless- USB context) resulting in a maximum clock rate of 264MHz instead of the expected 528MHz clock rate existing anywhere on the baseband codec die.
Resumo:
The creation of OFDM based Wireless Personal Area Networks (WPANs) has allowed high bit-rate wireless communication devices suitable for streaming High Definition video between consumer products as demonstrated in Wireless- USB. However, these devices need high clock rates, particularly for the OFDM sections resulting in high silicon cost and high electrical power. Acknowledging that electrical power in wireless consumer devices is more critical than the number of implemented logic gates, this paper presents a Double Data Rate (DDR) architecture to reduce the OFDM input and output clock rate by a factor of 2. The architecture has been implemented and tested for Wireless-USB (ECMA-368) resulting in a maximum clock of 264MHz instead of 528MHz existing anywhere on the die.
Resumo:
A method to map all the variants of the IEEE 802.11 MAC frames into the Multiband OFDM based ECMA-368 Physical standard is proposed, without contravening the standard. The transportation of IEEE 802.11 MAC frames over ECMA-368 allows for the migration current of Wireless LAN applications towards a Wireless Personal Area Network (WPAN) solution. This system benefits the Consumer Electronics Market as the high data-rate WPAN is capable of transporting broadcast-quality video while the same system can also transport existing applications available today, maintaining existing effort, products and backward-compatibility(1).
Resumo:
This paper presents a paralleled Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA., Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). We introduced hashtable into video processing and completed parallel implementation. We propose and evaluate parallel implementations of the LHMEA of TPA on clusters of workstations for real time video compression. It discusses how parallel video coding on load balanced multiprocessor systems can help, especially on motion estimation. The effect of load balancing for improved performance is discussed. The performance or the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.
Resumo:
This paper presents a paralleled Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA, Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). We introduced hashtable into video processing and completed parallel implementation. We propose and evaluate parallel implementations of the LHMEA of TPA on clusters of workstations for real time video compression. It discusses how parallel video coding on load balanced multiprocessor systems can help, especially on motion estimation. The effect of load balancing for improved performance is discussed. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.
Resumo:
This paper presents a paralleled Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA, Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). We introduced hashtable into video processing and completed parallel implementation. We propose and evaluate parallel implementations of the LHMEA of TPA on clusters of workstations for real time video compression. It discusses how parallel video coding on load balanced multiprocessor systems can help, especially on motion estimation. The effect of load balancing for improved performance is discussed. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.
Resumo:
Executive summary Nature of the problem (science/management/policy) • Freshwater ecosystems play a key role in the European nitrogen (N) cycle, both as a reactive agent that transfers, stores and processes N loadings from the atmosphere and terrestrial ecosystems, and as a natural environment severely impacted by the increase of these loadings. Approaches • This chapter is a review of major processes and factors controlling N transport and transformations for running waters, standing waters, groundwaters and riparian wetlands. Key findings/state of knowledge • The major factor controlling N processes in freshwater ecosystems is the residence time of water, which varies widely both in space and in time, and which is sensitive to changes in climate, land use and management. • The effects of increased N loadings to European freshwaters include acidification in semi-natural environments, and eutrophication in more disturbed ecosystems, with associated loss of biodiversity in both cases. • An important part of the nitrogen transferred by surface waters is in the form of organic N, as dissolved organic N (DON) and particulate organic N (PON). This part is dominant in semi-natural catchments throughout Europe and remains a significant component of the total N load even in nitrate enriched rivers. • In eutrophicated standing freshwaters N can be a factor limiting or co-limiting biological production, and control of both N and phosphorus (P) loading is oft en needed in impacted areas, if ecological quality is to be restored. Major uncertainties/challenges • The importance of storage and denitrifi cation in aquifers is a major uncertainty in the global N cycle, and controls in part the response of catchments to land use or management changes. In some aquifers, the increase of N concentrations will continue for decades even if efficient mitigation measures are implemented now. • Nitrate retention by riparian wetlands has oft en been highlighted. However, their use for mitigation must be treated with caution, since their effectiveness is difficult to predict, and side effects include increased DON emissions to adjacent open waters, N2O emissions to the atmosphere, and loss of biodiversity. • In fact, the character and specific spatial origins of DON are not fully understood, and similarly the quantitative importance of indirect N2O emissions from freshwater ecosystems as a result of N leaching losses from agricultural soils is still poorly known at the regional scale. • These major uncertainties remain due to the lack of adequate monitoring (all forms of N at a relevant frequency), especially – but not only – in the southern and eastern EU countries. Recommendations (research/policy) • The great variability of transfer pathways, buffering capacity and sensitivity of the catchments and of the freshwater ecosystems calls for site specific mitigation measures rather than standard ones applied at regional to national scale. • The spatial and temporal variations of the N forms, the processes controlling the transport and transformation of N within freshwaters, require further investigation if the role of N in influencing freshwater ecosystem health is to be better understood, underpinning the implementation of the EU Water Framework Directive for European freshwaters.
Resumo:
Load forecasting is an important task in the management of a power utility. The most recent developments in forecasting involve the use of artificial intelligence techniques, which offer powerful modelling capabilities. This paper discusses these techniques and provides a review of their application to load forecasting.