36 resultados para scalability


Relevância:

10.00% 10.00%

Publicador:

Resumo:

The brightest and most vivid colours in nature arise from the interaction of light with surfaces that exhibit periodic structure on the micro- and nanoscale. In the wings of butterflies, for example, a combination of multilayer interference, optical gratings, photonic crystals and other optical structures gives rise to complex colour mixing. Although the physics of structural colours is well understood, it remains a challenge to create artificial replicas of natural photonic structures(1-3). Here we use a combination of layer deposition techniques, including colloidal self-assembly, sputtering and atomic layer deposition, to fabricate photonic structures that mimic the colour mixing effect found on the wings of the Indonesian butterfly Papilio blumei. We also show that a conceptual variation to the natural structure leads to enhanced optical properties. Our approach offers improved efficiency, versatility and scalability compared with previous approaches(4-6).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we propose a multi-camera application capable of processing high resolution images and extracting features based on colors patterns over graphic processing units (GPU). The goal is to work in real time under the uncontrolled environment of a sport event like a football match. Since football players are composed for diverse and complex color patterns, a Gaussian Mixture Models (GMM) is applied as segmentation paradigm, in order to analyze sport live images and video. Optimization techniques have also been applied over the C++ implementation using profiling tools focused on high performance. Time consuming tasks were implemented over NVIDIA's CUDA platform, and later restructured and enhanced, speeding up the whole process significantly. Our resulting code is around 4-11 times faster on a low cost GPU than a highly optimized C++ version on a central processing unit (CPU) over the same data. Real time has been obtained processing until 64 frames per second. An important conclusion derived from our study is the scalability of the application to the number of cores on the GPU. © 2011 Springer-Verlag.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Cloud services are exploding, and organizations are converging their data centers in order to take advantage of the predictability, continuity, and quality of service delivered by virtualization technologies. In parallel, energy-efficient and high-security networking is of increasing importance. Network operators, and service and product providers require a new network solution to efficiently tackle the increasing demands of this changing network landscape. Software-defined networking has emerged as an efficient network technology capable of supporting the dynamic nature of future network functions and intelligent applications while lowering operating costs through simplified hardware, software, and management. In this article, the question of how to achieve a successful carrier grade network with software-defined networking is raised. Specific focus is placed on the challenges of network performance, scalability, security, and interoperability with the proposal of potential solution directions.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In intelligent video surveillance systems, scalability (of the number of simultaneous video streams) is important. Two key factors which hinder scalability are the time spent in decompressing the input video streams, and the limited computational power of the processor. This paper demonstrates how a combination of algorithmic and hardware techniques can overcome these limitations, and significantly increase the number of simultaneous streams. The techniques used are processing in the compressed domain, and exploitation of the multicore and vector processing capability of modern processors. The paper presents a system which performs background modeling, using a Mixture of Gaussians approach. This is an important first step in the segmentation of moving targets. The paper explores the effects of reducing the number of coefficients in the compressed domain, in terms of throughput speed and quality of the background modeling. The speedups achieved by exploiting compressed domain processing, multicore and vector processing are explored individually. Experiments show that a combination of all these techniques can give a speedup of 170 times on a single CPU compared to a purely serial, spatial domain implementation, with a slight gain in quality.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Processor architectures has taken a turn towards many-core processors, which integrate multiple processing cores on a single chip to increase overall performance, and there are no signs that this trend will stop in the near future. Many-core processors are harder to program than multi-core and single-core processors due to the need of writing parallel or concurrent programs with high degrees of parallelism. Moreover, many-cores have to operate in a mode of strong scaling because of memory bandwidth constraints. In strong scaling increasingly finer-grain parallelism must be extracted in order to keep all processing cores busy.

Task dataflow programming models have a high potential to simplify parallel program- ming because they alleviate the programmer from identifying precisely all inter-task de- pendences when writing programs. Instead, the task dataflow runtime system detects and enforces inter-task dependences during execution based on the description of memory each task accesses. The runtime constructs a task dataflow graph that captures all tasks and their dependences. Tasks are scheduled to execute in parallel taking into account dependences specified in the task graph.

Several papers report important overheads for task dataflow systems, which severely limits the scalability and usability of such systems. In this paper we study efficient schemes to manage task graphs and analyze their scalability. We assume a programming model that supports input, output and in/out annotations on task arguments, as well as commutative in/out and reductions. We analyze the structure of task graphs and identify versions and generations as key concepts for efficient management of task graphs. Then, we present three schemes to manage task graphs building on graph representations, hypergraphs and lists. We also consider a fourth edge-less scheme that synchronizes tasks using integers. Analysis using micro-benchmarks shows that the graph representation is not always scalable and that the edge-less scheme introduces least overhead in nearly all situations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Scalability and efficiency of on-chip communication of emerging Multiprocessor System-on-Chip (MPSoC) are critical design considerations. Conventional bus based interconnection schemes no longer fit for MPSoC with a large number of cores. Networks-on-Chip (NoC) is widely accepted as the next generation interconnection scheme for large scale MPSoC. The increase of MPSoC complexity requires fast and accurate system-level modeling techniques for rapid modeling and veri-fication of emerging MPSoCs. However, the existing modeling methods are limited in delivering the essentials of timing accuracy and simulation speed. This paper proposes a novel system-level Networks-on-Chip (NoC) modeling method, which is based on SystemC and TLM2.0 and capable of delivering timing accuracy close to cycle accurate modeling techniques at a significantly lower simulation cost. Experimental results are presented to demonstrate the proposed method. ©2010 IEEE.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper introduces hybrid address spaces as a fundamental design methodology for implementing scalable runtime systems on many-core architectures without hardware support for cache coherence. We use hybrid address spaces for an implementation of MapReduce, a programming model for large-scale data processing, and the implementation of a remote memory access (RMA) model. Both implementations are available on the Intel SCC and are portable to similar architectures. We present the design and implementation of HyMR, a MapReduce runtime system whereby different stages and the synchronization operations between them alternate between a distributed memory address space and a shared memory address space, to improve performance and scalability. We compare HyMR to a reference implementation and we find that HyMR improves performance by a factor of 1.71× over a set of representative MapReduce benchmarks. We also compare HyMR with Phoenix++, a state-of-art implementation for systems with hardware-managed cache coherence in terms of scalability and sustained to peak data processing bandwidth, where HyMR demon- strates improvements of a factor of 3.1× and 3.2× respectively. We further evaluate our hybrid remote memory access (HyRMA) programming model and assess its performance to be superior of that of message passing.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A new regime of relativistic high-order harmonic generation has been discovered (Pirozhkov 2012 Phys. Rev. Lett. 108 135004). Multi-terawatt relativistic-irradiance (>1018 W cm−2) femtosecond (~30–50 fs) lasers focused to underdense (few × 1019 cm−3) plasma formed in gas jet targets produce comb-like spectra with hundreds of even and odd harmonic orders reaching the photon energy of 360 eV, including the 'water window' spectral range. Harmonics are generated either by linearly or circularly polarized pulses from the J-KAREN (KPSI, JAEA) and Astra Gemini (CLF, RAL, UK) lasers. The photon number scalability has been demonstrated with a 120 TW laser, producing 40 μJ sr−1 per harmonic at 120 eV. The experimental results are explained using particle-in-cell simulations and catastrophe theory. A new mechanism of harmonic generation by sharp, structurally stable, oscillating electron spikes at the joint of the boundaries of the wake and bow waves excited by a laser pulse is introduced. In this paper, detailed descriptions of the experiments, simulations and model are provided and new features are shown, including data obtained with a two-channel spectrograph, harmonic generation by circularly polarized laser pulses and angular distribution.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we present a hybrid BDI-PGM framework, in which PGMs (Probabilistic Graphical Models) are incorporated into a BDI (belief-desire-intention) architecture. This work is motivated by the need to address the scalability and noisy sensing issues in SCADA (Supervisory Control And Data Acquisition) systems. Our approach uses the incorporated PGMs to model the uncertainty reasoning and decision making processes of agents situated in a stochastic environment. In particular, we use Bayesian networks to reason about an agent’s beliefs about the environment based on its sensory observations, and select optimal plans according to the utilities of actions defined in influence diagrams. This approach takes the advantage of the scalability of the BDI architecture and the uncertainty reasoning capability of PGMs. We present a prototype of the proposed approach using a transit scenario to validate its effectiveness.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We introduce a new parallel pattern derived from a specific application domain and show how it turns out to have application beyond its domain of origin. The pool evolution pattern models the parallel evolution of a population subject to mutations and evolving in such a way that a given fitness function is optimized. The pattern has been demonstrated to be suitable for capturing and modeling the parallel patterns underpinning various evolutionary algorithms, as well as other parallel patterns typical of symbolic computation. In this paper we introduce the pattern, we discuss its implementation on modern multi/many core architectures and finally present experimental results obtained with FastFlow and Erlang implementations to assess its feasibility and scalability.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The papers in this special issue focus on the topic of location awareness for radio and networks. Localization-awareness using radio signals stands to revolutionize the fields of navigation and communication engineering. It can be utilized to great effect in the next generation of cellular networks, mining applications, health-care monitoring, transportation and intelligent highways, multi-robot applications, first responders operations, military applications, factory automation, building and environmental controls, cognitive wireless networks, commercial and social network applications, and smart spaces. A multitude of technologies can be used in location-aware radios and networks, including GNSS, RFID, cellular, UWB, WLAN, Bluetooth, cooperative localization, indoor GPS, device-free localization, IR, Radar, and UHF. The performances of these technologies are measured by their accuracy, precision, complexity, robustness, scalability, and cost. Given the many application scenarios across different disciplines, there is a clear need for a broad, up-to-date and cogent treatment of radio-based location awareness. This special issue aims to provide a comprehensive overview of the state-of-the-art in technology, regulation, and theory. It also presents a holistic view of research challenges and opportunities in the emerging areas of localization.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Electron–positron (e–p) plasmas are widely thought to be emitted, in the form of ultra-relativistic winds or collimated jets, by some of the most energetic or powerful objects in the Universe, such as black-holes, pulsars, and quasars. These phenomena represent an unmatched astrophysical laboratory to test physics at its limit and, given their immense distance from Earth (some even farther than several billion light years), they also provide a unique window on the very early stages of our Universe. However, due to such gigantic distances, their properties are only inferred from the indirect interpretation of their radiative signatures and from matching numerical models: their generation mechanism and dynamics still pose complicated enigmas to the scientific community. Small-scale reproductions in the laboratory would represent a fundamental step towards a deeper understanding of this exotic state of matter. Here we present recent experimental results concerning the laser-driven production of ultra-relativistic e–p beams. In particular, we focus on the possibility of generating beams that present charge neutrality and that allow for collective effects in their dynamics, necessary ingredients for the testing pair-plasma physics in the laboratory. A brief discussion of the analytical and numerical modelling of the dynamics of these plasmas is also presented in order to provide a summary of the novel plasma physics that can be accessed with these objects. Finally, general considerations on the scalability of laboratory plasmas up to astrophysical scenarios are given.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Demand Response (DR) algorithms manipulate the energy consumption schedules of controllable loads so as to satisfy grid objectives. Implementation of DR algorithms using a centralised agent can be problematic for scalability reasons, and there are issues related to the privacy of data and robustness to communication failures. Thus it is desirable to use a scalable decentralised algorithm for the implementation of DR. In this paper, a hierarchical DR scheme is proposed for Peak Minimisation (PM) based on Dantzig-Wolfe Decomposition (DWD). In addition, a Time Weighted Maximisation option is included in the cost function which improves the Quality of Service for devices seeking to receive their desired energy sooner rather than later. The paper also demonstrates how the DWD algorithm can be implemented more efficiently through the calculation of the upper and lower cost bounds after each DWD iteration.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Economic dispatch (ED) problems often exhibit non-linear, non-convex characteristics due to the valve point effects. Further, various constraints and factors, such as prohibited operation zones, ramp rate limits and security constraints imposed by the generating units, and power loss in transmission make it even more challenging to obtain the global optimum using conventional mathematical methods. Meta-heuristic approaches are capable of solving non-linear, non-continuous and non-convex problems effectively as they impose no requirements on the optimization problems. However, most methods reported so far mainly focus on a specific type of ED problems, such as static or dynamic ED problems. This paper proposes a hybrid harmony search with arithmetic crossover operation, namely ACHS, for solving five different types of ED problems, including static ED with valve point effects, ED with prohibited operating zones, ED considering multiple fuel cells, combined heat and power ED, and dynamic ED. In this proposed ACHS, the global best information and arithmetic crossover are used to update the newly generated solution and speed up the convergence, which contributes to the algorithm exploitation capability. To balance the exploitation and exploration capabilities, the opposition based learning (OBL) strategy is employed to enhance the diversity of solutions. Further, four commonly used crossover operators are also investigated, and the arithmetic crossover shows its efficiency than the others when they are incorporated into HS. To make a comprehensive study on its scalability, ACHS is first tested on a group of benchmark functions with a 100 dimensions and compared with several state-of-the-art methods. Then it is used to solve seven different ED cases and compared with the results reported in literatures. All the results confirm the superiority of the ACHS for different optimization problems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A simple yet efficient harmony search (HS) method with a new pitch adjustment rule (NPAHS) is proposed for dynamic economic dispatch (DED) of electrical power systems, a large-scale non-linear real time optimization problem imposed by a number of complex constraints. The new pitch adjustment rule is based on the perturbation information and the mean value of the harmony memory, which is simple to implement and helps to enhance solution quality and convergence speed. A new constraint handling technique is also developed to effectively handle various constraints in the DED problem, and the violation of ramp rate limits between the first and last scheduling intervals that is often ignored by existing approaches for DED problems is effectively eliminated. To validate the effectiveness, the NPAHS is first tested on 10 popular benchmark functions with 100 dimensions, in comparison with four HS variants and five state-of-the-art evolutionary algorithms. Then, NPAHS is used to solve three 24-h DED systems with 5, 15 and 54 units, which consider the valve point effects, transmission loss, emission and prohibited operating zones. Simulation results on all these systems show the scalability and superiority of the proposed NPAHS on various large scale problems.