109 resultados para Distributed Bragg reflector
Resumo:
Presented is a design methodology which permits the application of distributed coupled resonator bandpass filter principles to form wideband small-aperture evanescent-mode waveguide antenna designs. This approach permits matching of the complex antenna aperture admittance of an evanescent-mode open-ended waveguide to a real impedance generator, and thereby to a coaxial feed probe. A simulated reflection coefficient of < - 10 dB was obtained over a bandwidth of 20%, from 2.0-2.45 GHz, in a 2.58 GHz cutoff waveguide. Dielectric-filled propagating waveguide and air-filled evanescent-mode waveguide sections are used to form the resonators/coupling elements of the antenna's coupled resonator matching sections. Simulated realised gain variation from 3.4-5.0 dBi is observed across the bandwidth. The antenna's maximum aperture dimension is < 0.47 wavelength at the upper operating frequency and so it is suitable for use in a wide angle scanning phased array.
Resumo:
A phenomenology of distributed passive intermodulation generation in coplanar waveguide transmission line is presented. The theoretical analysis is based upon the generalised nonlinear transmission line model, which accounts for the coupling of two propagating modes. The case of weak substrate nonlinearity is considered and the model is given qualitative verification through the mapping of passive intermodulation products generated in coplanar waveguide fabricated on a commercial laminate. Implications for future research are discussed. © 2012 IEEE.
Resumo:
The scheduling problem in distributed data-intensive computing environments has become an active research topic due to the tremendous growth in grid and cloud computing environments. As an innovative distributed intelligent paradigm, swarm intelligence provides a novel approach to solving these potentially intractable problems. In this paper, we formulate the scheduling problem for work-flow applications with security constraints in distributed data-intensive computing environments and present a novel security constraint model. Several meta-heuristic adaptations to the particle swarm optimization algorithm are introduced to deal with the formulation of efficient schedules. A variable neighborhood particle swarm optimization algorithm is compared with a multi-start particle swarm optimization and multi-start genetic algorithm. Experimental results illustrate that population based meta-heuristics approaches usually provide a good balance between global exploration and local exploitation and their feasibility and effectiveness for scheduling work-flow applications. © 2010 Elsevier Inc. All rights reserved.
Resumo:
We consider the problem of self-healing in peer-to-peer networks that are under repeated attack by an omniscient adversary. We assume that, over a sequence of rounds, an adversary either inserts a node with arbitrary connections or deletes an arbitrary node from the network. The network responds to each such change by quick “repairs,” which consist of adding or deleting a small number of edges. These repairs essentially preserve closeness of nodes after adversarial deletions, without increasing node degrees by too much, in the following sense. At any point in the algorithm, nodes v and w whose distance would have been l in the graph formed by considering only the adversarial insertions (not the adversarial deletions), will be at distance at most l log n in the actual graph, where n is the total number of vertices seen so far. Similarly, at any point, a node v whose degree would have been d in the graph with adversarial insertions only, will have degree at most 3d in the actual graph. Our distributed data structure, which we call the Forgiving Graph, has low latency and bandwidth requirements. The Forgiving Graph improves on the Forgiving Tree distributed data structure from Hayes et al. (2008) in the following ways: 1) it ensures low stretch over all pairs of nodes, while the Forgiving Tree only ensures low diameter increase; 2) it handles both node insertions and deletions, while the Forgiving Tree only handles deletions; 3) it requires only a very simple and minimal initialization phase, while the Forgiving Tree initially requires construction of a spanning tree of the network.
Resumo:
We analyze ways by which people decompose into groups in distributed systems. We are interested in systems in which an agent can increase its utility by connecting to other agents, but must also pay a cost that increases with the size of the sys- tem. The right balance is achieved by the right size group of agents. We formulate and analyze three intuitive and realistic games and show how simple changes in the protocol can dras- tically improve the price of anarchy of these games. In partic- ular, we identify two important properties for a low price of anarchy: agreement in joining the system, and the possibil- ity of appealing a rejection from a system. We show that the latter property is especially important if there are some pre- existing constraints regarding who may collaborate (or com- municate) with whom.
Resumo:
Focusing on the uplink, where mobile users (each with a single transmit antenna) communicate with a base station with multiple antennas, we treat multiple users as antennas to enable spatial multiplexing across users. Introducing distributed closed-loop spatial multiplexing with threshold-based user selection, we propose two uplink channel-assigning strategies with limited feedback. We prove that the proposed system also outperforms the standard greedy scheme with respect to the degree of fairness, measured by the variance of the time averaged throughput. For uplink multi-antenna systems, we show that the proposed scheduling is a better choice than the greedy scheme in terms of the average BER, feedback complexity, and fairness. The numerical results corroborate our findings
Resumo:
Achieving a clearer picture of categorial distinctions in the brain is essential for our understanding of the conceptual lexicon, but much more fine-grained investigations are required in order for this evidence to contribute to lexical research. Here we present a collection of advanced data-mining techniques that allows the category of individual concepts to be decoded from single trials of EEG data. Neural activity was recorded while participants silently named images of mammals and tools, and category could be detected in single trials with an accuracy well above chance, both when considering data from single participants, and when group-training across participants. By aggregating across all trials, single concepts could be correctly assigned to their category with an accuracy of 98%. The pattern of classifications made by the algorithm confirmed that the neural patterns identified are due to conceptual category, and not any of a series of processing-related confounds. The time intervals, frequency bands and scalp locations that proved most informative for prediction permit physiological interpretation: the widespread activation shortly after appearance of the stimulus (from 100. ms) is consistent both with accounts of multi-pass processing, and distributed representations of categories. These methods provide an alternative to fMRI for fine-grained, large-scale investigations of the conceptual lexicon. © 2010 Elsevier Inc.
Resumo:
The ability to distribute quantum entanglement is a prerequisite for many fundamental tests of quantum theory and numerous quantum information protocols. Two distant parties can increase the amount of entanglement between them by means of quantum communication encoded in a carrier that is sent from one party to the other. Intriguingly, entanglement can be increased even when the exchanged carrier is not entangled with the parties. However, in light of the defining property of entanglement stating that it cannot increase under classical communication, the carrier must be quantum. Here we show that, in general, the increase of relative entropy of entanglement between two remote parties is bounded by the amount of nonclassical correlations of the carrier with the parties as quantified by the relative entropy of discord. We study implications of this bound, provide new examples of entanglement distribution via unentangled states, and put further limits on this phenomenon.
Resumo:
This paper discusses methods of using the Internet as a communications media between distributed generator sites to provide new forms of loss-of-mains protection. An analysis of the quality of the communications channels between several nodes on the network was carried out experimentally. It is shown that Internet connections in urban environments are already capable of providing real-time power system protection, whilst rural Internet connections are borderline suitable but could not yet be recommended as a primary method of protection. Two strategies of providing loss-of-mains across Internet protocol are considered, broadcast of a reference frequency or phasor representing the utility and an Internet based inter-tripping scheme.
Resumo:
We consider the problem of self-healing in peer-to-peer networks that are under repeated attack by an omniscient adversary. We assume that the following process continues for up to n rounds where n is the total number of nodes initially in the network: the adversary deletesan arbitrary node from the network, then the network responds by quickly adding a small number of new edges.
We present a distributed data structure that ensures two key properties. First, the diameter of the network is never more than O(log Delta) times its original diameter, where Delta is the maximum degree of the network initially. We note that for many peer-to-peer systems, Delta is polylogarithmic, so the diameter increase would be a O(loglog n) multiplicative factor. Second, the degree of any node never increases by more than 3 over its original degree. Our data structure is fully distributed, has O(1) latency per round and requires each node to send and receive O(1) messages per round. The data structure requires an initial setup phase that has latency equal to the diameter of the original network, and requires, with high probability, each node v to send O(log n) messages along every edge incident to v. Our approach is orthogonal and complementary to traditional topology-based approaches to defending against attack.