988 resultados para distributed databases


Relevância:

20.00% 20.00%

Publicador:

Resumo:

We consider the problem of self-healing in peer-to-peer networks that are under repeated attack by an omniscient adversary. We assume that, over a sequence of rounds, an adversary either inserts a node with arbitrary connections or deletes an arbitrary node from the network. The network responds to each such change by quick “repairs,” which consist of adding or deleting a small number of edges. These repairs essentially preserve closeness of nodes after adversarial deletions, without increasing node degrees by too much, in the following sense. At any point in the algorithm, nodes v and w whose distance would have been l in the graph formed by considering only the adversarial insertions (not the adversarial deletions), will be at distance at most l log n in the actual graph, where n is the total number of vertices seen so far. Similarly, at any point, a node v whose degree would have been d in the graph with adversarial insertions only, will have degree at most 3d in the actual graph. Our distributed data structure, which we call the Forgiving Graph, has low latency and bandwidth requirements. The Forgiving Graph improves on the Forgiving Tree distributed data structure from Hayes et al. (2008) in the following ways: 1) it ensures low stretch over all pairs of nodes, while the Forgiving Tree only ensures low diameter increase; 2) it handles both node insertions and deletions, while the Forgiving Tree only handles deletions; 3) it requires only a very simple and minimal initialization phase, while the Forgiving Tree initially requires construction of a spanning tree of the network.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We analyze ways by which people decompose into groups in distributed systems. We are interested in systems in which an agent can increase its utility by connecting to other agents, but must also pay a cost that increases with the size of the sys- tem. The right balance is achieved by the right size group of agents. We formulate and analyze three intuitive and realistic games and show how simple changes in the protocol can dras- tically improve the price of anarchy of these games. In partic- ular, we identify two important properties for a low price of anarchy: agreement in joining the system, and the possibil- ity of appealing a rejection from a system. We show that the latter property is especially important if there are some pre- existing constraints regarding who may collaborate (or com- municate) with whom.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Focusing on the uplink, where mobile users (each with a single transmit antenna) communicate with a base station with multiple antennas, we treat multiple users as antennas to enable spatial multiplexing across users. Introducing distributed closed-loop spatial multiplexing with threshold-based user selection, we propose two uplink channel-assigning strategies with limited feedback. We prove that the proposed system also outperforms the standard greedy scheme with respect to the degree of fairness, measured by the variance of the time averaged throughput. For uplink multi-antenna systems, we show that the proposed scheduling is a better choice than the greedy scheme in terms of the average BER, feedback complexity, and fairness. The numerical results corroborate our findings

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Achieving a clearer picture of categorial distinctions in the brain is essential for our understanding of the conceptual lexicon, but much more fine-grained investigations are required in order for this evidence to contribute to lexical research. Here we present a collection of advanced data-mining techniques that allows the category of individual concepts to be decoded from single trials of EEG data. Neural activity was recorded while participants silently named images of mammals and tools, and category could be detected in single trials with an accuracy well above chance, both when considering data from single participants, and when group-training across participants. By aggregating across all trials, single concepts could be correctly assigned to their category with an accuracy of 98%. The pattern of classifications made by the algorithm confirmed that the neural patterns identified are due to conceptual category, and not any of a series of processing-related confounds. The time intervals, frequency bands and scalp locations that proved most informative for prediction permit physiological interpretation: the widespread activation shortly after appearance of the stimulus (from 100. ms) is consistent both with accounts of multi-pass processing, and distributed representations of categories. These methods provide an alternative to fMRI for fine-grained, large-scale investigations of the conceptual lexicon. © 2010 Elsevier Inc.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The ability to distribute quantum entanglement is a prerequisite for many fundamental tests of quantum theory and numerous quantum information protocols. Two distant parties can increase the amount of entanglement between them by means of quantum communication encoded in a carrier that is sent from one party to the other. Intriguingly, entanglement can be increased even when the exchanged carrier is not entangled with the parties. However, in light of the defining property of entanglement stating that it cannot increase under classical communication, the carrier must be quantum. Here we show that, in general, the increase of relative entropy of entanglement between two remote parties is bounded by the amount of nonclassical correlations of the carrier with the parties as quantified by the relative entropy of discord. We study implications of this bound, provide new examples of entanglement distribution via unentangled states, and put further limits on this phenomenon.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper discusses methods of using the Internet as a communications media between distributed generator sites to provide new forms of loss-of-mains protection. An analysis of the quality of the communications channels between several nodes on the network was carried out experimentally. It is shown that Internet connections in urban environments are already capable of providing real-time power system protection, whilst rural Internet connections are borderline suitable but could not yet be recommended as a primary method of protection. Two strategies of providing loss-of-mains across Internet protocol are considered, broadcast of a reference frequency or phasor representing the utility and an Internet based inter-tripping scheme.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Congenital Erythrocytosis (CE), or congenital polycythemia, represents a rare and heterogeneous clinical entity. It is caused by deregulated red blood cell production where erythrocyte overproduction results in elevated hemoglobin and hematocrit levels. Primary congenital familial erythrocytosis is associated with low erythropoietin (Epo) levels and results from mutations in the Epo receptor gene (EPOR). Secondary congenital erythrocytosis arises from conditions causing tissue hypoxia and results in increased Epo production. These include hemoglobin variants with increased affinity for oxygen (HBB, HBA mutations), decreased production of 2,3-bisphosphoglycerate due to BPGM mutations, or mutations in the genes involved in the hypoxia sensing pathway (VHL, EPAS1 and EGLN1). Depending on the affected gene, CE can be inherited either in an autosomal dominant or recessive mode, with sporadic cases arising de novo. Despite recent important discoveries in the molecular pathogenesis of CE, the molecular causes remain to be identified in about 70% of the patients. With the objective of collecting all the published and unpublished cases of CE the COST action MPN&MPNr-Euronet developed a comprehensive internet-based database focusing on the registration of clinical history, hematological, biochemical and molecular data (http://www.erythrocytosis.org/). In addition, unreported mutations are also curated in the corresponding Leiden Open Variation Database (LOVD). This article is protected by copyright. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:



We consider the problem of self-healing in peer-to-peer networks that are under repeated attack by an omniscient adversary. We assume that the following process continues for up to n rounds where n is the total number of nodes initially in the network: the adversary deletesan arbitrary node from the network, then the network responds by quickly adding a small number of new edges.

We present a distributed data structure that ensures two key properties. First, the diameter of the network is never more than O(log Delta) times its original diameter, where Delta is the maximum degree of the network initially. We note that for many peer-to-peer systems, Delta is polylogarithmic, so the diameter increase would be a O(loglog n) multiplicative factor. Second, the degree of any node never increases by more than 3 over its original degree. Our data structure is fully distributed, has O(1) latency per round and requires each node to send and receive O(1) messages per round. The data structure requires an initial setup phase that has latency equal to the diameter of the original network, and requires, with high probability, each node v to send O(log n) messages along every edge incident to v. Our approach is orthogonal and complementary to traditional topology-based approaches to defending against attack.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We consider the problem of self-healing in peer-to-peer networks that are under repeated attack by an omniscient adversary. We assume that, over a sequence of rounds, an adversary either inserts a node with arbitrary connections or deletes an arbitrary node from the network. The network responds to each such change by quick "repairs," which consist of adding or deleting a small number of edges. These repairs essentially preserve closeness of nodes after adversarial deletions,without increasing node degrees by too much, in the following sense. At any point in the algorithm, nodes v and w whose distance would have been - in the graph formed by considering only the adversarial insertions (not the adversarial deletions), will be at distance at most - log n in the actual graph, where n is the total number of vertices seen so far. Similarly, at any point, a node v whose degreewould have been d in the graph with adversarial insertions only, will have degree at most 3d in the actual graph. Our distributed data structure, which we call the Forgiving Graph, has low latency and bandwidth requirements. The Forgiving Graph improves on the Forgiving Tree distributed data structure from Hayes et al. (2008) in the following ways: 1) it ensures low stretch over all pairs of nodes, while the Forgiving Tree only ensures low diameter increase; 2) it handles both node insertions and deletions, while the Forgiving Tree only handles deletions; 3) it requires only a very simple and minimal initialization phase, while the Forgiving Tree initially requires construction of a spanning tree of the network. © Springer-Verlag 2012.