819 resultados para Power distribution network


Relevância:

40.00% 40.00%

Publicador:

Resumo:

The dissertation studies the general area of complex networked systems that consist of interconnected and active heterogeneous components and usually operate in uncertain environments and with incomplete information. Problems associated with those systems are typically large-scale and computationally intractable, yet they are also very well-structured and have features that can be exploited by appropriate modeling and computational methods. The goal of this thesis is to develop foundational theories and tools to exploit those structures that can lead to computationally-efficient and distributed solutions, and apply them to improve systems operations and architecture.

Specifically, the thesis focuses on two concrete areas. The first one is to design distributed rules to manage distributed energy resources in the power network. The power network is undergoing a fundamental transformation. The future smart grid, especially on the distribution system, will be a large-scale network of distributed energy resources (DERs), each introducing random and rapid fluctuations in power supply, demand, voltage and frequency. These DERs provide a tremendous opportunity for sustainability, efficiency, and power reliability. However, there are daunting technical challenges in managing these DERs and optimizing their operation. The focus of this dissertation is to develop scalable, distributed, and real-time control and optimization to achieve system-wide efficiency, reliability, and robustness for the future power grid. In particular, we will present how to explore the power network structure to design efficient and distributed market and algorithms for the energy management. We will also show how to connect the algorithms with physical dynamics and existing control mechanisms for real-time control in power networks.

The second focus is to develop distributed optimization rules for general multi-agent engineering systems. A central goal in multiagent systems is to design local control laws for the individual agents to ensure that the emergent global behavior is desirable with respect to the given system level objective. Ideally, a system designer seeks to satisfy this goal while conditioning each agent’s control on the least amount of information possible. Our work focused on achieving this goal using the framework of game theory. In particular, we derived a systematic methodology for designing local agent objective functions that guarantees (i) an equivalence between the resulting game-theoretic equilibria and the system level design objective and (ii) that the resulting game possesses an inherent structure that can be exploited for distributed learning, e.g., potential games. The control design can then be completed by applying any distributed learning algorithm that guarantees convergence to the game-theoretic equilibrium. One main advantage of this game theoretic approach is that it provides a hierarchical decomposition between the decomposition of the systemic objective (game design) and the specific local decision rules (distributed learning algorithms). This decomposition provides the system designer with tremendous flexibility to meet the design objectives and constraints inherent in a broad class of multiagent systems. Furthermore, in many settings the resulting controllers will be inherently robust to a host of uncertainties including asynchronous clock rates, delays in information, and component failures.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

FSodium phosphate tellurite glasses in the system (NaPO3)(x)(TeO2)(1-x) were prepared and structurally characterized by thermal analysis, vibrational spectroscopy, X-ray photoelectron spectroscopy (XPS) and a variety of complementary solid-state nuclear magnetic resonance (NMR) techniques. Unlike the situation in other mixed-network-former glasses, the interaction between the two network formers tellurium oxide and phosphorus oxide produces no new structural units, and no sharing of the network modifier Na2O takes place. The glass structure can be regarded as a network of interlinked metaphosphate-type P(2) tetrahedral and TeO4/2 antiprismotic units. The combined interpretation of the O 1s XPS data and the P-31 solid-state NMR spectra presents clear quantitative evidence for a nonstatistical connectivity distribution. Rather the formation of homootomic P-O-P and Te-O-Te linkages is favored over mixed P-O-Te connectivities. As a consequence of this chemical segregation effect, the spatial sodium distribution is not random, as also indicated by a detailed analysis of P-31/No-23 rotational echo double-resonance (REDOR) experiments. ACHTUNGTRENUNG(TeO2)1 x were prepared and structurally characterized by thermal analysis,vibrat ional spectroscopy,X-ray photoelectron spectroscopy (XPS) and a variety of complementary solid-state nuclear magnetic resonance (NMR) techniques. Unlike the situation in other mixed-network-former glasses,the interaction between the two network formers tellurium oxide and phosphorus oxide produces no new structural units,and no sharing of the network modifier Na2O takes place. The glass structure can be regarded as a network of interlinked metaphosphate-type P(2) tetrahedral and TeO4/2 antiprismatic units. The combined interpretation of the O 1s XPS data and the 31P solid-state NMR spectra presents clear quantitative evidence for a nonstatistical connectivity distribution. Rather,the formation of homoatomic P O P and Te O Te linkages is favored over mixed P O Te connectivities. As a consequence of this chemical segregation effect,the spatial sodium distribution is not random,as also indicated by a detailed analysis of 31P/23Na rotational echo double-resonance (REDOR) experiments.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Dynamic Power Management (DPM) is a technique to reduce power consumption of electronic system by selectively shutting down idle components. In this article we try to introduce back propagation network and radial basis network into the research of the system-level power management policies. We proposed two PM policies-Back propagation Power Management (BPPM) and Radial Basis Function Power Management (RBFPM) which are based on Artificial Neural Networks (ANN). Our experiments show that the two power management policies greatly lowered the system-level power consumption and have higher performance than traditional Power Management(PM) techniques-BPPM is 1.09-competitive and RBFPM is 1.08-competitive vs. 1.79, 1.45, 1.18-competitive separately for traditional timeout PM, adaptive predictive PM and stochastic PM.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Dynamic Power Management (DPM) is a technique to reduce power consumption of electronic system. by selectively shutting down idle components. In this article we try to introduce back propagation network and radial basis network into the research of the system-level policies. We proposed two PAY policies-Back propagation Power Management (BPPM) and Radial Basis Function Power management (RBFPM) which are based on Artificial Neural Networks (ANN). Our experiments show that the two power management policies greatly lowered the system-level power consumption and have higher performance than traditional Power Management(PM) techniques-BPPM is 1.09-competitive and RBFPM is 1.08-competitive vs. 1.79,145,1.18-competitive separately for traditional timeout PM, adaptive predictive PM and stochastic PM.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Dynamic Power Management (DPM) is a technique to reduce power consumption of electronic system by selectively shutting down idle components. In this article we try to introduce back propagation network and radial basis network into the research of the system-level power management policies. We proposed two PM policies-Back propagation Power Management (BPPM) and Radial Basis Function Power Management (RBFPM) which are based on Artificial Neural Networks (ANN). Our experiments show that the two power management policies greatly lowered the system-level power consumption and have higher performance than traditional Power Management(PM) techniques-BPPM is 1.09-competitive and RBFPM is 1.08-competitive vs. 1.79 . 1.45 . 1.18-competitive separately for traditional timeout PM . adaptive predictive PM and stochastic PM.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

We explored the origin of power law distribution observed in single-molecule conformational dynamics experiments. By establishing a kinetic master equation approach to study statistically the microscopic state dynamics, we show that the underlying landscape with exponentially distributed density of states leads to power law distribution of kinetics. The exponential density of states emerges when the system becomes glassy and landscape becomes rough with significant trapping.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Numerous problems exist that can be modeled as traffic through a network in which constraints exist to regulate flow. Vehicular road travel, computer networks, and cloud based resource distribution, among others all have natural representations in this manner. As these networks grow in size and/or complexity, analysis and certification of the safety invariants becomes increasingly costly. The NetSketch formalism introduces a lightweight verification framework that allows for greater scalability than traditional analysis methods. The NetSketch tool was developed to provide the power of this formalism in an easy to use and intuitive user interface.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper explores reasons for the high degree of variability in the sizes of ASes that have recently been observed, and the processes by which this variable distribution develops. AS size distribution is important for a number of reasons. First, when modeling network topologies, an AS size distribution assists in labeling routers with an associated AS. Second, AS size has been found to be positively correlated with the degree of the AS (number of peering links), so understanding the distribution of AS sizes has implications for AS connectivity properties. Our model accounts for AS births, growth, and mergers. We analyze two models: one incorporates only the growth of hosts and ASes, and a second extends that model to include mergers of ASes. We show analytically that, given reasonable assumptions about the nature of mergers, the resulting size distribution exhibits a power law tail with the exponent independent of the details of the merging process. We estimate parameters of the models from measurements obtained from Internet registries and from BGP tables. We then compare the models solutions to empirical AS size distribution taken from Mercator and Skitter datasets, and find that the simple growth-based model yields general agreement with empirical data. Our analysis of the model in which mergers occur in a manner independent of the size of the merging ASes suggests that more detailed analysis of merger processes is needed.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The Transmission Control Protocol (TCP) has been the protocol of choice for many Internet applications requiring reliable connections. The design of TCP has been challenged by the extension of connections over wireless links. We ask a fundamental question: What is the basic predictive power of TCP of network state, including wireless error conditions? The goal is to improve or readily exploit this predictive power to enable TCP (or variants) to perform well in generalized network settings. To that end, we use Maximum Likelihood Ratio tests to evaluate TCP as a detector/estimator. We quantify how well network state can be estimated, given network response such as distributions of packet delays or TCP throughput that are conditioned on the type of packet loss. Using our model-based approach and extensive simulations, we demonstrate that congestion-induced losses and losses due to wireless transmission errors produce sufficiently different statistics upon which an efficient detector can be built; distributions of network loads can provide effective means for estimating packet loss type; and packet delay is a better signal of network state than short-term throughput. We demonstrate how estimation accuracy is influenced by different proportions of congestion versus wireless losses and penalties on incorrect estimation.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper proposes the use of in-network caches (which we call Angels) to reduce the Minimum Distribution Time (MDT) of a file from a seeder – a node that possesses the file – to a set of leechers – nodes who are interested in downloading the file. An Angel is not a leecher in the sense that it is not interested in receiving the entire file, but rather it is interested in minimizing the MDT to all leechers, and as such uses its storage and up/down-link capacity to cache and forward parts of the file to other peers. We extend the analytical results by Kumar and Ross [1] to account for the presence of angels by deriving a new lower bound for the MDT. We show that this newly derived lower bound is tight by proposing a distribution strategy under assumptions of a fluid model. We present a GroupTree heuristic that addresses the impracticalities of the fluid model. We evaluate our designs through simulations that show that our Group-Tree heuristic outperforms other heuristics, that it scales well with the increase of the number of leechers, and that it closely approaches the optimal theoretical bounds.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This thesis proposes the use of in-network caches (which we call Angels) to reduce the Minimum Distribution Time (MDT) of a file from a seeder – a node that possesses the file – to a set of leechers – nodes who are interested in downloading the file. An Angel is not a leecher in the sense that it is not interested in receiving the entire file, but rather it is interested in minimizing the MDT to all leechers, and as such uses its storage and up/down-link capacity to cache and forward parts of the file to other peers. We extend the analytical results by Kumar and Ross (Kumar and Ross, 2006) to account for the presence of angels by deriving a new lower bound for the MDT. We show that this newly derived lower bound is tight by proposing a distribution strategy under assumptions of a fluid model. We present a GroupTree heuristic that addresses the impracticalities of the fluid model. We evaluate our designs through simulations that show that our GroupTree heuristic outperforms other heuristics, that it scales well with the increase of the number of leechers, and that it closely approaches the optimal theoretical bounds.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Recent empirical studies have shown that Internet topologies exhibit power laws of the form for the following relationships: (P1) outdegree of node (domain or router) versus rank; (P2) number of nodes versus outdegree; (P3) number of node pairs y = x^α within a neighborhood versus neighborhood size (in hops); and (P4) eigenvalues of the adjacency matrix versus rank. However, causes for the appearance of such power laws have not been convincingly given. In this paper, we examine four factors in the formation of Internet topologies. These factors are (F1) preferential connectivity of a new node to existing nodes; (F2) incremental growth of the network; (F3) distribution of nodes in space; and (F4) locality of edge connections. In synthetically generated network topologies, we study the relevance of each factor in causing the aforementioned power laws as well as other properties, namely diameter, average path length and clustering coefficient. Different kinds of network topologies are generated: (T1) topologies generated using our parametrized generator, we call BRITE; (T2) random topologies generated using the well-known Waxman model; (T3) Transit-Stub topologies generated using GT-ITM tool; and (T4) regular grid topologies. We observe that some generated topologies may not obey power laws P1 and P2. Thus, the existence of these power laws can be used to validate the accuracy of a given tool in generating representative Internet topologies. Power laws P3 and P4 were observed in nearly all considered topologies, but different topologies showed different values of the power exponent α. Thus, while the presence of power laws P3 and P4 do not give strong evidence for the representativeness of a generated topology, the value of α in P3 and P4 can be used as a litmus test for the representativeness of a generated topology. We also find that factors F1 and F2 are the key contributors in our study which provide the resemblance of our generated topologies to that of the Internet.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Rachit Agarwal, Rafael V. Martinez-Catala, Sean Harte, Cedric Segard, Brendan O'Flynn, "Modeling Power in Multi-functionality Sensor Network Applications," sensorcomm, pp.507-512, 2008 Proceedings of the Second International Conference on Sensor Technologies and Applications, August 25-August 31 2008, Cap Esterel, France