875 resultados para Distribution network reconfiguration problem


Relevância:

30.00% 30.00%

Publicador:

Resumo:

In aquatic systems, the ability of both the predator and prey to detect each other may be impaired by turbidity. This could lead to significant changes in the trophic interactions in the food web of lakes. Most fish use their vision for predation and the location of prey can be highly influenced by light level and clarity of the water environment. Turbidity is an optical property of water that causes light to be scattered and absorbed by particles and molecules. Turbidity is highly variable in lakes, due to seasonal changes in suspended sediments, algal blooms and wind-driven suspension of sediments especially in shallow waters. There is evidence that human activity has increased erosion leading to increased turbidity in aquatic systems. Turbidity could also play a significant role in distribution of fish. Turbidity could act as a cover for small fish and reduce predation risk. Diel horizontal migration by fish is common in shallow lakes and is considered as consequences of either optimal foraging behaviour for food or as a trade-off between foraging and predator avoidance. In turbid lakes, diel horizontal migration patterns could differ since turbidity can act as a refuge itself and affect the predator-prey interactions. Laboratory experiments were conducted with perch (Perca fluviatilis L.) and white bream (Abramis björkna (L.)) to clarify the effects of turbidity on their feeding. Additionally to clarify the effects of turbidity on predator preying on different types of prey, pikeperch larvae (Sander lucioperca (L.)), Daphnia pulex (Leydig), Sida crystallina (O.F. Müller), and Chaoborus flavicans (Meigen) were used as prey in different experiments. To clarify the role of turbidity in distribution and diel horizontal migration of perch, roach (Rutilus rutilus (L.)) and white bream, field studies were conducted in shallow turbid lakes. A clear and a turbid shallow lake were compared to investigate distribution of perch and roach in these two lakes in a 15-year study period. Feeding efficiency of perch and white bream was not significantly affected with increasing clay turbidity up to 50 NTU. The perch experiments with pikeperch larvae suggested that clay turbidity could act as a refuge especially at turbidity levels higher than 50 NTU. Perch experiments with different prey types suggested that pikeperch larvae probably use turbidity as a refuge better compared to Daphnia. Increase in turbidity probably has stronger affect on perch predating on plant-attached prey. The main findings of the thesis show that turbidity can play a significant role in distribution of fish. Perch and roach could use turbidity as refuge when macrophytes disappear while small perch may also use high turbidity as refuge when macrophytes are present. Floating-leaved macrophytes are probably good refuges for small fish in clay-turbid lakes and provide a certain level of turbidity and not too complex structure for refuge. The results give light to the predator-prey interactions in turbid environments. Turbidity of water should be taken in to account when studying the diel horizontal migrations and distribution of fish in shallow lakes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the liberalisation of electricity market it has become very important to determine the participants making use of the transmission network.Transmission line usage computation requires information of generator to load contributions and the path used by various generators to meet loads and losses. In this study relative electrical distance (RED) concept is used to compute reactive power contributions from various sources like generators, switchable volt-amperes reactive(VAR) sources and line charging susceptances that are scattered throughout the network, to meet the system demands. The transmission line charge susceptances contribution to the system reactive flows and its aid extended in reducing the reactive generation at the generator buses are discussed in this paper. Reactive power transmission cost evaluation is carried out in this study. The proposed approach is also compared with other approaches viz.,proportional sharing and modified Y-bus.Detailed case studies with base case and optimised results are carried out on a sample 8-bus system. IEEE 39-bus system and a practical 72-bus system, an equivalent of Indian Southern grid are also considered for illustration and results are discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Understanding the functioning of a neural system in terms of its underlying circuitry is an important problem in neuroscience. Recent d evelopments in electrophysiology and imaging allow one to simultaneously record activities of hundreds of neurons. Inferring the underlying neuronal connectivity patterns from such multi-neuronal spike train data streams is a challenging statistical and computational problem. This task involves finding significant temporal patterns from vast amounts of symbolic time series data. In this paper we show that the frequent episode mining methods from the field of temporal data mining can be very useful in this context. In the frequent episode discovery framework, the data is viewed as a sequence of events, each of which is characterized by an event type and its time of occurrence and episodes are certain types of temporal patterns in such data. Here we show that, using the set of discovered frequent episodes from multi-neuronal data, one can infer different types of connectivity patterns in the neural system that generated it. For this purpose, we introduce the notion of mining for frequent episodes under certain temporal constraints; the structure of these temporal constraints is motivated by the application. We present algorithms for discovering serial and parallel episodes under these temporal constraints. Through extensive simulation studies we demonstrate that these methods are useful for unearthing patterns of neuronal network connectivity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, numerical modelling of fracture in concrete using two-dimensional lattice model is presented and also a few issues related to lattice modelling technique applicable to concrete fracture are reviewed. A comparison is made with acoustic emission (AE) events with the number of fractured elements. To implement the heterogeneity of the plain concrete, two methods namely, by generating grain structure of the concrete using Fuller's distribution and the concrete material properties are randomly distributed following Gaussian distribution are used. In the first method, the modelling of the concrete at meso level is carried out following the existing methods available in literature. The shape of the aggregates present in the concrete are assumed as perfect spheres and shape of the same in two-dimensional lattice network is circular. A three-point bend (TPB) specimen is tested in the experiment under crack mouth opening displacement (CMOD) control at a rate of 0.0004 mm/sec and the fracture process in the same TPB specimen is modelled using regular triangular 2D lattice network. Load versus crack mouth opening isplacement (CMOD) plots thus obtained by using both the methods are compared with experimental results. It was observed that the number of fractured elements increases near the peak load and beyond the peak load. That is once the crack starts to propagate. AE hits also increase rapidly beyond the peak load. It is compulsory here to mention that although the lattice modelling of concrete fracture used in this present study is very similar to those already available in literature, the present work brings out certain finer details which are not available explicitly in the earlier works.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Previous studies have shown that buffering packets in DRAM is a performance bottleneck. In order to understand the impediments in accessing the DRAM, we developed a detailed Petri net model of IP forwarding application on IXP2400 that models the different levels of the memory hierarchy. The cell based interface used to receive and transmit packets in a network processor leads to some small size DRAM accesses. Such narrow accesses to the DRAM expose the bank access latency, reducing the bandwidth that can be realized. With real traces up to 30% of the accesses are smaller than the cell size, resulting in 7.7% reduction in DRAM bandwidth. To overcome this problem, we propose buffering these small chunks of data in the on chip scratchpad memory. This scheme also exploits greater degree of parallelism between different levels of the memory hierarchy. Using real traces from the internet, we show that the transmit rate can be improved by an average of 21% over the base scheme without the use of additional hardware. Further, the impact of different traffic patterns on the network processor resources is studied. Under real traffic conditions, we show that the data bus which connects the off-chip packet buffer to the micro-engines, is the obstacle in achieving higher throughput.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we propose a new security metric for measuring resilience of a symmetric key distribution scheme in wireless sensor network. A polynomial-based and a novel complete connectivity schemes are proposed and an analytical comparison, in terms of security and connectivity, between the schemes is shown. Motivated by the schemes, we derive general expressions for security and connectivity. A number of conclusions are made using these general expressions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We consider a dense, ad hoc wireless network confined to a small region, such that direct communication is possible between any pair of nodes. The physical communication model is that a receiver decodes the signal from a single transmitter, while treating all other signals as interference. Data packets are sent between source-destination pairs by multihop relaying. We assume that nodes self-organise into a multihop network such that all hops are of length d meters, where d is a design parameter. There is a contention based multiaccess scheme, and it is assumed that every node always has data to send, either originated from it or a transit packet (saturation assumption). In this scenario, we seek to maximize a measure of the transport capacity of the network (measured in bit-meters per second) over power controls (in a fading environment) and over the hop distance d, subject to an average power constraint. We first argue that for a dense collection of nodes confined to a small region, single cell operation is efficient for single user decoding transceivers. Then, operating the dense ad hoc network (described above) as a single cell, we study the optimal hop length and power control that maximizes the transport capacity for a given network power constraint. More specifically, for a fading channel and for a fixed transmission time strategy (akin to the IEEE 802.11 TXOP), we find that there exists an intrinsic aggregate bit rate (Theta(opt) bits per second, depending on the contention mechanism and the channel fading characteristics) carried by the network, when operating at the optimal hop length and power control. The optimal transport capacity is of the form d(opt)((P) over bar (t)) x Theta(opt) with d(opt) scaling as (P) over bar (1/eta)(t), where (P) over bar (t) is the available time average transmit power and eta is the path loss exponent. Under certain conditions on the fading distribution, we then provide a simple characterisation of the optimal operating point.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Solving large-scale all-to-all comparison problems using distributed computing is increasingly significant for various applications. Previous efforts to implement distributed all-to-all comparison frameworks have treated the two phases of data distribution and comparison task scheduling separately. This leads to high storage demands as well as poor data locality for the comparison tasks, thus creating a need to redistribute the data at runtime. Furthermore, most previous methods have been developed for homogeneous computing environments, so their overall performance is degraded even further when they are used in heterogeneous distributed systems. To tackle these challenges, this paper presents a data-aware task scheduling approach for solving all-to-all comparison problems in heterogeneous distributed systems. The approach formulates the requirements for data distribution and comparison task scheduling simultaneously as a constrained optimization problem. Then, metaheuristic data pre-scheduling and dynamic task scheduling strategies are developed along with an algorithmic implementation to solve the problem. The approach provides perfect data locality for all comparison tasks, avoiding rearrangement of data at runtime. It achieves load balancing among heterogeneous computing nodes, thus enhancing the overall computation time. It also reduces data storage requirements across the network. The effectiveness of the approach is demonstrated through experimental studies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Public key authentication is the verification of the identity-public key binding, and is foundational to the security of any network. The contribution of this thesis has been to provide public key authentication for a decentralised and resource challenged network such as an autonomous Delay Tolerant Network (DTN). It has resulted in the development and evaluation of a combined co-localisation trust system and key distribution scheme evaluated on a realistic large geographic scale mobility model. The thesis also addresses the problem of unplanned key revocation and replacement without any central authority.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we have first given a numerical procedure for the solution of second order non-linear ordinary differential equations of the type y″ = f (x;y, y′) with given initial conditions. The method is based on geometrical interpretation of the equation, which suggests a simple geometrical construction of the integral curve. We then translate this geometrical method to the numerical procedure adaptable to desk calculators and digital computers. We have studied the efficacy of this method with the help of an illustrative example with known exact solution. We have also compared it with Runge-Kutta method. We have then applied this method to a physical problem, namely, the study of the temperature distribution in a semi-infinite solid homogeneous medium for temperature-dependent conductivity coefficient.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Following Weisskopf, the kinematics of quantum mechanics is shown to lead to a modified charge distribution for a test electron embedded in the Fermi-Dirac vacuum with interesting consequences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Although empirical evidence suggests the contrary, many asset pricing models assume stock returns to be symmetrically distributed. In this paper it is argued that the occurrence of negative jumps in a firm's future earnings and, consequently, in its stock price, is positively related to the level of network externalities in the firm's product market. If the ex post frequency of these negative jumps in a sample does not equal the ex ante assessed probability of occurrence, the sample is subject to a peso problem. The hypothesis is tested for by regressing the skewness coefficient of a firm’s realised stock return distribution on the firm’s R&D intensity, i.e. the ratio of the firm’s research and development expenditure to its net sales. The empirical results support the technology-related peso problem hypothesis. In samples subject to such a peso problem, the returns are biased up and the variance is biased down.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we develop a novel auction algorithm for procuring wireless channel by a wireless node in a heterogeneous wireless network. We assume that the service providers of the heterogeneous wireless network are selfish and non-cooperative in the sense that they are only interested in maximizing their own utilities. The wireless user needs to procure wireless channels to execute multiple tasks. To solve the problem of the wireless user, we propose a reverse optimal (REVOPT) auction and derive an expression for the expected payment by the wireless user. The proposed auction mechanism REVOPT satisfies important game theoretic properties such as Bayesian incentive compatibility and individual rationality.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper we propose a general Linear Programming (LP) based formulation and solution methodology for obtaining optimal solution to the load distribution problem in divisible load scheduling. We exploit the power of the versatile LP formulation to propose algorithms that yield exact solutions to several very general load distribution problems for which either no solutions or only heuristic solutions were available. We consider both star (single-level tree) networks and linear daisy chain networks, having processors equipped with front-ends, that form the generic models for several important network topologies. We consider arbitrary processing node availability or release times and general models for communication delays and computation time that account for constant overheads such as start up times in communication and computation. The optimality of the LP based algorithms is proved rigorously.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The problem of detecting an unknown transient signal in noise is considered. The SNR of the observed data is first enhanced using wavelet domain filter The output of the wavelet domain filter is then transformed using a Wigner-Ville transform,which separates the spectrum of the observed signal into narrow frequency bands. Each subband signal at the output of the Wigner-ville block is subjected kto wavelet based level dependent denoising (WBLDD)to supress colored noise A weighted sum of the absolute value of outputs of WBLDD is passed through an energy detector, whose output is used as test statistic to take the final decision. By assigning weights proportional to the energy of the corresponding subband signals, the proposed detector approximates a frequency domain matched filter Simulation results are presented to show that the performance of the proposed detector is better than that of the wavelet packet transform based detector.