161 resultados para distributed nonlinearity


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Half sandwich complexes of the type [CpM(CO)(n)X] {X=Cl, Br, I; If, M=Fe, Ru; n=2 and if M=Mo; n=3} and [CpNiPPh3X] {X=Cl, Br, I} have been synthesized and their second order molecular nonlinearity (beta) measured at 1064 nm in CHCl3 by the hyper-Rayleigh scattering technique. Iron complexes consistently display larger beta values than ruthenium complexes while nickel complexes have marginally larger beta values than iron complexes. In the presence of an acceptor ligand such as CO or PPh3, the role of the halogen atom is that of a pi donor. The better overlap of Cl orbitals with Fe and Ni metal centres make Cl a better pi donor than Br or I in the respective complexes. Consequently, M-pi interaction is stronger in Fe/Ni-Cl complexes. The value of beta decreases as one goes down the halogen group. For the complexes of 4d metal ions where the metal-ligand distance is larger, the influence of pi orbital overlap appears to be less important, resulting in moderate changes in beta as a function of halogen substitution. (C) 2006 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Making use of the empirical potential functions for peptide NH .. O bonds, developed in this laboratory, the relative stabilities of the rightand left-handed α-helical structures of poly-L-alanine have been investigated, by calculating their conformational energies (V). The value of Vmin of the right-handed helix (αP) is about - 10.4 kcal/mole, and that of the left-handed helix (αM) is about - 9.6 kcal/mole, showing that the former is lower in energy by 0.8 kcal/mole. The helical parameters of the stable conformation of αP are n ∼ 3.6 and h ∼ 1.5 Å. The hydrogen bond of length 2.85 Å and nonlinearity of about 10° adds about 4.0 kcal/ mole to the stabilising energy of the helix in the minimum enregy region. The energy minimum is not sharply defined, but occurs over a long valley, suggesting that a distribution of conformations (φ{symbol}, ψ) of nearly the same energy may occur for the individual residues in a helix. The experimental data of a-helical fibres of poly-L-alanine are in good agreement with the theoretical results for αP. In the case of proteins, the mean values of (φ{symbol}, ψ) for different helices are distributed, but they invariably occur within the contour for V = Vmin + 2 kcal/mole for αP.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The scope of application of Laplace transforms presently limited to the study of linear partial differential equations, is extended to the nonlinear domain by this study. This has been achieved by modifying the definition of D transforms, put forth recently for the study of classes of nonlinear lumped parameter systems. The appropriate properties of the new D transforms are presented to bring out their applicability in the analysis of nonlinear distributed parameter systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Equivalence of certain classes of second-order non-linear distributed parameter systems and corresponding linear third-order systems is established through a differential transformation technique. As linear systems are amenable to analysis through existing techniques, this study is expected to offer a method of tackling certain classes of non-linear problems which may otherwise prove to be formidable in nature.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this two-part series of papers, a generalized non-orthogonal amplify and forward (GNAF) protocol which generalizes several known cooperative diversity protocols is proposed. Transmission in the GNAF protocol comprises of two phases - the broadcast phase and the cooperation phase. In the broadcast phase, the source broadcasts its information to the relays as well as the destination. In the cooperation phase, the source and the relays together transmit a space-time code in a distributed fashion. The GNAF protocol relaxes the constraints imposed by the protocol of Jing and Hassibi on the code structure. In Part-I of this paper, a code design criteria is obtained and it is shown that the GNAF protocol is delay efficient and coding gain efficient as well. Moreover GNAF protocol enables the use of sphere decoders at the destination with a non-exponential Maximum likelihood (ML) decoding complexity. In Part-II, several low decoding complexity code constructions are studied and a lower bound on the Diversity-Multiplexing Gain tradeoff of the GNAF protocol is obtained.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We address the problem of distributed space-time coding with reduced decoding complexity for wireless relay network. The transmission protocol follows a two-hop model wherein the source transmits a vector in the first hop and in the second hop the relays transmit a vector, which is a transformation of the received vector by a relay-specific unitary transformation. Design criteria is derived for this system model and codes are proposed that achieve full diversity. For a fixed number of relay nodes, the general system model considered in this paper admits code constructions with lower decoding complexity compared to codes based on some earlier system models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we describe an efficient coordinated-checkpointing and recovery algorithm which can work even when the channels are assumed to be non-FIFO, and messages may be lost. Nodes are assumed to be autonomous, and they do not block while taking checkpoints. Based on the local conditions, any process can request the previous coordinator for the 'permission' to initiate a new checkpoint. Allowing multiple initiators of checkpoints avoids the bottleneck associated with a single initiator, but the algorithm permits only a single instance of checkpointing process at any given time, thus reducing much of the overhead associated with multiple initiators of distributed algorithms.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The move towards IT outsourcing is the first step towards an environment where compute infrastructure is treated as a service. In utility computing this IT service has to honor Service Level Agreements (SLA) in order to meet the desired Quality of Service (QoS) guarantees. Such an environment requires reliable services in order to maximize the utilization of the resources and to decrease the Total Cost of Ownership (TCO). Such reliability cannot come at the cost of resource duplication, since it increases the TCO of the data center and hence the cost per compute unit. We, in this paper, look into aspects of projecting impact of hardware failures on the SLAs and techniques required to take proactive recovery steps in case of a predicted failure. By maintaining health vectors of all hardware and system resources, we predict the failure probability of resources based on observed hardware errors/failure events, at runtime. This inturn influences an availability aware middleware to take proactive action (even before the application is affected in case the system and the application have low recoverability). The proposed framework has been prototyped on a system running HP-UX. Our offline analysis of the prediction system on hardware error logs indicate no more than 10% false positives. This work to the best of our knowledge is the first of its kind to perform an end-to-end analysis of the impact of a hardware fault on application SLAs, in a live system.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The author presents adaptive control techniques for controlling the flow of real-time jobs from the peripheral processors (PPs) to the central processor (CP) of a distributed system with a star topology. He considers two classes of flow control mechanisms: (1) proportional control, where a certain proportion of the load offered to each PP is sent to the CP, and (2) threshold control, where there is a maximum rate at which each PP can send jobs to the CP. The problem is to obtain good algorithms for dynamically adjusting the control level at each PP in order to prevent overload of the CP, when the load offered by the PPs is unknown and varying. The author formulates the problem approximately as a standard system control problem in which the system has unknown parameters that are subject to change. Using well-known techniques (e.g., naive-feedback-controller and stochastic approximation techniques), he derives adaptive controls for the system control problem. He demonstrates the efficacy of these controls in the original problem by using the control algorithms in simulations of a queuing model of the CP and the load controls.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Erasure coding techniques are used to increase the reliability of distributed storage systems while minimizing storage overhead. Also of interest is minimization of the bandwidth required to repair the system following a node failure. In a recent paper, Wu et al. characterize the tradeoff between the repair bandwidth and the amount of data stored per node. They also prove the existence of regenerating codes that achieve this tradeoff. In this paper, we introduce Exact Regenerating Codes, which are regenerating codes possessing the additional property of being able to duplicate the data stored at a failed node. Such codes require low processing and communication overheads, making the system practical and easy to maintain. Explicit construction of exact regenerating codes is provided for the minimum bandwidth point on the storage-repair bandwidth tradeoff, relevant to distributed-mail-server applications. A sub-space based approach is provided and shown to yield necessary and sufficient conditions on a linear code to possess the exact regeneration property as well as prove the uniqueness of our construction. Also included in the paper, is an explicit construction of regenerating codes for the minimum storage point for parameters relevant to storage in peer-to-peer systems. This construction supports a variable number of nodes and can handle multiple, simultaneous node failures. All constructions given in the paper are of low complexity, requiring low field size in particular.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we address the problem of distributed transmission of functions of correlated sources over a fast fading multiple access channel (MAC). This is a basic building block in a hierarchical sensor network used in estimating a random field where the cluster head is interested only in estimating a function of the observations. The observations are transmitted to the cluster head through a fast fading MAC. We provide sufficient conditions for lossy transmission when the encoders and decoders are provided with partial information about the channel state. Furthermore signal side information maybe available at the encoders and the decoder. Various previous studies are shown as special cases. Efficient joint-source channel coding schemes are discussed for transmission of discrete and continuous alphabet sources to recover function values.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The problem of homogeneous solid propellant combustion instability is studied with a one-dimensional flame model, including the effects of gas-phase thermal inertia and nonlinearity. Computational results presented in this paper show nonlinear instabilities inherent in the equations, due to which periodic burning is found even under steady ambient conditions such as pressure. The stability boundary is obtained in terms of Denison-Baum parameters. It is found that inclusion of gas-phase thermal inertia stabilizes the combustion. Also, the effect of a distributed heat release in the gas phase, compared to the flame sheet model, is to destabilize the burning. Direct calculations for finite amplitude pressure disturbances show that two distinct resonant modes exist, the first one near the natural frequency as obtained from intrinsic instability analysis and a second mode occurring at a much higher driving frequency. It is found that er rn in the low frequency region, the response of the propellant is significantly affected by the specific type of gas-phase chemical heat-release model employed. Examination of frequency response function reveals that the role of gas-phase thermal inertia is to stabilize the burning near the first resonant mode. Calculations made for different amplitudes of driving pressure show that the mean burning rate decreases with increasing amplitude. Also, with an increase in the driving amplitude, higher harmonics are generated in the burning rate.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An important issue in the design of a distributed computing system (DCS) is the development of a suitable protocol. This paper presents an effort to systematize the protocol design procedure for a DCS. Protocol design and development can be divided into six phases: specification of the DCS, specification of protocol requirements, protocol design, specification and validation of the designed protocol, performance evaluation, and hardware/software implementation. This paper describes techniques for the second and third phases, while the first phase has been considered by the authors in their earlier work. Matrix and set theoretic based approaches are used for specification of a DCS and for specification of the protocol requirements. These two formal specification techniques form the basis of the development of a simple and straightforward procedure for the design of the protocol. The applicability of the above design procedure has been illustrated by considering an example of a computing system encountered on board a spacecraft. A Petri-net based approach has been adopted to model the protocol. The methodology developed in this paper can be used in other DCS applications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A detailed characterization of interference power statistics in CDMA systems is of considerable practical and theoretical interest. Such a characterization for uplink inter-cell interference has been difficult because of transmit power control, randomness in the number of interfering mobile stations, and randomness in their locations. We develop a new method to model the uplink inter-cell interference power as a lognormal distribution, and show that it is an order of magnitude more accurate than the conventional Gaussian approximation even when the average number of mobile stations per cell is relatively large and even outperforms the moment-matched lognormal approximation considered in the literature. The proposed method determines the lognormal parameters by matching its moment generating function with a new approximation of the moment generating function for the inter-cell interference. The method is tractable and exploits the elegant spatial Poisson process theory. Using several numerical examples, the accuracy of the proposed method in modeling the probability distribution of inter-cell interference is verified for both small and large values of interference.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There are a number of large networks which occur in many problems dealing with the flow of power, communication signals, water, gas, transportable goods, etc. Both design and planning of these networks involve optimization problems. The first part of this paper introduces the common characteristics of a nonlinear network (the network may be linear, the objective function may be non linear, or both may be nonlinear). The second part develops a mathematical model trying to put together some important constraints based on the abstraction for a general network. The third part deals with solution procedures; it converts the network to a matrix based system of equations, gives the characteristics of the matrix and suggests two solution procedures, one of them being a new one. The fourth part handles spatially distributed networks and evolves a number of decomposition techniques so that we can solve the problem with the help of a distributed computer system. Algorithms for parallel processors and spatially distributed systems have been described.There are a number of common features that pertain to networks. A network consists of a set of nodes and arcs. In addition at every node, there is a possibility of an input (like power, water, message, goods etc) or an output or none. Normally, the network equations describe the flows amoungst nodes through the arcs. These network equations couple variables associated with nodes. Invariably, variables pertaining to arcs are constants; the result required will be flows through the arcs. To solve the normal base problem, we are given input flows at nodes, output flows at nodes and certain physical constraints on other variables at nodes and we should find out the flows through the network (variables at nodes will be referred to as across variables).The optimization problem involves in selecting inputs at nodes so as to optimise an objective function; the objective may be a cost function based on the inputs to be minimised or a loss function or an efficiency function. The above mathematical model can be solved using Lagrange Multiplier technique since the equalities are strong compared to inequalities. The Lagrange multiplier technique divides the solution procedure into two stages per iteration. Stage one calculates the problem variables % and stage two the multipliers lambda. It is shown that the Jacobian matrix used in stage one (for solving a nonlinear system of necessary conditions) occurs in the stage two also.A second solution procedure has also been imbedded into the first one. This is called total residue approach. It changes the equality constraints so that we can get faster convergence of the iterations.Both solution procedures are found to coverge in 3 to 7 iterations for a sample network.The availability of distributed computer systems — both LAN and WAN — suggest the need for algorithms to solve the optimization problems. Two types of algorithms have been proposed — one based on the physics of the network and the other on the property of the Jacobian matrix. Three algorithms have been deviced, one of them for the local area case. These algorithms are called as regional distributed algorithm, hierarchical regional distributed algorithm (both using the physics properties of the network), and locally distributed algorithm (a multiprocessor based approach with a local area network configuration). The approach used was to define an algorithm that is faster and uses minimum communications. These algorithms are found to converge at the same rate as the non distributed (unitary) case.