36 resultados para expenditure constraint


Relevância:

20.00% 20.00%

Publicador:

Resumo:

An approach to the constraint counting theory of glasses is applied to many glass systems which include an oxide, chalcohalide, and chalcogenides. In this, shifting of the percolation threshold due to noncovalent bonding interactions in a basically covalent network and other recent extensions of the theory appear natural. This is particularly insightful and reveals that the chemical threshold signifies another structural transition along with the rigidity percolation threshold, thus unifying these two seemingly disparate toplogical concepts. [S0163-1829(99)11441-3].

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The half-duplex constraint, which mandates that a cooperative relay cannot transmit and receive simultaneously, considerably simplifies the demands made on the hardware and signal processing capabilities of a relay. However, the very inability of a relay to transmit and receive simultaneously leads to a potential under-utilization of time and bandwidth resources available to the system. We analyze the impact of the half-duplex constraint on the throughput of a cooperative relay system that uses rateless codes to harness spatial diversity and efficiently transmit information from a source to a destination. We derive closed-form expressions for the throughput of the system, and show that as the number of relays increases, the throughput approaches that of a system that uses more sophisticated full-duplex nodes. Thus, half-duplex nodes are well suited for cooperation using rateless codes despite the simplicity of both the cooperation protocol and the relays.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The objectives of this paper are to examine the loss of crack tip constraint in dynamically loaded fracture specimens and to assess whether it can lead to enhancement in the fracture toughness at high loading rates which has been observed in several experimental studies. To this end, 2-D plane strain finite element analyses of single edge notched (tension) specimen and three point bend specimen subjected to time varying loads are performed. The material is assumed to obey the small strain J(2) flow theory of plasticity with rate independent behaviour. The results demonstrate that a valid J-Q field exists under dynamic loading irrespective of the crack length and specimen geometry. Further, the constraint parameter Q becomes strongly negative at high loading rates, particularly in deeply cracked specimens. The variation of dynamic fracture toughness K-dc with stress intensity rate K for cleavage cracking is predicted using a simple critical stress criterion. It is found that inertia-driven constraint loss can substantially enhance K-dc for (K) over dot > 10(5) MPa rootm/s.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this work, the effect of lattice orientation on the fields prevailing near a notch tip is investigated pertaining to various constraint levels in FCC single crystals. A modified boundary layer formulation is employed and numerical solutions under mode I, plane strain conditions are generated by assuming an elastic-perfectly plastic FCC single crystal. The analysis is carried out corresponding to different lattice orientations with respect to the notch line. It is found that the near-tip deformation field, especially the development of kink or slip shear bands is sensitive to the constraint level. The stress distribution and the size and shape of the plastic zone near the notch tip are also strongly influenced by the level of T-stress. The present results clearly establish that ductile single crystal fracture geometries would progressively lose crack tip constraint as the T-stress becomes more negative irrespective of lattice orientation. Also, the near-tip field for a range of constraint levels can be characterized by two-parameters such as K-T or J-Q as in isotropic plastic solids.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Building flexible constraint length Viterbi decoders requires us to be able to realize de Bruijn networks of various sizes on the physically provided interconnection network. This paper considers the case when the physical network is itself a de Bruijn network and presents a scalable technique for realizing any n-node de Bruijn network on an N-node de Bruijn network, where n < N. The technique ensures that the length of the longest path realized on the network is minimized and that each physical connection is utilized to send only one data item, both of which are desirable in order to reduce the hardware complexity of the network and to obtain the best possible performance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new technique named as model predictive spread acceleration guidance (MPSAG) is proposed in this paper. It combines nonlinear model predictive control and spread acceleration guidance philosophies. This technique is then used to design a nonlinear suboptimal guidance law for a constant speed missile against stationary target with impact angle constraint. MPSAG technique can be applied to a class of nonlinear problems, which leads to a closed form solution of the lateral acceleration (latax) history update. Guidance command assumed is the lateral acceleration (latax), applied normal to the velocity vector. The new guidance law is validated by considering the nonlinear kinematics with both lag-free as well as first order autopilot delay. The simulation results show that the proposed technique is quite promising to come up with a nonlinear guidance law that leads to both very small miss distance as well as the desired impact angle.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new technique named as model predictive spread acceleration guidance (MPSAG) is proposed in this paper. It combines nonlinear model predictive control and spread acceleration guidance philosophies. This technique is then used to design a nonlinear suboptimal guidance law for a constant speed missile against stationary target with impact angle constraint. MPSAG technique can be applied to a class of nonlinear problems, which leads to a closed form solution of the lateral acceleration (latax) history update. Guidance command assumed is the lateral acceleration (latax), applied normal to the velocity vector. The new guidance law is validated by considering the nonlinear kinematics with both lag-free as well as first order autopilot delay. The simulation results show that the proposed technique is quite promising to come up with a nonlinear guidance law that leads to both very small miss distance as well as the desired impact angle.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Using a recently developed method named as model predictive static programming (MPSP), a nonlinear suboptimal guidance law for a constant speed missile against a slow moving target with impact angle constraint is proposed. In this paper MPSP technique leads to a closed form solution of the latax history update for the given problem. Guidance command is the latax,which is normal to the missile velocity and the terminal constraints are miss distance and impact angle. The new guidance law is validated by considering the nonlinear kinematics with both lag-free and first order autopilot delay.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Fuzzy multiobjective programming for a deterministic case involves maximizing the minimum goal satisfaction level among conflicting goals of different stakeholders using Max-min approach. Uncertainty due to randomness in a fuzzy multiobjective programming may be addressed by modifying the constraints using probabilistic inequality (e.g., Chebyshev’s inequality) or by addition of new constraints using statistical moments (e.g., skewness). Such modifications may result in the reduction of the optimal value of the system performance. In the present study, a methodology is developed to allow some violation in the newly added and modified constraints, and then minimizing the violation of those constraints with the objective of maximizing the minimum goal satisfaction level. Fuzzy goal programming is used to solve the multiobjective model. The proposed methodology is demonstrated with an application in the field of Waste Load Allocation (WLA) in a river system.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article compares the land use in solar energy technologies with conventional energy sources. This has been done by introducing two parameters called land transformation and land occupation. It has been shown that the land area transformed by solar energy power generation is small compared to hydroelectric power generation, and is comparable with coal and nuclear energy power generation when life-cycle transformations are considered. We estimate that 0.97% of total land area or 3.1% of the total uncultivable land area of India would be required to generate 3400 TWh/yr from solar energy power systems in conjunction with other renewable energy sources.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Points-to analysis is a key compiler analysis. Several memory related optimizations use points-to information to improve their effectiveness. Points-to analysis is performed by building a constraint graph of pointer variables and dynamically updating it to propagate more and more points-to information across its subset edges. So far, the structure of the constraint graph has been only trivially exploited for efficient propagation of information, e.g., in identifying cyclic components or to propagate information in topological order. We perform a careful study of its structure and propose a new inclusion-based flow-insensitive context-sensitive points-to analysis algorithm based on the notion of dominant pointers. We also propose a new kind of pointer-equivalence based on dominant pointers which provides significantly more opportunities for reducing the number of pointers tracked during the analysis. Based on this hitherto unexplored form of pointer-equivalence, we develop a new context-sensitive flow-insensitive points-to analysis algorithm which uses incremental dominator update to efficiently compute points-to information. Using a large suite of programs consisting of SPEC 2000 benchmarks and five large open source programs we show that our points-to analysis is 88% faster than BDD-based Lazy Cycle Detection and 2x faster than Deep Propagation. We argue that our approach of detecting dominator-based pointer-equivalence is a key to improve points-to analysis efficiency.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Pervasive use of pointers in large-scale real-world applications continues to make points-to analysis an important optimization-enabler. Rapid growth of software systems demands a scalable pointer analysis algorithm. A typical inclusion-based points-to analysis iteratively evaluates constraints and computes a points-to solution until a fixpoint. In each iteration, (i) points-to information is propagated across directed edges in a constraint graph G and (ii) more edges are added by processing the points-to constraints. We observe that prioritizing the order in which the information is processed within each of the above two steps can lead to efficient execution of the points-to analysis. While earlier work in the literature focuses only on the propagation order, we argue that the other dimension, that is, prioritizing the constraint processing, can lead to even higher improvements on how fast the fixpoint of the points-to algorithm is reached. This becomes especially important as we prove that finding an optimal sequence for processing the points-to constraints is NP-Complete. The prioritization scheme proposed in this paper is general enough to be applied to any of the existing points-to analyses. Using the prioritization framework developed in this paper, we implement prioritized versions of Andersen's analysis, Deep Propagation, Hardekopf and Lin's Lazy Cycle Detection and Bloom Filter based points-to analysis. In each case, we report significant improvements in the analysis times (33%, 47%, 44%, 20% respectively) as well as the memory requirements for a large suite of programs, including SPEC 2000 benchmarks and five large open source programs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we study the diversity-multiplexing-gain tradeoff (DMT) of wireless relay networks under the half-duplex constraint. It is often unclear what penalty if any, is imposed by the half-duplex constraint on the DMT of such networks. We study two classes of networks; the first class, called KPP(I) networks, is the class of networks with the relays organized in K parallel paths between the source and the destination. While we assume that there is no direct source-destination path, the K relaying paths can interfere with each other. The second class, termed as layered networks, is comprised of relays organized in layers, where links exist only between adjacent layers. We present a communication scheme based on static schedules and amplify-and-forward relaying for these networks. We also show that for KPP(I) networks with K >= 3, the proposed schemes can achieve full-duplex DMT performance, thus demonstrating that there is no performance hit on the DMT due to the half-duplex constraint. We also show that, for layered networks, a linear DMT of d(max)(1 - r)(+) between the maximum diversity d(max) and the maximum MG, r(max) = 1 is achievable. We adapt existing DMT optimal coding schemes to these networks, thus specifying the end-to-end communication strategy explicitly.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Before installation, a voltage source converter is usually subjected to heat-run test to verify its thermal design and performance under load. For heat-run test, the converter needs to be operated at rated voltage and rated current for a substantial length of time. Hence, such tests consume huge amount of energy in case of high-power converters. Also, the capacities of the source and loads available in the research and development (R&D) centre or the production facility could be inadequate to conduct such tests. This paper proposes a method to conduct heat-run tests on high-power, pulse width modulated (PWM) converters with low energy consumption. The experimental set-up consists of the converter under test and another converter (of similar or higher rating), both connected in parallel on the ac side and open on the dc side. Vector-control or synchronous reference frame control is employed to control the converters such that one draws certain amount of reactive power and the other supplies the same; only the system losses are drawn from the mains. The performance of the controller is validated through simulation and experiments. Experimental results, pertaining to heat-run tests on a high-power PWM converter, are presented at power levels of 25 kVA to 150 kVA.