774 resultados para Paper addresses
Resumo:
The paper addresses certain issues pertaining to the technology of lime-stabilised steam-cured blocks used for masonry construction. Properties of lime-stabilised steam-cured blocks using expansive soils and tank bed soils have been examined. Influence of parameters like steam curing period, lime content and fly ash content on wet strength of blocks is studied. Steam curing of lime stabilised blocks at 80degreesC for about 20 hours at atmospheric pressure leads to considerably higher strengths when compared with curing under wet cloth at ambient temperatures. Clay-fly ash fractions of the mix control the optimum lime content yielding maximum strength. Long-term strength behaviour of steam-cured blocks has been monitored. The results indicate a favourable lime-clay ratio for stable long-term strength. A small-scale steam cured block production system has been designed and implemented to construct a load bearing masonry structure, thus demonstrating the potential of steam-cured block as a material for masonry construction.
Resumo:
This paper addresses the behaviour of compacted expansive soils under swell-shrink cycles. Laboratory cyclic swell-shrink tests were conducted on compacted specimens of two expansive soils at surcharge pressures of 6.25, 50.00, and 100.00 kPa. The void ratio and water content of the specimens at several intermediate stages during swelling until the end of swelling and during shrinkage until the end of shrinkage were determined to trace the water content versus void ratio paths with an increasing number of swell-shrink cycles. The test results showed that the swell-shrink path was reversible once the soil reached an equilibrium stage where the vertical deformations during swelling and shrinkage were the same. This usually occurred after about four swell-shrink cycles. The swelling and shrinkage path of each specimen subjected to full swelling - full shrinkage cycles showed an S-shaped curve (two curvilinear portions and a linear portion). However, the swelling and shrinkage path occurred as a part of the S-shaped curve, when the specimen was subjected to full swelling - partial shrinkage cycles. More than 80% of the total volumetric change and more than 50% of the total vertical deformation occurred in the central linear portion of the S-shaped curve. The volumetric change was essentially parallel to the saturation line within a degree of saturation range of 50-80% for the equilibrium cycle. The primary value of the swell-shrink path is to provide information regarding the void ratio change that would occur for a given change in water content for any possible swell-shrink pattern. It is suggested that these swell-shrink paths can be established with a limited number of tests in the laboratory.
Resumo:
Sensor network applications such as environmental monitoring demand that the data collection process be carried out for the longest possible time. Our paper addresses this problem by presenting a routing scheme that ensures that the monitoring network remains connected and hence the live sensor nodes deliver data for a longer duration. We analyze the role of relay nodes (neighbours of the base-station) in maintaining network connectivity and present a routing strategy that, for a particular class of networks, approaches the optimal as the set of relay nodes becomes larger. We then use these findings to develop an appropriate distributed routing protocol using potential-based routing. The basic idea of potential-based routing is to define a (scalar) potential value at each node in the network and forward data to the neighbor with the highest potential. We propose a potential function and evaluate its performance through simulations. The results show that our approach performs better than the well known lifetime maximization policy proposed by Chang and Tassiulas (2004), as well as AODV [Adhoc on demand distance vector routing] proposed by Perkins (1997).
Resumo:
This paper addresses the problem of secure path key establishment in wireless sensor networks that uses the random key pre-distribution technique. Inspired by the recent proxy-based scheme in the work of Ling and Znati (2005) and Li et al. (2005), we introduce a friend-based scheme for establishing pairwise keys securely. We show that the chances of finding friends in a neighbourhood are considerably more than that of finding proxies, leading to lower communication overhead. Further, we prove that the friend-based scheme performs better than the proxy-based scheme both in terms of resilience against node capture as well as in energy consumption for pairwise key establishment, making our scheme more feasible.
Resumo:
Statistical information about the wireless channel can be used at the transmitter side to enhance the performance of MIMO systems. This paper addresses how the concept of channel precoding can be used to enhance the performance of STBCs from Generalized Pseudo Orthogonal Designs which were first introduced by Zhu and Jafarkhani. Such designs include some important classes of STBCs that are directly derivable from Quasi-Orthogonal Designs and Co-ordinate Interleaved Orthogonal Designs.
Resumo:
This paper addresses the problem of maximum margin classification given the moments of class conditional densities and the false positive and false negative error rates. Using Chebyshev inequalities, the problem can be posed as a second order cone programming problem. The dual of the formulation leads to a geometric optimization problem, that of computing the distance between two ellipsoids, which is solved by an iterative algorithm. The formulation is extended to non-linear classifiers using kernel methods. The resultant classifiers are applied to the case of classification of unbalanced datasets with asymmetric costs for misclassification. Experimental results on benchmark datasets show the efficacy of the proposed method.
Resumo:
This paper addresses the problem of multiagent search in an unknown environment. The agents are autonomous in nature and are equipped with necessary sensors to carry out the search operation. The uncertainty, or lack of information about the search area is known a priori as a probability density function. The agents are deployed in an optimal way so as to maximize the one step uncertainty reduction. The agents continue to deploy themselves and reduce uncertainty till the uncertainty density is reduced over the search space below a minimum acceptable level. It has been shown, using LaSalle’s invariance principle, that a distributed control law which moves each of the agents towards the centroid of its Voronoi partition, modified by the sensor range leads to single step optimal deployment. This principle is now used to devise search trajectories for the agents. The simulations were carried out in 2D space with saturation on speeds of the agents. The results show that the control strategy per step indeed moves the agents to the respective centroid and the algorithm reduces the uncertainty distribution to the required level within a few steps.
Resumo:
This paper addresses a search problem with multiple limited capability search agents in a partially connected dynamical networked environment under different information structures. A self assessment-based decision-making scheme for multiple agents is proposed that uses a modified negotiation scheme with low communication overheads. The scheme has attractive features of fast decision-making and scalability to large number of agents without increasing the complexity of the algorithm. Two models of the self assessment schemes are developed to study the effect of increase in information exchange during decision-making. Some analytical results on the maximum number of self assessment cycles, effect of increasing communication range, completeness of the algorithm, lower bound and upper bound on the search time are also obtained. The performance of the various self assessment schemes in terms of total uncertainty reduction in the search region, using different information structures is studied. It is shown that the communication requirement for self assessment scheme is almost half of the negotiation schemes and its performance is close to the optimal solution. Comparisons with different sequential search schemes are also carried out. Note to Practitioners-In the futuristic military and civilian applications such as search and rescue, surveillance, patrol, oil spill, etc., a swarm of UAVs can be deployed to carry out the mission for information collection. These UAVs have limited sensor and communication ranges. In order to enhance the performance of the mission and to complete the mission quickly, cooperation between UAVs is important. Designing cooperative search strategies for multiple UAVs with these constraints is a difficult task. Apart from this, another requirement in the hostile territory is to minimize communication while making decisions. This adds further complexity to the decision-making algorithms. In this paper, a self-assessment-based decision-making scheme, for multiple UAVs performing a search mission, is proposed. The agents make their decisions based on the information acquired through their sensors and by cooperation with neighbors. The complexity of the decision-making scheme is very low. It can arrive at decisions fast with low communication overheads, while accommodating various information structures used for increasing the fidelity of the uncertainty maps. Theoretical results proving completeness of the algorithm and the lower and upper bounds on the search time are also provided.
Resumo:
Mobile ad-hoc networks (MANETs) have recently drawn significant research attention since they offer unique benefits and versatility with respect to bandwidth spatial reuse, intrinsic fault tolerance, and low-cost rapid deployment. This paper addresses the issue of delay sensitive realtime data transport in these type of networks. An effective QoS mechanism is thereby required for the speedy transport of the realtime data. QoS issue in MANET is an open-end problem. Various QoS measures are incorporated in the upperlayers of the network, but a few techniques addresses QoS techniques in the MAC layer. There are quite a few QoS techniques in the MAC layer for the infrastructure based wireless network. The goal and the challenge is to achieve a QoS delivery and a priority access to the real time traffic in adhoc wireless environment, while maintaining democracy in the resource allocation. We propose a MAC layer protocol called "FCP based FAMA protocol", which allocates the channel resources to the needy in a more democratic way, by examining the requirements, malicious behavior and genuineness of the request. We have simulated both the FAMA as well as FCP based FAMA and tested in various MANET conditions. Simulated results have clearly shown a performance improvement in the channel utilization and a decrease in the delay parameters in the later case. Our new protocol outperforms the other QoS aware MAC layer protocols.
Resumo:
This paper addresses the problem of how to select the optimal number of sensors and how to determine their placement in a given monitored area for multimedia surveillance systems. We propose to solve this problem by obtaining a novel performance metric in terms of a probability measure for accomplishing the task as a function of set of sensors and their placement. This measure is then used to find the optimal set. The same measure can be used to analyze the degradation in system 's performance with respect to the failure of various sensors. We also build a surveillance system using the optimal set of sensors obtained based on the proposed design methodology. Experimental results show the effectiveness of the proposed design methodology in selecting the optimal set of sensors and their placement.
Resumo:
Superscalar processors currently have the potential to fetch multiple basic blocks per cycle by employing one of several recently proposed instruction fetch mechanisms. However, this increased fetch bandwidth cannot be exploited unless pipeline stages further downstream correspondingly improve. In particular,register renaming a large number of instructions per cycle is diDcult. A large instruction window, needed to receive multiple basic blocks per cycle, will slow down dependence resolution and instruction issue. This paper addresses these and related issues by proposing (i) partitioning of the instruction window into multiple blocks, each holding a dynamic code sequence; (ii) logical partitioning of the registerjle into a global file and several local jles, the latter holding registers local to a dynamic code sequence; (iii) the dynamic recording and reuse of register renaming information for registers local to a dynamic code sequence. Performance studies show these mechanisms improve performance over traditional superscalar processors by factors ranging from 1.5 to a little over 3 for the SPEC Integer programs. Next, it is observed that several of the loops in the benchmarks display vector-like behavior during execution, even if the static loop bodies are likely complex for compile-time vectorization. A dynamic loop vectorization mechanism that builds on top of the above mechanisms is briefly outlined. The mechanism vectorizes up to 60% of the dynamic instructions for some programs, albeit the average number of iterations per loop is quite small.
Resumo:
The quest for prosperity has been a central motive in the life of man from the moment of his entrance into the worldly scence. And certain, it is that the issue of prosperity at the very cost of his very existence has mounted in intensity and urgency with the unforeseen evolution of industrialization. The traditional paradigm of prosperity has been resting on the classical theory of production economics.With increasing empiricism it is obvious that the rational model fail to grapple with the complexity of the concept. The paper addresses the prosperity as an goal state resting on the conviction of harmony between present generation and generation of humans to come. Sustainable prosperity involves more than growth in services/goods. It requires a change in the content of the growth, to make it less material-and energy-intensive and more equitable in its impact. The process of economic prosperity must be more soundly based upon the realities of the stock of capital that sustains it.
Resumo:
Traffic Engineering has been the prime concern for Internet Service Providers (ISPs), with the main focus being minimization of over-utilization of network capacity even though additional capacity is available which is under-utilized, Furthermore, requirements of timely delivery of digitized audiovisual information raises a new challenge of finding a path meeting these requirements. This paper addresses the issue of (a) distributing load to achieve global efficiency in resource utilization. (b) Finding a path satisfying the real time requirements of, delay and bandwidth requested by the applications. In this paper we do a critical study of the link utilization that varies over time and determine the time interval during which the link occupancy remains constant across days. This information helps in pre-determining link utilization that is useful in balancing load in the network Finally, we run simulations that use a dynamic time interval for profiling traffic and show improvement in terms number of calls admitted/blocked.
Resumo:
This paper addresses the problem of curtailing the number of control actions using fuzzy expert approach for voltage/reactive power dispatch. It presents an approach using fuzzy set theory for reactive power control with the purpose of improving the voltage profile of a power system. To minimize the voltage deviations from pre-desired values of all the load buses, using the sensitivities with respect to reactive power control variables form the basis of the proposed Fuzzy Logic Control (FLC). Control variables considered are switchable VAR compensators, On Load Tap Changing (OLTC) transformers and generator excitations. Voltage deviations and controlling variables are translated into fuzzy set notations to formulate the relation between voltage deviations and controlling ability of controlling devices. The developed fuzzy system is tested on a few simulated practical Indian power systems and modified IEEE-30 bus system. The performance of the fuzzy system is compared with conventional optimization technique and results obtained are encouraging. Results obtained for a modified IEEE-30 bus test system and a 205-node equivalent EHV system a part of Indian southern grid are presented for illustration purposes. The proposed fuzzy-expert technique is found suitable for on-line applications in energy control centre as the solution is obtained fast with significant speedups with few number of controllers.
Resumo:
This paper addresses the problem of multiple unmanned aerial vehicle (UAV) rendezvous when the UAVs have to perform maneuvers to avoid collisions with other UAVs. The proposed solution consists of using velocity control and a wandering maneuver, if needed, of the UAVs based on a consensus among them on the estimated time of arrival at the point of the rendezvous. This algorithm, with a slight modification is shown to be useful in tracking stationary or slowly moving targets with a standoff distance. The proposed algorithm is simple and computationally efficient. The simulation results demonstrate the efficacy of the proposed approach. DOI: 10.1061/(ASCE)AS.1943-5525.0000145. (C) 2012 American Society of Civil Engineers.