783 resultados para Performance evaluation
Resumo:
Current Internet transport protocols make end-to-end measurements and maintain per-connection state to regulate the use of shared network resources. When two or more such connections share a common endpoint, there is an opportunity to correlate the end-to-end measurements made by these protocols to better diagnose and control the use of shared resources. We develop packet probing techniques to determine whether a pair of connections experience shared congestion. Correct, efficient diagnoses could enable new techniques for aggregate congestion control, QoS admission control, connection scheduling and mirror site selection. Our extensive simulation results demonstrate that the conditional (Bayesian) probing approach we employ provides superior accuracy, converges faster, and tolerates a wider range of network conditions than recently proposed memoryless (Markovian) probing approaches for addressing this opportunity.
Resumo:
This thesis presents a framework for aggregated congestion management for TCP flows and shows how to integrate such an approach in an existing TCP protocol stack. The thesis presents an initial implementation of this congestion management scheme in Linux, with performance evaluation in ns as well.
Resumo:
Although cooperation generally increases the amount of resources available to a community of nodes, thus improving individual and collective performance, it also allows for the appearance of potential mistreatment problems through the exposition of one node's resources to others. We study such concerns by considering a group of independent, rational, self-aware nodes that cooperate using on-line caching algorithms, where the exposed resource is the storage at each node. Motivated by content networking applications -- including web caching, CDNs, and P2P -- this paper extends our previous work on the on-line version of the problem, which was conducted under a game-theoretic framework, and limited to object replication. We identify and investigate two causes of mistreatment: (1) cache state interactions (due to the cooperative servicing of requests) and (2) the adoption of a common scheme for cache management policies. Using analytic models, numerical solutions of these models, as well as simulation experiments, we show that on-line cooperation schemes using caching are fairly robust to mistreatment caused by state interactions. To appear in a substantial manner, the interaction through the exchange of miss-streams has to be very intense, making it feasible for the mistreated nodes to detect and react to exploitation. This robustness ceases to exist when nodes fetch and store objects in response to remote requests, i.e., when they operate as Level-2 caches (or proxies) for other nodes. Regarding mistreatment due to a common scheme, we show that this can easily take place when the "outlier" characteristics of some of the nodes get overlooked. This finding underscores the importance of allowing cooperative caching nodes the flexibility of choosing from a diverse set of schemes to fit the peculiarities of individual nodes. To that end, we outline an emulation-based framework for the development of mistreatment-resilient distributed selfish caching schemes. Our framework utilizes a simple control-theoretic approach to dynamically parameterize the cache management scheme. We show performance evaluation results that quantify the benefits from instantiating such a framework, which could be substantial under skewed demand profiles.
Resumo:
Emerging configurable infrastructures such as large-scale overlays and grids, distributed testbeds, and sensor networks comprise diverse sets of available computing resources (e.g., CPU and OS capabilities and memory constraints) and network conditions (e.g., link delay, bandwidth, loss rate, and jitter) whose characteristics are both complex and time-varying. At the same time, distributed applications to be deployed on these infrastructures exhibit increasingly complex constraints and requirements on resources they wish to utilize. Examples include selecting nodes and links to schedule an overlay multicast file transfer across the Grid, or embedding a network experiment with specific resource constraints in a distributed testbed such as PlanetLab. Thus, a common problem facing the efficient deployment of distributed applications on these infrastructures is that of "mapping" application-level requirements onto the network in such a manner that the requirements of the application are realized, assuming that the underlying characteristics of the network are known. We refer to this problem as the network embedding problem. In this paper, we propose a new approach to tackle this combinatorially-hard problem. Thanks to a number of heuristics, our approach greatly improves performance and scalability over previously existing techniques. It does so by pruning large portions of the search space without overlooking any valid embedding. We present a construction that allows a compact representation of candidate embeddings, which is maintained by carefully controlling the order via which candidate mappings are inserted and invalid mappings are removed. We present an implementation of our proposed technique, which we call NETEMBED – a service that identify feasible mappings of a virtual network configuration (the query network) to an existing real infrastructure or testbed (the hosting network). We present results of extensive performance evaluation experiments of NETEMBED using several combinations of real and synthetic network topologies. Our results show that our NETEMBED service is quite effective in identifying one (or all) possible embeddings for quite sizable queries and hosting networks – much larger than what any of the existing techniques or services are able to handle.
Resumo:
In professional sports there are in general three steps required to improve performance namely task definition, training and performance assessment. This process is iteratively repeated and feedback generated from quantitative performance measurement is in turn used for task redefinition. Task definition can be achieved in a number of ways including via video streaming or indeed and as is more common, by listening to coaching staff. However non-subjective performance evaluation is difficult due to the complexity of the movements involved. When considering the subset of sports where precision accuracy and repeatability are a necessity this problem becomes inherently more difficult to solve. Until recently sports such as martial arts, fencing and darts, where the smallest deviation from a prescribed movement goal can result in large outcome error, were deemed too difficult to characterise fully. Advances in technology, as illustrated by this study, now make this type of physiometry possible.
Resumo:
Dense deployment of wireless local area network (WLAN) access points (APs) is an important part of the next generation Wi-Fi and standardization (802.11ax) efforts are underway. Increasing demand for WLAN connectivity motivates such dense deployments, especially in geographical areas with large numbers of users, such as stadiums, large enterprises, multi-tenant buildings, and urban cities. Although densification of WLAN APs guarantees coverage, it is susceptible to increased interference and uncoordinated association of stations (STAs) to APs, which degrade network throughput. Therefore, to improve network throughput, algorithms are proposed in this thesis to optimally coordinate AP associations in the presence of interference. In essence, coordination of APs in dense WLANs (DWLANs) is achieved through coordination of STAs' associations with APs. While existing approaches suggest tuning of APs' beacon powers or using transmit power control (TPC) for association control, here, the signal-to-interference-plus-noise ratio (SINRs) of STAs and the clear channel assessment (CCA) threshold of the 802.11 MAC protocol are employed. The proposed algorithms in this thesis enhance throughput and minimize coverage holes inherent in cell breathing and TPC techniques by not altering the transmit powers of APs, which determine cell coverage. Besides uncoordinated AP associations, unnecessary frequent transmission deferment is envisaged as another problem in DWLANs due to the clear channel assessment aspect of the carrier sensing multiple access collision avoidance (CSMA/CA) scheme in 802.11 standards and the short spatial reuse distance between co-channel APs. To address this problem in addition to AP association coordination, an algorithm is proposed for CCA threshold adjustment in each AP cell, such that CCA threshold used in one cell mitigates transmission deferment in neighboring cells. Performance evaluation reveals that the proposed association optimization algorithms achieve significant gain in throughput when compared with the default strongest signal first (SSF) association scheme in the current 802.11 standard. Also, further gain in throughput is observed when the CCA threshold adjustment is combined with the optimized association. Results show that when STA-AP association is optimized and CCA threshold is adjusted in each cell, throughput improves. Finally, transmission delay and the number of packet re-transmissions due to collision and contention significantly decrease.
Resumo:
Performance evaluation of parallel software and architectural exploration of innovative hardware support face a common challenge with emerging manycore platforms: they are limited by the slow running time and the low accuracy of software simulators. Manycore FPGA prototypes are difficult to build, but they offer great rewards. Software running on such prototypes runs orders of magnitude faster than current simulators. Moreover, researchers gain significant architectural insight during the modeling process. We use the Formic FPGA prototyping board [1], which specifically targets scalable and cost-efficient multi-board prototyping, to build and test a 64-board model of a 512-core, MicroBlaze-based, non-coherent hardware prototype with a full network-on-chip in a 3D-mesh topology. We expand the hardware architecture to include the ARM Versatile Express platforms and build a 520-core heterogeneous prototype of 8 Cortex-A9 cores and 512 MicroBlaze cores. We then develop an MPI library for the prototype and evaluate it extensively using several bare-metal and MPI benchmarks. We find that our processor prototype is highly scalable, models faithfully single-chip multicore architectures, and is a very efficient platform for parallel programming research, being 50,000 times faster than software simulation.
Resumo:
This letter investigates performance enhancement by the concept of multi-carrier index keying in orthogonal frequency division multiplexing (OFDM) systems. For the performance evaluation, a tight closed-form approximation of the bit error rate (BER) is derived introducing the expression for the number of bit errors occurring in both the index domain and the complex domain, in the presence of both imperfect and perfect detection of active multi-carrier indices. The accuracy of the derived BER results for various cases are validated using simulations, which can provide accuracy within 1 dB at favorable channels.
Resumo:
In this paper, the impact of hardware impairments on secrecy performance of cognitive MIMO schemes is investigated. In addition, the relay which helps the source forward the source signal to the destination can operate either half-duplex mode or full-duplex mode. For performance evaluation, we give the expressions of average secrecy rate over Rayleigh fading channel. Monte-Carlo simulations are presented to compare and optimize the performance of the proposed schemes.
Resumo:
This paper describes middleware-level support for agent mobility, targeted at hierarchically structured wireless sensor and actuator network applications. Agent mobility enables a dynamic deployment and adaptation of the application on top of the wireless network at runtime, while allowing the middleware to optimize the placement of agents, e.g., to reduce wireless network traffic, transparently to the application programmer. The paper presents the design of the mechanisms and protocols employed to instantiate agents on nodes and to move agents between nodes. It also gives an evaluation of a middleware prototype running on Imote2 nodes that communicate over ZigBee. The results show that our implementation is reasonably efficient and fast enough to support the envisioned functionality on top of a commodity multi-hop wireless technology. Our work is to a large extent platform-neutral, thus it can inform the design of other systems that adopt a hierarchical structuring of mobile components. © 2012 ICST Institute for Computer Science, Social Informatics and Telecommunications Engineering.
Resumo:
Multicarrier Index Keying (MCIK) is a recently developed technique that modulates subcarriers but also indices of the subcarriers. In this paper a novel low-complexity detection scheme of subcarrier indices is proposed for an MCIK system and addresses a substantial reduction in complexity over the optimalmaximum likelihood (ML) detection. For the performance evaluation, a closed-form expression for the pairwise error probability (PEP) of an active subcarrier index, and a tight approximation of the average PEP of multiple subcarrier indices are derived in closed-form. The theoretical outcomes are validated usingsimulations, at a difference of less than 0.1dB. Compared to the optimal ML, the proposed detection achieves a substantial reduction in complexity with small loss in error performance (<= 0.6dB).
Resumo:
This paper presents a thorough performance analysis of dual-hop cognitive amplify-and-forward (AF) relaying networks under spectrum-sharing mechanism over independent non-identically distributed (i.n.i.d.) 􀀀 fading channels. In order to guarantee the quality-of-service (QoS) of primary networks, both maximum tolerable peak interference power Q at the primary users (PUs) and maximum allowable transmit power P at secondary users (SUs) are considered to constrain transmit power at the cognitive transmitters. For integer-valued fading parameters, a closed-form lower bound for the outage probability (OP) of the considered networks is obtained. Moreover, assuming arbitrary-valued fading parameters, the lower bound in integral form for the OP is derived. In order to obtain further insights on the OP performance, asymptotic expressions for the OP at high SNRs are derived, from which the diversity/coding gains and the diversity-multiplexing gain tradeoff (DMT) of the secondary network can be readily deduced. It is shown that the diversity gain and also the DMT are solely determined by the fading parameters of the secondary network whereas the primary network only affects the coding gain. The derived results include several others available in previously published works as special cases, such as those for Nakagami-m fading channels. In addition, performance evaluation results have been obtained by Monte Carlo computer simulations which have verified the accuracy of the theoretical analysis.