903 resultados para Access Network Performance


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Calibration process in micro-simulation is an extremely complicated phenomenon. The difficulties are more prevalent if the process encompasses fitting aggregate and disaggregate parameters e.g. travel time and headway. The current practice in calibration is more at aggregate level, for example travel time comparison. Such practices are popular to assess network performance. Though these applications are significant there is another stream of micro-simulated calibration, at disaggregate level. This study will focus on such micro-calibration exercise-key to better comprehend motorway traffic risk level, management of variable speed limit (VSL) and ramp metering (RM) techniques. Selected section of Pacific Motorway in Brisbane will be used as a case study. The discussion will primarily incorporate the critical issues encountered during parameter adjustment exercise (e.g. vehicular, driving behaviour) with reference to key traffic performance indicators like speed, land distribution and headway; at specific motorway points. The endeavour is to highlight the utility and implications of such disaggregate level simulation for improved traffic prediction studies. The aspects of calibrating for points in comparison to that for whole of the network will also be briefly addressed to examine the critical issues such as the suitability of local calibration at global scale. The paper will be of interest to transport professionals in Australia/New Zealand where micro-simulation in particular at point level, is still comparatively a less explored territory in motorway management.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper presents the benefits and issues related to travel time prediction on urban network. Travel time information quantifies congestion and is perhaps the most important network performance measure. Travel time prediction has been an active area of research for the last five decades. The activities related to ITS have increased the attention of researchers for better and accurate real-time prediction of travel time. Majority of the literature on travel time prediction is applicable to freeways where, under non-incident conditions, traffic flow is not affected by external factors such as traffic control signals and opposing traffic flows. On urban environment the problem is more complicated due to conflicting areas (intersections), mid-link sources and sinks etc. and needs to be addressed.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Just as telecommunications has played a key role in the global economy,1 high-speed broadband will have a significant role to play in the future of the digital economy. In particular high-speed broadband will have a role to play in the delivery of applications and services necessary for acquiring, and maintaining into the future Australia and Australians’ appropriate education level; community; health services, information provision and support; government services and engagement and participation by the public in the political process.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The Macroscopic Fundamental Diagram (MFD) relates space-mean density and flow, and the existence with dynamic features was confirmed in congested urban network in downtown Yokohama with real data set. Since the MFD represents the area-wide network traffic performances, studies on perimeter control strategies and an area traffic state estimation utilizing the MFD concept has been reported. However, limited works have been reported on real world example from signalised arterial network. This paper fuses data from multiple sources (Bluetooth, Loops and Signals) and develops a framework for the development of the MFD for Brisbane, Australia. Existence of the MFD in Brisbane arterial network is confirmed. Different MFDs (from whole network and several sub regions) are evaluated to discover the spatial partitioning in network performance representation. The findings confirmed the usefulness of appropriate network partitioning for traffic monitoring and incident detections. The discussion addressed future research directions

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The Macroscopic Fundamental Diagram (MFD) relates space-mean density and flow, and the existence with dynamic features was confirmed in congested urban network in downtown Yokohama with real data set. Since the MFD represents the area-wide network traffic performances, studies on perimeter control strategies and an area traffic state estimation utilizing the MFD concept has been reported. However, limited works have been reported on real world example from signalised arterial network. This paper fuses data from multiple sources (Bluetooth, Loops and Signals) and presents a framework for the development of the MFD for Brisbane, Australia. Existence of the MFD in Brisbane arterial network is confirmed. Different MFDs (from whole network and several sub regions) are evaluated to discover the spatial partitioning for network performance representation. The findings confirmed the usefulness of appropriate network partitioning for traffic monitoring and incident detections. The discussion addressed future research directions.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This thesis explored traffic characteristics at the aggregate level for area-wide traffic monitoring of large urban area. It focused on three aspects: understanding a macroscopic network performance under real-time traffic information provision, measuring traffic performance of a signalised arterial network using available data sets, and discussing network zoning for monitoring purposes in the case of Brisbane, Australia. This work presented the use of probe vehicle data for estimating traffic state variables, and illustrated dynamic features of regional traffic performance of Brisbane. The results confirmed the viability and effectiveness of area-wide traffic monitoring.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The ever expanding growth of the wireless access to the Internet in recent years has led to the proliferation of wireless and mobile devices to connect to the Internet. This has created the possibility of mobile devices equipped with multiple radio interfaces to connect to the Internet using any of several wireless access network technologies such as GPRS, WLAN and WiMAX in order to get the connectivity best suited for the application. These access networks are highly heterogeneous and they vary widely in their characteristics such as bandwidth, propagation delay and geographical coverage. The mechanism by which a mobile device switches between these access networks during an ongoing connection is referred to as vertical handoff and it often results in an abrupt and significant change in the access link characteristics. The most common Internet applications such as Web browsing and e-mail make use of the Transmission Control Protocol (TCP) as their transport protocol and the behaviour of TCP depends on the end-to-end path characteristics such as bandwidth and round-trip time (RTT). As the wireless access link is most likely the bottleneck of a TCP end-to-end path, the abrupt changes in the link characteristics due to a vertical handoff may affect TCP behaviour adversely degrading the performance of the application. The focus of this thesis is to study the effect of a vertical handoff on TCP behaviour and to propose algorithms that improve the handoff behaviour of TCP using cross-layer information about the changes in the access link characteristics. We begin this study by identifying the various problems of TCP due to a vertical handoff based on extensive simulation experiments. We use this study as a basis to develop cross-layer assisted TCP algorithms in handoff scenarios involving GPRS and WLAN access networks. We then extend the scope of the study by developing cross-layer assisted TCP algorithms in a broader context applicable to a wide range of bandwidth and delay changes during a handoff. And finally, the algorithms developed here are shown to be easily extendable to the multiple-TCP flow scenario. We evaluate the proposed algorithms by comparison with standard TCP (TCP SACK) and show that the proposed algorithms are effective in improving TCP behavior in vertical handoff involving a wide range of bandwidth and delay of the access networks. Our algorithms are easy to implement in real systems and they involve modifications to the TCP sender algorithm only. The proposed algorithms are conservative in nature and they do not adversely affect the performance of TCP in the absence of cross-layer information.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper optical code-division multiple-access (O-CDMA) packet network is considered, which offers inherent security in the access networks. Two types of random access protocols are proposed for packet transmission. In protocol 1, all distinct codes and in protocol 2, distinct codes as well as shifted versions of all these codes are used. O-CDMA network performance using optical orthogonal codes (OOCs) 1-D and two-dimensional (2-D) wavelength/time single-pulse-per-row (W/T SPR) codes are analyzed. The main advantage of using 2-D codes instead of one-dimensional (1-D) codes is to reduce the errors due to multiple access interference among different users. SPR codes are chosen as they have nearly ideal correlation properties. In this paper, correlation receiver is considered in the analysis. Using analytical model, we compare the OOC and SPR code performances in O-CDMA networks. We compute packet-success probability and throughput for both the types of codes. The analysis shows improved performance with SPR codes as compared to OOC codes.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper optical code-division multiple-access (O-CDMA) packet network is considered. Two types of random access protocols are proposed for packet transmission. In protocol 1, all distinct codes and in protocol 2, distinct codes as well as shifted versions of all these codes are used. O-CDMA network performance using optical orthogonal codes (OOCs) 1-D and twodimensional (2-D) wavelength/time single-pulse-per-row (W/TSPR) codes are analyzed. The main advantage of using 2-D codes instead of one-dimensional (1-D) codes is to reduce the errors due to multiple access interference among different users. In this paper, correlation receiver is considered in the analysis. Using analytical model, we compute and compare packet-success probability for 1-D and 2-D codes in an O-CDMA network and the analysis shows improved performance with 2-D codes as compared to 1-D codes.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The key requirements for enabling real-time remote healthcare service on a mobile platform, in the present day heterogeneous wireless access network environment, are uninterrupted and continuous access to the online patient vital medical data, monitor the physical condition of the patient through video streaming, and so on. For an application, this continuity has to be sufficiently transparent both from a performance perspective as well as a Quality of Experience (QoE) perspective. While mobility protocols (MIPv6, HIP, SCTP, DSMIP, PMIP, and SIP) strive to provide both and do so, limited or non-availability (deployment) of these protocols on provider networks and server side infrastructure has impeded adoption of mobility on end user platforms. Add to this, the cumbersome OS configuration procedures required to enable mobility protocol support on end user devices and the user's enthusiasm to add this support is lost. Considering the lack of proper mobility implementations that meet the remote healthcare requirements above, we propose SeaMo+ that comprises a light-weight application layer framework, termed as the Virtual Real-time Multimedia Service (VRMS) for mobile devices to provide an uninterrupted real-time multimedia information access to the mobile user. VRMS is easy to configure, platform independent, and does not require additional network infrastructure unlike other existing schemes. We illustrate the working of SeaMo+ in two realistic remote patient monitoring application scenarios.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

O câncer de colo de útero é o terceiro tipo de neoplasia mais comum em mulheres no estado do Rio de Janeiro, perdendo apenas para câncer de mama e pulmão. Não há informação disponível sobre a qualidade dos exames utilizados para prevenção do câncer do colo do útero nos municípios fluminenses. O trabalho teve como objetivo avaliar a distribuição geográfica das unidades laboratoriais e a cobertura de exames e a performance da rede de laboratórios de citopatologia do estado do Rio de Janeiro, segundo regiões mesopolitanas e municípios, no que diz respeito aos exames citopatológicos realizados no âmbito do SUS, considerando a qualidade, a normalização dos procedimentos e a gestão do laboratório e propor melhorias. A pesquisa usa um desenho de estudo do tipo transversal incluindo dados relativos ao funcionamento dos laboratórios que fazem exame papanicolaou no âmbito do SUS no estado do Rio de Janeiro. Trata-se de um estudo com dados secundários, já coletados para atender a ação de avaliação dos laboratórios de citopatologia no âmbito do SUS, nesse trabalho também é usada a abordagem ecológica para estimar a cobertura dos exames em relação à população-alvo do programa de rastreamento de câncer de colo de útero. A rede de laboratórios de citopatologia do estado do Rio de Janeiro não teve desempenho satisfatório. As regiões que tiveram laboratórios com melhor desempenho foram Baia da Ilha Grande e Baixada Litorânea e aquelas com laboratórios de pior desempenho foram Médio Paraíba e Noroeste. Os critérios avaliados com melhor desempenho foram da dimensão qualidade e o mais fraco desempenho foi observado para os critérios da dimensão normalização. A dimensão de gestão de laboratórios teve desempenho regular. Um relevante achado desse estudo foi a insuficiente qualidade da leitura de lâminas, inclusive procedimentos de releitura, que pode ser explicado pela suposta falta de capacitação dos profissionais em todo o processo exigido desde a identificação, fixação, e formas de encaminhamento do material até a chegada aos laboratórios. Com relação à cobertura, alguns municípios se aproximam do parâmetro (0,30) porém, embora este dado revele a capacidade da rede estadual do rio de Janeiro de ofertar exames, é preciso que ele seja analisado em conjunto com as situações de citologia anterior e tempo da citologia anterior para verificação da periodicidade da oferta e o melhor dimensionamento do alcance da população alvo assim como a abrangência da rede laboratorial de cada município. Conclusão: Os dados mostram repetição desnecessária de exames citopatológicos, o que implica custos injustificados e uma situação ainda mais deficitária de alcance das ações do que tem revelado o indicador razão. Ocorrendo principalmente em regiões com predominância de laboratórios privados. A expansão da cobertura com base na periodicidade recomendada do exame é relevante no quadro estadual encontrado e deve vir associada a iniciativas que garantam a qualidade no processo de coleta e análise do material, bem como a adequada capacitação dos profissionais para adoção de condutas recomendadas para as lesões identificadas.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The paper describes different research concepts of which the combination leads to an improvement of the power consumption per subscriber in a wireline access network by 10× and the energy efficiency per transferred bit by 100×. © 2012 IEEE.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Recent measurements of local-area and wide-area traffic have shown that network traffic exhibits variability at a wide range of scales self-similarity. In this paper, we examine a mechanism that gives rise to self-similar network traffic and present some of its performance implications. The mechanism we study is the transfer of files or messages whose size is drawn from a heavy-tailed distribution. We examine its effects through detailed transport-level simulations of multiple TCP streams in an internetwork. First, we show that in a "realistic" client/server network environment i.e., one with bounded resources and coupling among traffic sources competing for resources the degree to which file sizes are heavy-tailed can directly determine the degree of traffic self-similarity at the link level. We show that this causal relationship is not significantly affected by changes in network resources (bottleneck bandwidth and buffer capacity), network topology, the influence of cross-traffic, or the distribution of interarrival times. Second, we show that properties of the transport layer play an important role in preserving and modulating this relationship. In particular, the reliable transmission and flow control mechanisms of TCP (Reno, Tahoe, or Vegas) serve to maintain the long-range dependency structure induced by heavy-tailed file size distributions. In contrast, if a non-flow-controlled and unreliable (UDP-based) transport protocol is used, the resulting traffic shows little self-similar characteristics: although still bursty at short time scales, it has little long-range dependence. If flow-controlled, unreliable transport is employed, the degree of traffic self-similarity is positively correlated with the degree of throttling at the source. Third, in exploring the relationship between file sizes, transport protocols, and self-similarity, we are also able to show some of the performance implications of self-similarity. We present data on the relationship between traffic self-similarity and network performance as captured by performance measures including packet loss rate, retransmission rate, and queueing delay. Increased self-similarity, as expected, results in degradation of performance. Queueing delay, in particular, exhibits a drastic increase with increasing self-similarity. Throughput-related measures such as packet loss and retransmission rate, however, increase only gradually with increasing traffic self-similarity as long as reliable, flow-controlled transport protocol is used.