24 resultados para service performance


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The focus of this paper is to outline a method for consolidating and implementing the work on performance-based specification and testing. First part of the paper will review the mathematical significance of the variables used in common service life models. The aim is to identify a set of significant variables that influence the ingress of chloride ions into concrete. These variables are termed as Key Performance Indicators (KPI’s). This will also help to reduce the complexity of some of the service life models and make them more appealing for practicing engineers. The second part of the paper presents a plan for developing a database based on these KPI’s so that relationships can then be drawn between common concrete mix parameters and KPI’s. This will assist designers in specifying a concrete with adequate performance for a particular environment. This, collectively, is referred to as the KPI based approach and the concluding remarks will outline how the authors envisage the KPI theory to relate to performance assessment and monitoring.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Best concrete research paper by a student - Research has shown that the cost of managing structures puts high strain on the infrastructure budget, with
estimates of over 50% of the European construction budget being dedicated to repair and maintenance. If reinforced concrete
structures are not suitably designed and adequately maintained, their service life is compromised, resulting in the full economic
value of the investment not realised. The issue is more prevalent in coastal structures as a result of combinations of aggressive
actions, such as those caused by chlorides, sulphates and cyclic freezing and thawing.
It is a common practice nowadays to ensure durability of reinforced concrete structures by specifying a concrete mix and a
nominal cover at the design stage to cater for the exposure environment. This in theory should produce the performance required
to achieve a specified service life. Although the European Standard EN 206-1 specifies variations in the exposure environment,
it does not take into account the macro and micro climates surrounding structures, which have a significant influence on their
performance and service life. Therefore, in order to construct structures which will perform satisfactorily in different exposure
environments, the following two aspects need to be developed: a performance based specification to supplement EN 206-1
which will outline the expected performance of the structure in a given environment; and a simple yet transferrable procedure
for assessing the performance of structures in service termed KPI Theory. This will allow the asset managers not only to design
structures for the intended service life, but also to take informed maintenance decisions should the performance in service fall
short of what was specified. This paper aims to discuss this further.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present a mathematically rigorous Quality-of-Service (QoS) metric which relates the achievable quality of service metric (QoS) for a real-time analytics service to the server energy cost of offering the service. Using a new iso-QoS evaluation methodology, we scale server resources to meet QoS targets and directly rank the servers in terms of their energy-efficiency and by extension cost of ownership. Our metric and method are platform-independent and enable fair comparison of datacenter compute servers with significant architectural diversity, including micro-servers. We deploy our metric and methodology to compare three servers running financial option pricing workloads on real-life market data. We find that server ranking is sensitive to data inputs and desired QoS level and that although scale-out micro-servers can be up to two times more energy-efficient than conventional heavyweight servers for the same target QoS, they are still six times less energy efficient than high-performance computational accelerators.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Uncertainty profiles are used to study the effects of contention within cloud and service-based environments. An uncertainty profile provides a qualitative description of an environment whose quality of service (QoS) may fluctuate unpredictably. Uncertain environments are modelled by strategic games with two agents; a daemon is used to represent overload and high resource contention; an angel is used to represent an idealised resource allocation situation with no underlying contention. Assessments of uncertainty profiles are useful in two ways: firstly, they provide a broad understanding of how environmental stress can effect an application’s performance (and reliability); secondly, they allow the effects of introducing redundancy into a computation to be assessed

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A new niche of densely populated, unprotected networks is becoming more prevalent in public areas such as Shopping Malls, defined here as independent open-access networks, which have attributes that make attack detection more challenging than in typical enterprise networks. To address these challenges, new detection systems which do not rely on knowledge of internal device state are investigated here. This paper shows that this lack of state information requires an additional metric (The exchange timeout window) for detection of WLAN Denial of Service Probe Flood attacks. Variability in this metric has a significant influence on the ability of a detection system to reliably detect the presence of attacks. A parameter selection method is proposed which is shown to provide reliability and repeatability in attack detection in WLANs. Results obtained from ongoing live trials are presented that demonstrate the importance of accurately estimating probe request and probe response timeouts in future Independent Intrusion Detection Systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Rationale, aims and objectives: This study aimed to determine the value of using a mix of clinical pharmacy data and routine hospital admission spell data in the development of predictive algorithms. Exploration of risk factors in hospitalized patients, together with the targeting strategies devised, will enable the prioritization of clinical pharmacy services to optimize patient outcomes. 

Methods: Predictive algorithms were developed using a number of detailed steps using a 75% sample of integrated medicines management (IMM) patients, and validated using the remaining 25%. IMM patients receive targeted clinical pharmacy input throughout their hospital stay. The algorithms were applied to the validation sample, and predicted risk probability was generated for each patient from the coefficients. Risk threshold for the algorithms were determined by identifying the cut-off points of risk scores at which the algorithm would have the highest discriminative performance. Clinical pharmacy staffing levels were obtained from the pharmacy department staffing database. 

Results: Numbers of previous emergency admissions and admission medicines together with age-adjusted co-morbidity and diuretic receipt formed a 12-month post-discharge and/or readmission risk algorithm. Age-adjusted co-morbidity proved to be the best index to predict mortality. Increased numbers of clinical pharmacy staff at ward level was correlated with a reduction in risk-adjusted mortality index (RAMI). 

Conclusions: Algorithms created were valid in predicting risk of in-hospital and post-discharge mortality and risk of hospital readmission 3, 6 and 12 months post-discharge. The provision of ward-based clinical pharmacy services is a key component to reducing RAMI and enabling the full benefits of pharmacy input to patient care to be realized.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Outwith clinical trials, patient outcomes specifically related to SACT (systemic anti-cancer therapy) are not well reported despite a significant proportion of patients receiving active treatment at the end of life. The NCEPOD reviewing deaths within 30 days of SACT found SACT caused or hastened death in 27% of cases.

Method: Across the Northern Ireland cancer network, 95 patients who died within 30 days of SACT for solid tumours were discussed at the Morbidity and Mortality monthly meeting during 2013. Using a structured template, each case was independently reviewed, with particular focus on whether SACT caused or hastened death.

Results: Lung, GI and breast cancers were the most common sites. Performance status was recorded in 92% at time of final SACT cycle (ECOG PS 0-2 89%).

In 57% the cause of death was progressive disease. Other causes included thromboembolism (13%) and infection (5% neutropenic sepsis, 6% non-neutropenic sepsis). In 26% with death from progressive disease, the patient was first cycle of first line treatment for metastatic disease. In the majority discussion regarding treatment aims and risks was documented. Only one patient was receiving SACT with curative intent, who died from appropriately managed neutropenic sepsis.

A definitive decision regarding SACT's role in death was made in 60%: in 49% SACT was deemed non-contributory and in 11% SACT was deemed the cause of death. In 40% SACT did not play a major role, but a definitive negative association could not be made.

Conclusion: Development of a robust review process of 30-day mortality after SACT established a benchmark for SACT delivery for future comparisons and identified areas for SACT service organisation improvement. Moreover it encourages individual practice reflection and highlights the importance of balancing patients' needs and concerns with realistic outcomes and risks, particularly in heavily pre-treated patients or those of poor performance status.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

How can GPU acceleration be obtained as a service in a cluster? This question has become increasingly significant due to the inefficiency of installing GPUs on all nodes of a cluster. The research reported in this paper is motivated to address the above question by employing rCUDA (remote CUDA), a framework that facilitates Acceleration-as-a-Service (AaaS), such that the nodes of a cluster can request the acceleration of a set of remote GPUs on demand. The rCUDA framework exploits virtualisation and ensures that multiple nodes can share the same GPU. In this paper we test the feasibility of the rCUDA framework on a real-world application employed in the financial risk industry that can benefit from AaaS in the production setting. The results confirm the feasibility of rCUDA and highlight that rCUDA achieves similar performance compared to CUDA, provides consistent results, and more importantly, allows for a single application to benefit from all the GPUs available in the cluster without loosing efficiency.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The integral variability of raw materials, lack of awareness and appreciation of the technologies for achieving quality control and lack of appreciation of the micro and macro environmental conditions that the structures will be subjected, makes modern day concreting a challenge. This also makes Designers and Engineers adhere more closely to prescriptive standards developed for relatively less aggressive environments. The data from exposure sites and real structures prove, categorically, that the prescriptive specifications are inadequate for chloride environments. In light of this shortcoming, a more pragmatic approach would be to adopt performance-based specifications which are familiar to industry in the form of specification for mechanical strength. A recently completed RILEM technical committee made significant advances in making such an approach feasible.
Furthering a performance-based specification requires establishment of reliable laboratory and on-site test methods, as well as easy to perform service-life models. This article highlights both laboratory and on-site test methods for chloride diffusivity/electrical resistivity and the relationship between these tests for a range of concretes. Further, a performance-based approach using an on-site diffusivity test is outlined that can provide an easier to apply/adopt practice for Engineers and asset managers for specifying/testing concrete structures.