919 resultados para vibro-impact system
Resumo:
In this paper ensembles of forecasts (of up to six hours) are studied from a convection-permitting model with a representation of model error due to unresolved processes. The ensemble prediction system (EPS) used is an experimental convection-permitting version of the UK Met Office’s 24- member Global and Regional Ensemble Prediction System (MOGREPS). The method of representing model error variability, which perturbs parameters within the model’s parameterisation schemes, has been modified and we investigate the impact of applying this scheme in different ways. These are: a control ensemble where all ensemble members have the same parameter values; an ensemble where the parameters are different between members, but fixed in time; and ensembles where the parameters are updated randomly every 30 or 60 min. The choice of parameters and their ranges of variability have been determined from expert opinion and parameter sensitivity tests. A case of frontal rain over the southern UK has been chosen, which has a multi-banded rainfall structure. The consequences of including model error variability in the case studied are mixed and are summarised as follows. The multiple banding, evident in the radar, is not captured for any single member. However, the single band is positioned in some members where a secondary band is present in the radar. This is found for all ensembles studied. Adding model error variability with fixed parameters in time does increase the ensemble spread for near-surface variables like wind and temperature, but can actually decrease the spread of the rainfall. Perturbing the parameters periodically throughout the forecast does not further increase the spread and exhibits “jumpiness” in the spread at times when the parameters are perturbed. Adding model error variability gives an improvement in forecast skill after the first 2–3 h of the forecast for near-surface temperature and relative humidity. For precipitation skill scores, adding model error variability has the effect of improving the skill in the first 1–2 h of the forecast, but then of reducing the skill after that. Complementary experiments were performed where the only difference between members was the set of parameter values (i.e. no initial condition variability). The resulting spread was found to be significantly less than the spread from initial condition variability alone.
Resumo:
We present a benchmark system for global vegetation models. This system provides a quantitative evaluation of multiple simulated vegetation properties, including primary production; seasonal net ecosystem production; vegetation cover; composition and height; fire regime; and runoff. The benchmarks are derived from remotely sensed gridded datasets and site-based observations. The datasets allow comparisons of annual average conditions and seasonal and inter-annual variability, and they allow the impact of spatial and temporal biases in means and variability to be assessed separately. Specifically designed metrics quantify model performance for each process, and are compared to scores based on the temporal or spatial mean value of the observations and a "random" model produced by bootstrap resampling of the observations. The benchmark system is applied to three models: a simple light-use efficiency and water-balance model (the Simple Diagnostic Biosphere Model: SDBM), the Lund-Potsdam-Jena (LPJ) and Land Processes and eXchanges (LPX) dynamic global vegetation models (DGVMs). In general, the SDBM performs better than either of the DGVMs. It reproduces independent measurements of net primary production (NPP) but underestimates the amplitude of the observed CO2 seasonal cycle. The two DGVMs show little difference for most benchmarks (including the inter-annual variability in the growth rate and seasonal cycle of atmospheric CO2), but LPX represents burnt fraction demonstrably more accurately. Benchmarking also identified several weaknesses common to both DGVMs. The benchmarking system provides a quantitative approach for evaluating how adequately processes are represented in a model, identifying errors and biases, tracking improvements in performance through model development, and discriminating among models. Adoption of such a system would do much to improve confidence in terrestrial model predictions of climate change impacts and feedbacks.
Resumo:
The aim of this paper is to develop a comprehensive taxonomy of green supply chain management (GSCM) practices and develop a structural equation modelling-driven decision support system following GSCM taxonomy for managers to provide better understanding of the complex relationship between the external and internal factors and GSCM operational practices. Typology and/or taxonomy play a key role in the development of social science theories. The current taxonomies focus on a single or limited component of the supply chain. Furthermore, they have not been tested using different sample compositions and contexts, yet replication is a prerequisite for developing robust concepts and theories. In this paper, we empirically replicate one such taxonomy extending the original study by (a) developing broad (containing the key components of supply chain) taxonomy; (b) broadening the sample by including a wider range of sectors and organisational size; and (c) broadening the geographic scope of the previous studies. Moreover, we include both objective measures and subjective attitudinal measurements. We use a robust two-stage cluster analysis to develop our GSCM taxonomy. The main finding validates the taxonomy previously proposed and identifies size, attitude and level of environmental risk and impact as key mediators between internal drivers, external drivers and GSCM operational practices.
Resumo:
Catastrophe risk models used by the insurance industry are likely subject to significant uncertainty, but due to their proprietary nature and strict licensing conditions they are not available for experimentation. In addition, even if such experiments were conducted, these would not be repeatable by other researchers because commercial confidentiality issues prevent the details of proprietary catastrophe model structures from being described in public domain documents. However, such experimentation is urgently required to improve decision making in both insurance and reinsurance markets. In this paper we therefore construct our own catastrophe risk model for flooding in Dublin, Ireland, in order to assess the impact of typical precipitation data uncertainty on loss predictions. As we consider only a city region rather than a whole territory and have access to detailed data and computing resources typically unavailable to industry modellers, our model is significantly more detailed than most commercial products. The model consists of four components, a stochastic rainfall module, a hydrological and hydraulic flood hazard module, a vulnerability module, and a financial loss module. Using these we undertake a series of simulations to test the impact of driving the stochastic event generator with four different rainfall data sets: ground gauge data, gauge-corrected rainfall radar, meteorological reanalysis data (European Centre for Medium-Range Weather Forecasts Reanalysis-Interim; ERA-Interim) and a satellite rainfall product (The Climate Prediction Center morphing method; CMORPH). Catastrophe models are unusual because they use the upper three components of the modelling chain to generate a large synthetic database of unobserved and severe loss-driving events for which estimated losses are calculated. We find the loss estimates to be more sensitive to uncertainties propagated from the driving precipitation data sets than to other uncertainties in the hazard and vulnerability modules, suggesting that the range of uncertainty within catastrophe model structures may be greater than commonly believed.
Resumo:
Despite significant progress in climate impacts research, the narratives that science can presently piece together of a 2-, 3-, 4-, or 5-degree warmer world remain fragmentary. Here we briefly review past undertakings to comprehensively characterize and quantify climate impacts based on multi-model approaches. We then report on the Inter-Sectoral Impact Model Intercomparison Project (ISI-MIP), a community-driven effort to systematically compare impacts models across sectors and scales, and to quantify the uncertainties along the chain from greenhouse gas emissions and climate input data to the modelling of climate impacts themselves. We show how ISI-MIP and similar efforts can substantially advance the science relevant to impacts, adaptation and vulnerability, and we outline the steps that need to be taken in order to make the most of available modelling tools. We discuss pertinent limitations of these methods and how they could be tackled. We argue that it is time to consolidate the current patchwork of impacts knowledge through integrated cross-sectoral assessments, and that the climate impacts community is now in a favourable position to do so.
Resumo:
Dynamic electricity pricing can produce efficiency gains in the electricity sector and help achieve energy policy goals such as increasing electric system reliability and supporting renewable energy deployment. Retail electric companies can offer dynamic pricing to residential electricity customers via smart meter-enabled tariffs that proxy the cost to procure electricity on the wholesale market. Current investments in the smart metering necessary to implement dynamic tariffs show policy makers’ resolve for enabling responsive demand and realizing its benefits. However, despite these benefits and the potential bill savings these tariffs can offer, adoption among residential customers remains at low levels. Using a choice experiment approach, this paper seeks to determine whether disclosing the environmental and system benefits of dynamic tariffs to residential customers can increase adoption. Although sampling and design issues preclude wide generalization, we found that our environmentally conscious respondents reduced their required discount to switch to dynamic tariffs around 10% in response to higher awareness of environmental and system benefits. The perception that shifting usage is easy to do also had a significant impact, indicating the potential importance of enabling technology. Perhaps the targeted communication strategy employed by this study is one way to increase adoption and achieve policy goals.
Resumo:
Atmospheric methane concentrations decreased during the early to middle Holocene; however, the governing mechanisms remain controversial. Although it has been suggested that the mid-Holocene minimum methane emissions are associated with hydrological change, direct evidence is lacking. Here we report a new independent approach, linking hydrological change in peat sediments from the Tibetan Plateau to changes in archaeal diether concentrations and diploptene delta C-13 values as tracers for methanogenesis and methanotrophy, respectively. A minimum in inferred methanogenesis occurred during the mid-Holocene, which, locally, corresponds with the driest conditions of the Holocene, reflecting a minimum in Asian monsoon precipitation. The close coupling between precipitation and methanogenesis is validated by climate simulations, which also suggest a regionally widespread impact. Importantly, the minimum in methanogenesis is associated with a maximum in methanotrophy. Therefore, methane emissions in the Tibetan Plateau region were apparently lower during the mid-Holocene and partially controlled by interactions of large-scale atmospheric circulation.
Resumo:
Personalised nutrition (PN) has the potential to reduce disease risk and optimise health and performance. Although previous research has shown good acceptance of the concept of PN in the UK, preferences regarding the delivery of a PN service (e.g. online v. face-to-face) are not fully understood. It is anticipated that the presence of a free at point of delivery healthcare system, the National Health Service (NHS), in the UK may have an impact on end-user preferences for deliverances. To determine this, supplementary analysis of qualitative data obtained from focus group discussions on PN service delivery, collected as part of the Food4Me project in the UK and Ireland, was undertaken. Irish data provided comparative analysis of a healthcare system that is not provided free of charge at the point of delivery to the entire population. Analyses were conducted using the 'framework approach' described by Rabiee (Focus-group interview and data analysis. Proc Nutr Soc 63, 655-660). There was a preference for services to be led by the government and delivered face-to-face, which was perceived to increase trust and transparency, and add value. Both countries associated paying for nutritional advice with increased commitment and motivation to follow guidelines. Contrary to Ireland, however, and despite the perceived benefit of paying, UK discussants still expected PN services to be delivered free of charge by the NHS. Consideration of this unique challenge of free healthcare that is embedded in the NHS culture will be crucial when introducing PN to the UK.
Resumo:
This report provides case studies of Early Warning Systems (EWSs) and risk assessments encompassing three main hazard types: drought; flood and cyclone. The case studies are taken from ten countries across three continents (focusing on Africa, South Asia and the Caribbean). The case studies have been developed to assist the UK Department for International Development (DFID) to prioritise areas for Early Warning System (EWS) related research under their ‘Science for Humanitarian Emergencies and Resilience’ (SHEAR) programme. The aim of these case studies is to ensure that DFID SHEAR research is informed by the views of Non-Governmental Organisations (NGOs) and communities engaged with Early Warning Systems and risk assessments (including community-based Early Warning Systems). The case studies highlight a number of challenges facing Early Warning Systems (EWSs). These challenges relate to financing; integration; responsibilities; community interpretation; politics; dissemination; accuracy; capacity and focus. The case studies summarise a number of priority areas for EWS related research: • Priority 1: Contextualising and localising early warning information • Priority 2: Climate proofing current EWSs • Priority 3: How best to sustain effective EWSs between hazard events? • Priority 4: Optimising the dissemination of risk and warning information • Priority 5: Governance and financing of EWSs • Priority 6: How to support EWSs under challenging circumstances • Priority 7: Improving EWSs through monitoring and evaluating the impact and effectiveness of those systems
Resumo:
The UK house building sector is facing dual pressures to expand supply, along with delivering against tougher Building Regulations’ requirements, predominantly in the areas of sustainability. A review of current literature has highlighted that the pressures the UK house building industry is currently under may be having a negative impact on build quality, causing an increase in defects. A review and synthesis of the current defect literature with respect to new-build housing and the wider construction sector has found that the prevailing emphasis is limited to the classification, causes, pathology and statistical analysis of defects. There is thus a need to better understand the overall impact of individual defects on key stakeholders within the new-build housing defect detection and remediation process. As part of ongoing research to develop and verify a defect impact assessment rating system, this paper seeks to contribute to our understanding of the impact of individual defects from a key stakeholder perspective by undertaking the literature review and synthesis phase. The literature review identifies the three distinct, but interrelated, dominant impact factors: cost, disruption, and health and safety. By pulling the strands of defect literature together the theoretical lens and key stakeholder sampling strategy is formed as the basis for the subsequent impact weighting development phase.
Resumo:
Current methods for initialising coupled atmosphere-ocean forecasts often rely on the use of separate atmosphere and ocean analyses, the combination of which can leave the coupled system imbalanced at the beginning of the forecast, potentially accelerating the development of errors. Using a series of experiments with the European Centre for Medium-range Weather Forecasts coupled system, the magnitude and extent of these so-called initialisation shocks is quantified, and their impact on forecast skill measured. It is found that forecasts initialised by separate ocean and atmospheric analyses do exhibit initialisation shocks in lower atmospheric temperature, when compared to forecasts initialised using a coupled data assimilation method. These shocks result in as much as a doubling of root-mean-square error on the first day of the forecast in some regions, and in increases that are sustained for the duration of the 10-day forecasts performed here. However, the impacts of this choice of initialisation on forecast skill, assessed using independent datasets, were found to be negligible, at least over the limited period studied. Larger initialisation shocks are found to follow a change in either the atmospheric or ocean model component between the analysis and forecast phases: changes in the ocean component can lead to sea surface temperature shocks of more than 0.5K in some equatorial regions during the first day of the forecast. Implications for the development of coupled forecast systems, particularly with respect to coupled data assimilation methods, are discussed.
Resumo:
Objectives In this study a prototype of a new health forecasting alert system is developed, which is aligned to the approach used in the Met Office’s (MO) National Severe Weather Warning Service (NSWWS). This is in order to improve information available to responders in the health and social care system by linking temperatures more directly to risks of mortality, and developing a system more coherent with other weather alerts. The prototype is compared to the current system in the Cold Weather and Heatwave plans via a case-study approach to verify its potential advantages and shortcomings. Method The prototype health forecasting alert system introduces an “impact vs likelihood matrix” for the health impacts of hot and cold temperatures which is similar to those used operationally for other weather hazards as part of the NSWWS. The impact axis of this matrix is based on existing epidemiological evidence, which shows an increasing relative risk of death at extremes of outdoor temperature beyond a threshold which can be identified epidemiologically. The likelihood axis is based on a probability measure associated with the temperature forecast. The new method is tested for two case studies (one during summer 2013, one during winter 2013), and compared to the performance of the current alert system. Conclusions The prototype shows some clear improvements over the current alert system. It allows for a much greater degree of flexibility, provides more detailed regional information about the health risks associated with periods of extreme temperatures, and is more coherent with other weather alerts which may make it easier for front line responders to use. It will require validation and engagement with stakeholders before it can be considered for use.
Resumo:
Abstract: A new methodology was created to measure the energy consumption and related green house gas (GHG) emissions of a computer operating system (OS) across different device platforms. The methodology involved the direct power measurement of devices under different activity states. In order to include all aspects of an OS, the methodology included measurements in various OS modes, whilst uniquely, also incorporating measurements when running an array of defined software activities, so as to include OS application management features. The methodology was demonstrated on a laptop and phone that could each run multiple OSs, results confirmed that OS can significantly impact the energy consumption of devices. In particular, the new versions of the Microsoft Windows OS were tested and highlighted significant differences between the OS versions on the same hardware. The developed methodology could enable a greater awareness of energy consumption, during both the software development and software marketing processes.
Resumo:
Replacement and upgrading of assets in the electricity network requires financial investment for the distribution and transmission utilities. The replacement and upgrading of network assets also represents an emissions impact due to the carbon embodied in the materials used to manufacture network assets. This paper uses investment and asset data for the GB system for 2015-2023 to assess the suitability of using a proxy with peak demand data and network investment data to calculate the carbon impacts of network investments. The proxies are calculated on a regional basis and applied to calculate the embodied carbon associated with current network assets by DNO region. The proxies are also applied to peak demand data across the 2015-2023 period to estimate the expected levels of embodied carbon that will be associated with network investment during this period. The suitability of these proxies in different contexts are then discussed, along with initial scenario analysis to calculate the impact of avoiding or deferring network investments through distributed generation projects. The proxies were found to be effective in estimating the total embodied carbon of electricity system investment in order to compare investment strategies in different regions of the GB network.
Resumo:
The predictability of high impact weather events on multiple time scales is a crucial issue both in scientific and socio-economic terms. In this study, a statistical-dynamical downscaling (SDD) approach is applied to an ensemble of decadal hindcasts obtained with the Max-Planck-Institute Earth System Model (MPI-ESM) to estimate the decadal predictability of peak wind speeds (as a proxy for gusts) over Europe. Yearly initialized decadal ensemble simulations with ten members are investigated for the period 1979–2005. The SDD approach is trained with COSMO-CLM regional climate model simulations and ERA-Interim reanalysis data and applied to the MPI-ESM hindcasts. The simulations for the period 1990–1993, which was characterized by several windstorm clusters, are analyzed in detail. The anomalies of the 95 % peak wind quantile of the MPI-ESM hindcasts are in line with the positive anomalies in reanalysis data for this period. To evaluate both the skill of the decadal predictability system and the added value of the downscaling approach, quantile verification skill scores are calculated for both the MPI-ESM large-scale wind speeds and the SDD simulated regional peak winds. Skill scores are predominantly positive for the decadal predictability system, with the highest values for short lead times and for (peak) wind speeds equal or above the 75 % quantile. This provides evidence that the analyzed hindcasts and the downscaling technique are suitable for estimating wind and peak wind speeds over Central Europe on decadal time scales. The skill scores for SDD simulated peak winds are slightly lower than those for large-scale wind speeds. This behavior can be largely attributed to the fact that peak winds are a proxy for gusts, and thus have a higher variability than wind speeds. The introduced cost-efficient downscaling technique has the advantage of estimating not only wind speeds but also estimates peak winds (a proxy for gusts) and can be easily applied to large ensemble datasets like operational decadal prediction systems.