49 resultados para Variables from CGTMSE

em CentAUR: Central Archive University of Reading - UK


Relevância:

80.00% 80.00%

Publicador:

Resumo:

The distribution and variability of water vapor and its links with radiative cooling and latent heating via precipitation are crucial to understanding feedbacks and processes operating within the climate system. Column-integrated water vapor (CWV) and additional variables from the European Centre for Medium-Range Weather Forecasts (ECMWF) 40-year reanalysis (ERA40) are utilized to quantify the spatial and temporal variability in tropical water vapor over the period 1979–2001. The moisture variability is partitioned between dynamical and thermodynamic influences and compared with variations in precipitation provided by the Climate Prediction Center Merged Analysis of Precipitation (CMAP) and the Global Precipitation Climatology Project (GPCP). The spatial distribution of CWV is strongly determined by thermodynamic constraints. Spatial variability in CWV is dominated by changes in the large-scale dynamics, in particular associated with the El Niño–Southern Oscillation (ENSO). Trends in CWV are also dominated by dynamics rather than thermodynamics over the period considered. However, increases in CWV associated with changes in temperature are significant over the equatorial east Pacific when analyzing interannual variability and over the north and northwest Pacific when analyzing trends. Significant positive trends in CWV tend to predominate over the oceans while negative trends in CWV are found over equatorial Africa and Brazil. Links between changes in CWV and vertical motion fields are identified over these regions and also the equatorial Atlantic. However, trends in precipitation are generally incoherent and show little association with the CWV trends. This may in part reflect the inadequacies of the precipitation data sets and reanalysis products when analyzing decadal variability. Though the dynamic component of CWV is a major factor in determining precipitation variability in the tropics, in some regions/seasons the thermodynamic component cancels its effect on precipitation variability.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The rheological properties of dough and gluten are important for end-use quality of flour but there is a lack of knowledge of the relationships between fundamental and empirical tests and how they relate to flour composition and gluten quality. Dough and gluten from six breadmaking wheat qualities were subjected to a range of rheological tests. Fundamental (small-deformation) rheological characterizations (dynamic oscillatory shear and creep recovery) were performed on gluten to avoid the nonlinear influence of the starch component, whereas large deformation tests were conducted on both dough and gluten. A number of variables from the various curves were considered and subjected to a principal component analysis (PCA) to get an overview of relationships between the various variables. The first component represented variability in protein quality, associated with elasticity and tenacity in large deformation (large positive loadings for resistance to extension and initial slope of dough and gluten extension curves recorded by the SMS/Kieffer dough and gluten extensibility rig, and the tenacity and strain hardening index of dough measured by the Dobraszczyk/Roberts dough inflation system), the elastic character of the hydrated gluten proteins (large positive loading for elastic modulus [G'], large negative loadings for tan delta and steady state compliance [J(e)(0)]), the presence of high molecular weight glutenin subunits (HMW-GS) 5+10 vs. 2+12, and a size distribution of glutenin polymers shifted toward the high-end range. The second principal component was associated with flour protein content. Certain rheological data were influenced by protein content in addition to protein quality (area under dough extension curves and dough inflation curves [W]). The approach made it possible to bridge the gap between fundamental rheological properties, empirical measurements of physical properties, protein composition, and size distribution. The interpretation of this study gave indications of the molecular basis for differences in breadmaking performance.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Aim  We provide a new quantitative analysis of lizard reproductive ecology. Comparative studies of lizard reproduction to date have usually considered life-history components separately. Instead, we examine the rate of production (productivity hereafter) calculated as the total mass of offspring produced in a year. We test whether productivity is influenced by proxies of adult mortality rates such as insularity and fossorial habits, by measures of temperature such as environmental and body temperatures, mode of reproduction and activity times, and by environmental productivity and diet. We further examine whether low productivity is linked to high extinction risk. Location  World-wide. Methods  We assembled a database containing 551 lizard species, their phylogenetic relationships and multiple life history and ecological variables from the literature. We use phylogenetically informed statistical models to estimate the factors related to lizard productivity. Results  Some, but not all, predictions of metabolic and life-history theories are supported. When analysed separately, clutch size, relative clutch mass and brood frequency are poorly correlated with body mass, but their product – productivity – is well correlated with mass. The allometry of productivity scales similarly to metabolic rate, suggesting that a constant fraction of assimilated energy is allocated to production irrespective of body size. Island species were less productive than continental species. Mass-specific productivity was positively correlated with environmental temperature, but not with body temperature. Viviparous lizards were less productive than egg-laying species. Diet and primary productivity were not associated with productivity in any model. Other effects, including lower productivity of fossorial, nocturnal and active foraging species were confounded with phylogeny. Productivity was not lower in species at risk of extinction. Main conclusions  Our analyses show the value of focusing on the rate of annual biomass production (productivity), and generally supported associations between productivity and environmental temperature, factors that affect mortality and the number of broods a lizard can produce in a year, but not with measures of body temperature, environmental productivity or diet.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Excessive salt intake is linked to cardiovascular disease and several other health problems around the world. The UK Food Standards Agency initiated a campaign at the end of 2004 to reduce salt intake in the population. There is disagreement over whether the campaign was effective in curbing salt intake or not. We provide fresh evidence on the impact of the campaign, by using data on spot urinary sodium readings and socio-demographic variables from the Health Survey for England over 2003–2007 and combining it with food price information from the Expenditure and Food Survey. Aggregating the data into a pseudo-panel, we estimate fixed effects models to examine the trend in salt intake over the period and to deduce the heterogeneous effects of the policy on the intake of socio-demographic groups. Our results are consistent with a previous hypothesis that the campaign reduced salt intakes by approximately 10%. The impact is shown to be stronger among women than among men. Older cohorts of men show a larger response to the salt campaign compared to younger cohorts, while among women, younger cohorts respond more strongly than older cohorts.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The evidence provided by modelled assessments of future climate impact on flooding is fundamental to water resources and flood risk decision making. Impact models usually rely on climate projections from global and regional climate models (GCM/RCMs). However, challenges in representing precipitation events at catchment-scale resolution mean that decisions must be made on how to appropriately pre-process the meteorological variables from GCM/RCMs. Here the impacts on projected high flows of differing ensemble approaches and application of Model Output Statistics to RCM precipitation are evaluated while assessing climate change impact on flood hazard in the Upper Severn catchment in the UK. Various ensemble projections are used together with the HBV hydrological model with direct forcing and also compared to a response surface technique. We consider an ensemble of single-model RCM projections from the current UK Climate Projections (UKCP09); multi-model ensemble RCM projections from the European Union's FP6 ‘ENSEMBLES’ project; and a joint probability distribution of precipitation and temperature from a GCM-based perturbed physics ensemble. The ensemble distribution of results show that flood hazard in the Upper Severn is likely to increase compared to present conditions, but the study highlights the differences between the results from different ensemble methods and the strong assumptions made in using Model Output Statistics to produce the estimates of future river discharge. The results underline the challenges in using the current generation of RCMs for local climate impact studies on flooding. Copyright © 2012 Royal Meteorological Society

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The variogram is essential for local estimation and mapping of any variable by kriging. The variogram itself must usually be estimated from sample data. The sampling density is a compromise between precision and cost, but it must be sufficiently dense to encompass the principal spatial sources of variance. A nested, multi-stage, sampling with separating distances increasing in geometric progression from stage to stage will do that. The data may then be analyzed by a hierarchical analysis of variance to estimate the components of variance for every stage, and hence lag. By accumulating the components starting from the shortest lag one obtains a rough variogram for modest effort. For balanced designs the analysis of variance is optimal; for unbalanced ones, however, these estimators are not necessarily the best, and the analysis by residual maximum likelihood (REML) will usually be preferable. The paper summarizes the underlying theory and illustrates its application with data from three surveys, one in which the design had four stages and was balanced and two implemented with unbalanced designs to economize when there were more stages. A Fortran program is available for the analysis of variance, and code for the REML analysis is listed in the paper. (c) 2005 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The study reported presents the findings relating to commercial growing of genetically-modified Bt cotton in South Africa by a large sample of smallholder farmers over three seasons (1998/99, 1999/2000, 2000/01) following adoption. The analysis presents constructs and compares groupwise differences for key variables in Bt v. non-Bt technology and uses regressions to further analyse the production and profit impacts of Bt adoption. Analysis of the distribution of benefits between farmers due to the technology is also presented. In parallel with these socio-economic measures, the toxic loads being presented to the environment following the introduction of Bt cotton are monitored in terms of insecticide active ingredient (ai) and the Biocide Index. The latter adjusts ai to allow for differing persistence and toxicity of insecticides. Results show substantial and significant financial benefits to smallholder cotton growers of adopting Bt cotton over three seasons in terms of increased yields, lower insecticide spray costs and higher gross margins. This includes one particularly wet, poor growing season. In addition, those with the smaller holdings appeared to benefit proportionately more from the technology (in terms of higher gross margins) than those with larger holdings. Analysis using the Gini-coefficient suggests that the Bt technology has helped to reduce inequality amongst smallholder cotton growers in Makhathini compared to what may have been the position if they had grown conventional cotton. However, while Bt growers applied lower amounts of insecticide and had lower Biocide Indices (per ha) than growers of non-Bt cotton, some of this advantage was due to a reduction in non-bollworm insecticide. Indeed, the Biocide Index for all farmers in the population actually increased with the introduction of Bt cotton. The results indicate the complexity of such studies on the socio-economic and environmental impacts of GM varieties in the developing world.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The variogram is essential for local estimation and mapping of any variable by kriging. The variogram itself must usually be estimated from sample data. The sampling density is a compromise between precision and cost, but it must be sufficiently dense to encompass the principal spatial sources of variance. A nested, multi-stage, sampling with separating distances increasing in geometric progression from stage to stage will do that. The data may then be analyzed by a hierarchical analysis of variance to estimate the components of variance for every stage, and hence lag. By accumulating the components starting from the shortest lag one obtains a rough variogram for modest effort. For balanced designs the analysis of variance is optimal; for unbalanced ones, however, these estimators are not necessarily the best, and the analysis by residual maximum likelihood (REML) will usually be preferable. The paper summarizes the underlying theory and illustrates its application with data from three surveys, one in which the design had four stages and was balanced and two implemented with unbalanced designs to economize when there were more stages. A Fortran program is available for the analysis of variance, and code for the REML analysis is listed in the paper. (c) 2005 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A semi-distributed model, INCA, has been developed to determine the fate and distribution of nutrients in terrestrial and aquatic systems. The model simulates nitrogen and phosphorus processes in soils, groundwaters and river systems and can be applied in a semi-distributed manner at a range of scales. In this study, the model has been applied at field to sub-catchment to whole catchment scale to evaluate the behaviour of biosolid-derived losses of P in agricultural systems. It is shown that process-based models such as INCA, applied at a wide range of scales, reproduce field and catchment behaviour satisfactorily. The INCA model can also be used to generate generic information for risk assessment. By adjusting three key variables: biosolid application rates, the hydrological connectivity of the catchment and the initial P-status of the soils within the model, a matrix of P loss rates can be generated to evaluate the behaviour of the model and, hence, of the catchment system. The results, which indicate the sensitivity of the catchment to flow paths, to application rates and to initial soil conditions, have been incorporated into a Nutrient Export Risk Matrix (NERM).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background Pharmacy aseptic units prepare and supply injectables to minimise risks. The UK National Aseptic Error Reporting Scheme has been collecting data on pharmacy compounding errors, including near-misses, since 2003. Objectives The cumulative reports from January 2004 to December 2007, inclusive, were analysed. Methods The different variables of product types, error types, staff making and detecting errors, stage errors detected, perceived contributory factors, and potential or actual outcomes were presented by cross-tabulation of data. Results A total of 4691 reports were submitted against an estimated 958 532 items made, returning 0.49% as the overall error rate. Most of the errors were detected before reaching patients, with only 24 detected during or after administration. The highest number of reports related to adult cytotoxic preparations (40%) and the most frequently recorded error was a labelling error (34.2%). Errors were mostly detected at first check in assembly area (46.6%). Individual staff error contributed most (78.1%) to overall errors, while errors with paediatric parenteral nutrition appeared to be blamed on low staff levels more than other products were. The majority of errors (68.6%) had no potential patient outcomes attached, while it appeared that paediatric cytotoxic products and paediatric parenteral nutrition were associated with greater levels of perceived patient harm. Conclusions The majority of reports were related to near-misses, and this study highlights scope for examining current arrangements for checking and releasing products, certainly for paediatric cytotoxic and paediatric parenteral nutrition preparations within aseptic units, but in the context of resource and capacity constraints.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ozone and temperature profiles from the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS) have been assimilated, using three-dimensional variational assimilation, into a stratosphere troposphere version of the Met Office numerical weather-prediction system. Analyses are made for the month of September 2002, when there was an unprecedented split in the southern hemisphere polar vortex. The analyses are validated against independent ozone observations from sondes, limb-occultation and total column ozone satellite instruments. Through most of the stratosphere, precision varies from 5 to 15%, and biases are 15% or less of the analysed field. Problems remain in the vortex and below the 60 hPa. level, especially at the tropopause where the analyses have too much ozone and poor agreement with independent data. Analysis problems are largely a result of the model rather than the data, giving confidence in the MIPAS ozone retrievals, though there may be a small high bias in MIPAS ozone in the lower stratosphere. Model issues include an excessive Brewer-Dobson circulation, which results both from known problems with the tracer transport scheme and from the data assimilation of dynamical variables. The extreme conditions of the vortex split reveal large differences between existing linear ozone photochemistry schemes. Despite these issues, the ozone analyses are able to successfully describe the ozone hole split and compare well to other studies of this event. Recommendations are made for the further development of the ozone assimilation system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

During a 4-week run in October–November 2006, a pilot experiment was performed at the CERN Proton Synchrotron in preparation for the Cosmics Leaving OUtdoor Droplets (CLOUD) experiment, whose aim is to study the possible influence of cosmic rays on clouds. The purpose of the pilot experiment was firstly to carry out exploratory measurements of the effect of ionising particle radiation on aerosol formation from trace H2SO4 vapour and secondly to provide technical input for the CLOUD design. A total of 44 nucleation bursts were produced and recorded, with formation rates of particles above the 3 nm detection threshold of between 0.1 and 100 cm−3 s−1, and growth rates between 2 and 37 nm h−1. The corresponding H2SO4 concentrations were typically around 106 cm−3 or less. The experimentally-measured formation rates and H2SO4 concentrations are comparable to those found in the atmosphere, supporting the idea that sulphuric acid is involved in the nucleation of atmospheric aerosols. However, sulphuric acid alone is not able to explain the observed rapid growth rates, which suggests the presence of additional trace vapours in the aerosol chamber, whose identity is unknown. By analysing the charged fraction, a few of the aerosol bursts appear to have a contribution from ion-induced nucleation and ion-ion recombination to form neutral clusters. Some indications were also found for the accelerator beam timing and intensity to influence the aerosol particle formation rate at the highest experimental SO2 concentrations of 6 ppb, although none was found at lower concentrations. Overall, the exploratory measurements provide suggestive evidence for ion-induced nucleation or ion-ion recombination as sources of aerosol particles. However in order to quantify the conditions under which ion processes become significant, improvements are needed in controlling the experimental variables and in the reproducibility of the experiments. Finally, concerning technical aspects, the most important lessons for the CLOUD design include the stringent requirement of internal cleanliness of the aerosol chamber, as well as maintenance of extremely stable temperatures (variations below 0.1 _C).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The study reported presents the findings relating to commercial growing of genetically-modified Bt cotton in South Africa by a large sample of smallholder farmers over three seasons (1998/99, 1999/2000, 2000/01) following adoption. The analysis presents constructs and compares groupwise differences for key variables in Bt v. non-Bt technology and uses regressions to further analyse the production and profit impacts of Bt adoption. Analysis of the distribution of benefits between farmers due to the technology is also presented. In parallel with these socio-economic measures, the toxic loads being presented to the environment following the introduction of Bt cotton are monitored in terms of insecticide active ingredient (ai) and the Biocide Index. The latter adjusts ai to allow for differing persistence and toxicity of insecticides. Results show substantial and significant financial benefits to smallholder cotton growers of adopting Bt cotton over three seasons in terms of increased yields, lower insecticide spray costs and higher gross margins. This includes one particularly wet, poor growing season. In addition, those with the smaller holdings appeared to benefit proportionately more from the technology (in terms of higher gross margins) than those with larger holdings. Analysis using the Gini-coefficient suggests that the Bt technology has helped to reduce inequality amongst smallholder cotton growers in Makhathini compared to what may have been the position if they had grown conventional cotton. However, while Bt growers applied lower amounts of insecticide and had lower Biocide Indices (per ha) than growers of non-Bt cotton, some of this advantage was due to a reduction in non-bollworm insecticide. Indeed, the Biocide Index for all farmers in the population actually increased with the introduction of Bt cotton. The results indicate the complexity of such studies on the socio-economic and environmental impacts of GM varieties in the developing world.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the Biodiversity World (BDW) project we have created a flexible and extensible Web Services-based Grid environment for biodiversity researchers to solve problems in biodiversity and analyse biodiversity patterns. In this environment, heterogeneous and globally distributed biodiversity-related resources such as data sets and analytical tools are made available to be accessed and assembled by users into workflows to perform complex scientific experiments. One such experiment is bioclimatic modelling of the geographical distribution of individual species using climate variables in order to predict past and future climate-related changes in species distribution. Data sources and analytical tools required for such analysis of species distribution are widely dispersed, available on heterogeneous platforms, present data in different formats and lack interoperability. The BDW system brings all these disparate units together so that the user can combine tools with little thought as to their availability, data formats and interoperability. The current Web Servicesbased Grid environment enables execution of the BDW workflow tasks in remote nodes but with a limited scope. The next step in the evolution of the BDW architecture is to enable workflow tasks to utilise computational resources available within and outside the BDW domain. We describe the present BDW architecture and its transition to a new framework which provides a distributed computational environment for mapping and executing workflows in addition to bringing together heterogeneous resources and analytical tools.