219 resultados para load planning

em CentAUR: Central Archive University of Reading - UK


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The development of a combined engineering and statistical Artificial Neural Network model of UK domestic appliance load profiles is presented. The model uses diary-style appliance use data and a survey questionnaire collected from 51 suburban households and 46 rural households during the summer of 2010 and2011 respectively. It also incorporates measured energy data and is sensitive to socioeconomic, physical dwelling and temperature variables. A prototype model is constructed in MATLAB using a two layer feed forward network with back propagation training which has a 12:10:24 architecture. Model outputs include appliance load profiles which can be applied to the fields of energy planning (microrenewables and smart grids), building simulation tools and energy policy.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Planning a project with proper considerations of all necessary factors and managing a project to ensure its successful implementation will face a lot of challenges. Initial stage in planning a project for bidding a project is costly, time consuming and usually with poor accuracy on cost and effort predictions. On the other hand, detailed information for previous projects may be buried in piles of archived documents which can be increasingly difficult to learn from the previous experiences. Project portfolio has been brought into this field aiming to improve the information sharing and management among different projects. However, the amount of information that could be shared is still limited to generic information. This paper, we report a recently developed software system COBRA to automatically generate a project plan with effort estimation of time and cost based on data collected from previous completed projects. To maximise the data sharing and management among different projects, we proposed a method of using product based planning from PRINCE2 methodology. (Automated Project Information Sharing and Management System -�COBRA) Keywords: project management, product based planning, best practice, PRINCE2

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The study reported presents the findings relating to commercial growing of genetically-modified Bt cotton in South Africa by a large sample of smallholder farmers over three seasons (1998/99, 1999/2000, 2000/01) following adoption. The analysis presents constructs and compares groupwise differences for key variables in Bt v. non-Bt technology and uses regressions to further analyse the production and profit impacts of Bt adoption. Analysis of the distribution of benefits between farmers due to the technology is also presented. In parallel with these socio-economic measures, the toxic loads being presented to the environment following the introduction of Bt cotton are monitored in terms of insecticide active ingredient (ai) and the Biocide Index. The latter adjusts ai to allow for differing persistence and toxicity of insecticides. Results show substantial and significant financial benefits to smallholder cotton growers of adopting Bt cotton over three seasons in terms of increased yields, lower insecticide spray costs and higher gross margins. This includes one particularly wet, poor growing season. In addition, those with the smaller holdings appeared to benefit proportionately more from the technology (in terms of higher gross margins) than those with larger holdings. Analysis using the Gini-coefficient suggests that the Bt technology has helped to reduce inequality amongst smallholder cotton growers in Makhathini compared to what may have been the position if they had grown conventional cotton. However, while Bt growers applied lower amounts of insecticide and had lower Biocide Indices (per ha) than growers of non-Bt cotton, some of this advantage was due to a reduction in non-bollworm insecticide. Indeed, the Biocide Index for all farmers in the population actually increased with the introduction of Bt cotton. The results indicate the complexity of such studies on the socio-economic and environmental impacts of GM varieties in the developing world.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this work is to study the hydrochemical variations during flood events in the Rio Tinto, SW Spain. Three separate rainfall/flood events were monitored in October 2004 following the dry season. In general, concentrations markedly increased following the first event (Fe from 99 to 1130 mg/L; Q(max) = 0.78 m(3)/s) while dissolved loads peaked in the second event (Fe = 7.5 kg/s, Cu = 0.83 kg/s, Zn = 0.82 kg/s; Q(max) = 77 m(3)/s) and discharge in the third event (Q(max) = 127 m(3)/s). This pattern reflects a progressive depletion of metals and sulphate stored in the dry summer as soluble evaporitic salt minerals and concentrated pore fluids, with dilution by freshwater becoming increasingly dominant as the month progressed. Variations in relative concentrations were attributed to oxyhydroxysulphate Fe precipitation, to relative changes in the sources of acid mine drainage (e.g. salt minerals, mine tunnels, spoil heaps etc.) and to differences in the rainfall distributions along the catchment. The contaminant load carried by the river during October 2004 was enormous, totalling some 770 t of Fe, 420 t of Al, 100 t of Cu, 100 t of Zn and 71 t of Mn. This represents the largest recorded example of this flush-out process in an acid mine drainage setting. Approximately 1000 times more water and 1408 200 times more dissolved elements were carried by the river during October 2004 than during the dry, low-flow conditions of September 2004, highlighting the key role of flood Events in the annual pollutant transport budget of semi-arid and and systems and the need to monitor these events in detail in order to accurately quantify pollutant transport. (c) 2007 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

1. The UK Biodiversity Action Plan (UKBAP) identifies invertebrate species in danger of national extinction. For many of these species, targets for recovery specify the number of populations that should exist by a specific future date but offer no procedure to plan strategically to achieve the target for any species. 2. Here we describe techniques based upon geographic information systems (GIS) that produce conservation strategy maps (CSM) to assist with achieving recovery targets based on all available and relevant information. 3. The heath fritillary Mellicta athalia is a UKBAP species used here to illustrate the use of CSM. A phase 1 habitat survey was used to identify habitat polygons across the county of Kent, UK. These were systematically filtered using relevant habitat, botanical and autecological data to identify seven types of polygon, including those with extant colonies or in the vicinity of extant colonies, areas managed for conservation but without colonies, and polygons that had the appropriate habitat structure and may therefore be suitable for reintroduction. 4. Five clusters of polygons of interest were found across the study area. The CSM of two of them are illustrated here: the Blean Wood complex, which contains the existing colonies of heath fritillary in Kent, and the Orlestone Forest complex, which offers opportunities for reintroduction. 5. Synthesis and applications. Although the CSM concept is illustrated here for the UK, we suggest that CSM could be part of species conservation programmes throughout the world. CSM are dynamic and should be stored in electronic format, preferably on the world-wide web, so that they can be easily viewed and updated. CSM can be used to illustrate opportunities and to develop strategies with scientists and non-scientists, enabling the engagement of all communities in a conservation programme. CSM for different years can be presented to illustrate the progress of a plan or to provide continuous feedback on how a field scenario develops.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Models developed to identify the rates and origins of nutrient export from land to stream require an accurate assessment of the nutrient load present in the water body in order to calibrate model parameters and structure. These data are rarely available at a representative scale and in an appropriate chemical form except in research catchments. Observational errors associated with nutrient load estimates based on these data lead to a high degree of uncertainty in modelling and nutrient budgeting studies. Here, daily paired instantaneous P and flow data for 17 UK research catchments covering a total of 39 water years (WY) have been used to explore the nature and extent of the observational error associated with nutrient flux estimates based on partial fractions and infrequent sampling. The daily records were artificially decimated to create 7 stratified sampling records, 7 weekly records, and 30 monthly records from each WY and catchment. These were used to evaluate the impact of sampling frequency on load estimate uncertainty. The analysis underlines the high uncertainty of load estimates based on monthly data and individual P fractions rather than total P. Catchments with a high baseflow index and/or low population density were found to return a lower RMSE on load estimates when sampled infrequently than those with a tow baseflow index and high population density. Catchment size was not shown to be important, though a limitation of this study is that daily records may fail to capture the full range of P export behaviour in smaller catchments with flashy hydrographs, leading to an underestimate of uncertainty in Load estimates for such catchments. Further analysis of sub-daily records is needed to investigate this fully. Here, recommendations are given on load estimation methodologies for different catchment types sampled at different frequencies, and the ways in which this analysis can be used to identify observational error and uncertainty for model calibration and nutrient budgeting studies. (c) 2006 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Amman the primate capital city of the Hashemite Kingdom of Jordan currently has a population in excess of 2 million, but in 1924 it consisted of little more than a collection of dwellings and some 2000-3000 inhabitants. The present paper sets out to document and explain the phenomenal expansion of "ever-growing Amman". The physical geography of the urban region and the early growth of the city are considered at the outset and this leads directly to consideration of the highly polarised social structuring that characterises contemporary Amman. In doing this, original data derived from the recent Greater Amman Municipality's Geographical Information System are presented. In this respect, the essential modernity of the city is exemplified. The employment and industrial bases of the city and a range of pressing contemporary issues are then considered, including transport and congestion, the provision of urban water under conditions of water stress and privatisation, and urban and regional development planning for the city. The paper concludes by emphasizing the growing regional and international geopolitical salience of the city of Amman at the start of the 21st century. (C) 2008 Elsevier Ltd. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In molecular biology, it is often desirable to find common properties in large numbers of drug candidates. One family of methods stems from the data mining community, where algorithms to find frequent graphs have received increasing attention over the past years. However, the computational complexity of the underlying problem and the large amount of data to be explored essentially render sequential algorithms useless. In this paper, we present a distributed approach to the frequent subgraph mining problem to discover interesting patterns in molecular compounds. This problem is characterized by a highly irregular search tree, whereby no reliable workload prediction is available. We describe the three main aspects of the proposed distributed algorithm, namely, a dynamic partitioning of the search space, a distribution process based on a peer-to-peer communication framework, and a novel receiverinitiated load balancing algorithm. The effectiveness of the distributed method has been evaluated on the well-known National Cancer Institute’s HIV-screening data set, where we were able to show close-to linear speedup in a network of workstations. The proposed approach also allows for dynamic resource aggregation in a non dedicated computational environment. These features make it suitable for large-scale, multi-domain, heterogeneous environments, such as computational grids.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we present a distributed computing framework for problems characterized by a highly irregular search tree, whereby no reliable workload prediction is available. The framework is based on a peer-to-peer computing environment and dynamic load balancing. The system allows for dynamic resource aggregation, does not depend on any specific meta-computing middleware and is suitable for large-scale, multi-domain, heterogeneous environments, such as computational Grids. Dynamic load balancing policies based on global statistics are known to provide optimal load balancing performance, while randomized techniques provide high scalability. The proposed method combines both advantages and adopts distributed job-pools and a randomized polling technique. The framework has been successfully adopted in a parallel search algorithm for subgraph mining and evaluated on a molecular compounds dataset. The parallel application has shown good calability and close-to linear speedup in a distributed network of workstations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Explanations are an important by-product of medical decisionsupport activities, as they have proved to favour compliance and correct treatment performance. To achieve this purpose, these texts should have a strong argumentation content and should adapt to emotional, as well as to rational attitudes of the Addressee. This paper describes how Rhetorical Sentence Planning can contribute to this aim: the rulebased plan discourse revision is introduced between Text Planning and Linguistic Realization, and exploits knowledge about the user personality and emotions and about the potential impact of domain items on user compliance and memory recall. The proposed approach originates from analytical and empirical evaluation studies of computer generated explanation texts in the domain of drug prescription. This work was partially supported by a British-Italian Collaboration in Research and Higher Education Project, which involved the Universities of Reading and of Bari, in 1996.