199 resultados para Simulation package
Resumo:
Common approaches to the simulation of borehole heat exchangers (BHEs) assume heat transfer in circulating fluid and grout to be in a quasi-steady state and ignore fluctuations in fluid temperature due to transport of the fluid around the loop. However, in domestic ground source heat pump (GSHP) systems, the heat pump and circulating pumps switch on and off during a given hour; therefore, the effect of the thermal mass of the circulating fluid and the dynamics of fluid transport through the loop has important implications for system design. This may also be important in commercial systems that are used intermittently. This article presents transient simulation of a domestic GSHP system with a single BHE using a dynamic three-dimensional (3D) numerical BHE model. The results show that delayed response associated with the transit of fluid along the pipe loop is of some significance in moderating swings in temperature during heat pump operation. In addition, when 3D effects are considered, a lower heat transfer rate is predicted during steady operations. These effects could be important when considering heat exchanger design and system control. The results will be used to develop refined two-dimensional models.
Resumo:
Daylighting systems can offer energy savings primarily by reducing electric lighting usage. Accurate predictive models of daylighting system performances are crucial for effective design and implementation of this renewable energy technology. A comparative study of predictive methods was performed and the use of a commercial raytracing software program was validated as a method of predicting light pipe performance. Raytracing simulation was shown to more accurately predict transmission effi ciency than existing analytical methods.
Resumo:
Recent activity in the development of future weather data for building performance simulation follows recognition of the limitations of traditional methods, which have been based on a stationary (observed) climate. In the UK, such developments have followed on from the availability of regional climate models as delivered in UKCIP02 and recently the probabilistic projections released under UKCP09. One major area of concern is the future performance and adaptability of buildings which employ exclusively passive or low-energy cooling systems. One such method which can be employed in an integral or retrofit situation is direct or indirect evaporative cooling. The effectiveness of evaporative cooling is most strongly influenced by the wet-bulb depression of the ambient air, hence is generally regarded as most suited to hot, dry climates. However, this technology has been shown to be effective in the UK, primarily in mixed-mode buildings or as a retrofit to industrial/commercial applications. Climate projections for the UK generally indicate an increase in the summer wet-bulb depression, suggesting an enhanced potential for the application of evaporative cooling. The paper illustrates this potential by an analysis of the probabilistic scenarios released under UKCP09, together with a detailed building/plant simulation of case study building located in the South-East of England. The results indicate a high probability that evaporative cooling will still be a viable low-energy technique in the 2050s.
Resumo:
Design summer years representing near-extreme hot summers have been used in the United Kingdom for the evaluation of thermal comfort and overheating risk. The years have been selected from measured weather data basically representative of an assumed stationary climate. Recent developments have made available ‘morphed’ equivalents of these years by shifting and stretching the measured variables using change factors produced by the UKCIP02 climate projections. The release of the latest, probabilistic, climate projections of UKCP09 together with the availability of a weather generator that can produce plausible daily or hourly sequences of weather variables has opened up the opportunity for generating new design summer years which can be used in risk-based decision-making. There are many possible methods for the production of design summer years from UKCP09 output: in this article, the original concept of the design summer year is largely retained, but a number of alternative methodologies for generating the years are explored. An alternative, more robust measure of warmth (weighted cooling degree hours) is also employed. It is demonstrated that the UKCP09 weather generator is capable of producing years for the baseline period, which are comparable with those in current use. Four methodologies for the generation of future years are described, and their output related to the future (deterministic) years that are currently available. It is concluded that, in general, years produced from the UKCP09 projections are warmer than those generated previously. Practical applications: The methodologies described in this article will facilitate designers who have access to the output of the UKCP09 weather generator (WG) to generate Design Summer Year hourly files tailored to their needs. The files produced will differ according to the methodology selected, in addition to location, emissions scenario and timeslice.
Resumo:
The type and thickness of insulation on the topside horizontal of cold pitched roofs has a significant role in controlling air movement, energy conservation and moisture transfer reduction through the ceiling to the loft (roof void) space. To investigate its importance, a numerical model using a HAM software package on a Matlab platform with a Simulink simulation tool has been developed using insitu measurements of airflows from the dwelling space through the ceiling to the loft of three houses of different configurations and loft space. Considering typical UK roof underlay (i.e. bituminous felt and a vapour permeable underlay), insitu measurements of the 3 houses were tested using a calibrated passive sampling technique. Using the measured airflows, the effect of air movement on three types of roof insulation (i.e. fibreglass, cellulose and foam) was modelled to investigate associated energy losses and moisture transport. The thickness of the insulation materials were varied but the ceiling airtightness and eaves gap size were kept constant. These instances were considered in order to visualize the effects of the changing parameters. In addition, two different roof underlays of varying resistances were considered and compared to access the influence of the underlay, if any, on energy conservation. The comparison of these insulation materials in relation to the other parameters showed that the type of insulation material and thickness, contributes significantly to energy conservation and moisture transfer reduction through the roof and hence of the building as a whole.
Resumo:
The Chartered Institute of Building Service Engineers (CIBSE) produced a technical memorandum (TM36) presenting research on future climate impacting building energy use and thermal comfort. One climate projection for each of four CO2 emissions scenario were used in TM36, so providing a deterministic outlook. As part of the UK Climate Impacts Programme (UKCIP) probabilistic climate projections are being studied in relation to building energy simulation techniques. Including uncertainty in climate projections is considered an important advance to climate impacts modelling and is included in the latest UKCIP data (UKCP09). Incorporating the stochastic nature of these new climate projections in building energy modelling requires a significant increase in data handling and careful statistical interpretation of the results to provide meaningful conclusions. This paper compares the results from building energy simulations when applying deterministic and probabilistic climate data. This is based on two case study buildings: (i) a mixed-mode office building with exposed thermal mass and (ii) a mechanically ventilated, light-weight office building. Building (i) represents an energy efficient building design that provides passive and active measures to maintain thermal comfort. Building (ii) relies entirely on mechanical means for heating and cooling, with its light-weight construction raising concern over increased cooling loads in a warmer climate. Devising an effective probabilistic approach highlighted greater uncertainty in predicting building performance, depending on the type of building modelled and the performance factors under consideration. Results indicate that the range of calculated quantities depends not only on the building type but is strongly dependent on the performance parameters that are of interest. Uncertainty is likely to be particularly marked with regard to thermal comfort in naturally ventilated buildings.
Resumo:
OBJECTIVES: The prediction of protein structure and the precise understanding of protein folding and unfolding processes remains one of the greatest challenges in structural biology and bioinformatics. Computer simulations based on molecular dynamics (MD) are at the forefront of the effort to gain a deeper understanding of these complex processes. Currently, these MD simulations are usually on the order of tens of nanoseconds, generate a large amount of conformational data and are computationally expensive. More and more groups run such simulations and generate a myriad of data, which raises new challenges in managing and analyzing these data. Because the vast range of proteins researchers want to study and simulate, the computational effort needed to generate data, the large data volumes involved, and the different types of analyses scientists need to perform, it is desirable to provide a public repository allowing researchers to pool and share protein unfolding data. METHODS: To adequately organize, manage, and analyze the data generated by unfolding simulation studies, we designed a data warehouse system that is embedded in a grid environment to facilitate the seamless sharing of available computer resources and thus enable many groups to share complex molecular dynamics simulations on a more regular basis. RESULTS: To gain insight into the conformational fluctuations and stability of the monomeric forms of the amyloidogenic protein transthyretin (TTR), molecular dynamics unfolding simulations of the monomer of human TTR have been conducted. Trajectory data and meta-data of the wild-type (WT) protein and the highly amyloidogenic variant L55P-TTR represent the test case for the data warehouse. CONCLUSIONS: Web and grid services, especially pre-defined data mining services that can run on or 'near' the data repository of the data warehouse, are likely to play a pivotal role in the analysis of molecular dynamics unfolding data.
Resumo:
The plume of Ice Shelf Water (ISW) flowing into the Weddell Sea over the Filchner sill contributes to the formation of Antarctic Bottom Water. The Filchner overflow is simulated using a hydrostatic, primitive equation three-dimensional ocean model with a 0.5–2 Sv ISW influx above the Filchner sill. The best fit to mooring temperature observations is found with influxes of 0.5 and 1 Sv, below a previous estimate of 1.6 ± 0.5 Sv based on sparse mooring velocities. The plume first moves north over the continental shelf, and then turns west, along slope of the continental shelf break where it breaks up into subplumes and domes, some of which then move downslope. Other subplumes run into the eastern submarine ridge and propagate along the ridge downslope in a chaotic manner. The next, western ridge is crossed by the plume through several paths. Despite a number of discrepancies with observational data, the model reproduces many attributes of the flow. In particular, we argue that the temporal variability shown by the observations can largely be attributed to the unstable structure of the flow, where the temperature fluctuations are determined by the motion of the domes past the moorings. Our sensitivity studies show that while thermobaricity plays a role, its effect is small for the flows considered. Smoothing the ridges out demonstrate that their presence strongly affects the plume shape around the ridges. An increase in the bottom drag or viscosity leads to slowing down, and hence thickening and widening of the plume
Resumo:
This study investigates the possibilities and limitations of using Regional Climate Model (RCM) output for the simulation of alpine permafrost scenarios. It focuses on the general problem of scale mismatch between RCMs and impact models and, in particular, the special challenges that arise when driving an impact model in topographically complex high-mountain environments with the output of an RCM. Two approaches are introduced that take into account the special difficulties in such areas, and thus enable the use of RCM for alpine permafrost scenario modelling. Intended as an initial example, they are applied at the area of Corvatsch (Upper Engadine, Switzerland) in order to demonstrate and discuss the application of the two approaches, rather than to provide an assessment of future changes in permafrost occurrence. There are still many uncertainties and inaccuracies inherent in climate and impact models, which increase when driving one model with the output of the other. Nevertheless, our study shows that the use of RCMs offers new and promising perspectives for the simulation of high-mountain permafrost scenarios
Resumo:
This paper presents results obtained from a numerical simulation for the horizontal slinky-loop heat exchanger of a ground-source heat pump system. A three-dimensional numerical model was developed and the results of the thermal performance of various heat exchanger configurations are presented. The investigation was carried out on five types of loop pitch (loop spacing), three types of loop diameter, three values of soil thermal properties, and allowing continuous and intermittent operation. Comparison was made for the heat transfer rate, the amount of pipe material needed, as well as excavation work required for the horizontal slinky-loop heat exchanger. The results indicate that system parameters have a significant effect on the thermal performance of the system
Resumo:
This Atlas presents statistical analyses of the simulations submitted to the Aqua-Planet Experiment (APE) data archive. The simulations are from global Atmospheric General Circulation Models (AGCM) applied to a water-covered earth. The AGCMs include ones actively used or being developed for numerical weather prediction or climate research. Some are mature, application models and others are more novel and thus less well tested in Earth-like applications. The experiment applies AGCMs with their complete parameterization package to an idealization of the planet Earth which has a greatly simplified lower boundary that consists of an ocean only. It has no land and its associated orography, and no sea ice. The ocean is represented by Sea Surface Temperatures (SST) which are specified everywhere with simple, idealized distributions. Thus in the hierarchy of tests available for AGCMs, APE falls between tests with simplified forcings such as those proposed by Held and Suarez (1994) and Boer and Denis (1997) and Earth-like simulations of the Atmospheric Modeling Intercomparison Project (AMIP, Gates et al., 1999). Blackburn and Hoskins (2013) summarize the APE and its aims. They discuss where the APE fits within a modeling hierarchy which has evolved to evaluate complete models and which provides a link between realistic simulation and conceptual models of atmospheric phenomena. The APE bridges a gap in the existing hierarchy. The goals of APE are to provide a benchmark of current model behaviors and to stimulate research to understand the cause of inter-model differences., APE is sponsored by the World Meteorological Organization (WMO) joint Commission on Atmospheric Science (CAS), World Climate Research Program (WCRP) Working Group on Numerical Experimentation (WGNE). Chapter 2 of this Atlas provides an overview of the specification of the eight APE experiments and of the data collected. Chapter 3 lists the participating models and includes brief descriptions of each. Chapters 4 through 7 present a wide variety of statistics from the 14 participating models for the eight different experiments. Additional intercomparison figures created by Dr. Yukiko Yamada in AGU group are available at http://www.gfd-dennou.org/library/ape/comparison/. This Atlas is intended to present and compare the statistics of the APE simulations but does not contain a discussion of interpretive analyses. Such analyses are left for journal papers such as those included in the Special Issue of the Journal of the Meteorological Society of Japan (2013, Vol. 91A) devoted to the APE. Two papers in that collection provide an overview of the simulations. One (Blackburn et al., 2013) concentrates on the CONTROL simulation and the other (Williamson et al., 2013) on the response to changes in the meridional SST profile. Additional papers provide more detailed analysis of the basic simulations, while others describe various sensitivities and applications. The APE experiment data base holds a wealth of data that is now publicly available from the APE web site: http://climate.ncas.ac.uk/ape/. We hope that this Atlas will stimulate future analyses and investigations to understand the large variation seen in the model behaviors.
Resumo:
As a major mode of intraseasonal variability, which interacts with weather and climate systems on a near-global scale, the Madden – Julian Oscillation (MJO) is a crucial source of predictability for numerical weather prediction (NWP) models. Despite its global significance and comprehensive investigation, improvements in the representation of the MJO in an NWP context remain elusive. However, recent modifications to the model physics in the ECMWF model led to advances in the representation of atmospheric variability and the unprecedented propagation of the MJO signal through the entire integration period. In light of these recent advances, a set of hindcast experiments have been designed to assess the sensitivity of MJO simulation to the formulation of convection. Through the application of established MJO diagnostics, it is shown that the improvements in the representation of the MJO can be directly attributed to the modified convective parametrization. Furthermore, the improvements are attributed to the move from a moisture-convergent- to a relative-humidity-dependent formulation for organized deep entrainment. It is concluded that, in order to understand the physical mechanisms through which a relative-humidity-dependent formulation for entrainment led to an improved simulation of the MJO, a more process-based approach should be taken. T he application of process-based diagnostics t o t he hindcast experiments presented here will be the focus of Part II of this study.
Resumo:
In Part I of this study it was shown that moving from a moisture-convergent- to a relative-humidity-dependent organized entrainment rate in the formulation for deep convection was responsible for significant advances in the simulation of the Madden – Julian Oscillation (MJO) in the ECMWF model. However, the application of traditional MJO diagnostics were not adequate to understand why changing the control on convection had such a pronounced impact on the representation of the MJO. In this study a set of process-based diagnostics are applied to the hindcast experiments described in Part I to identify the physical mechanisms responsible for the advances in MJO simulation. Increasing the sensitivity of the deep convection scheme to environmental moisture is shown to modify the relationship between precipitation and moisture in the model. Through dry-air entrainment, convective plumes ascending in low-humidity environments terminate lower in the atmosphere. As a result, there is an increase in the occurrence of cumulus congestus, which acts to moisten the mid troposphere. Due to the modified precipitation – moisture relationship more moisture is able to build up, which effectively preconditions the tropical atmosphere for the t ransition t o d eep convection. R esults from this study suggest that a tropospheric moisture control on convection is key to simulating the interaction between the convective heating and the large-scale wave forcing associated with the MJO.
Resumo:
This work presents two schemes of measuring the linear and angular kinematics of a rigid body using a kinematically redundant array of triple-axis accelerometers with potential applications in biomechanics. A novel angular velocity estimation algorithm is proposed and evaluated that can compensate for angular velocity errors using measurements of the direction of gravity. Analysis and discussion of optimal sensor array characteristics are provided. A damped 2 axis pendulum was used to excite all 6 DoF of the a suspended accelerometer array through determined complex motion and is the basis of both simulation and experimental studies. The relationship between accuracy and sensor redundancy is investigated for arrays of up to 100 triple axis (300 accelerometer axes) accelerometers in simulation and 10 equivalent sensors (30 accelerometer axes) in the laboratory test rig. The paper also reports on the sensor calibration techniques and hardware implementation.