993 resultados para Fire Dynamic Simulator
Resumo:
Manufacturing firms are driven by competitive pressures to continually improve the effectiveness and efficiency of their organisations. For this reason, manufacturing engineers often implement changes to existing processes, or design new production facilities, with the expectation of making further gains in manufacturing system performance. This thesis relates to how the likely outcome of this type of decision should be predicted prior to its implementation. The thesis argues that since manufacturing systems must also interact with many other parts of an organisation, the expected performance improvements can often be significantly hampered by constraints that arise elsewhere in the business. As a result, decision-makers should attempt to predict just how well a proposed design will perform when these other factors, or 'support departments', are taken into consideration. However, the thesis also demonstrates that, in practice, where quantitative analysis is used to evaluate design decisions, the analysis model invariably ignores the potential impact of support functions on a system's overall performance. A more comprehensive modelling approach is therefore required. A study of how various business functions interact establishes that to properly represent the kind of delays that give rise to support department constraints, a model should actually portray the dynamic and stochastic behaviour of entities in both the manufacturing and non-manufacturing aspects of a business. This implies that computer simulation be used to model design decisions but current simulation software does not provide a sufficient range of functionality to enable the behaviour of all of these entities to be represented in this way. The main objective of the research has therefore been the development of a new simulator that will overcome limitations of existing software and so enable decision-makers to conduct a more holistic evaluation of design decisions. It is argued that the application of object-oriented techniques offers a potentially better way of fulfilling both the functional and ease-of-use issues relating to development of the new simulator. An object-oriented analysis and design of the system, called WBS/Office, are therefore presented that extends to modelling a firm's administrative and other support activities in the context of the manufacturing system design process. A particularly novel feature of the design is the ability for decision-makers to model how a firm's specific information and document processing requirements might hamper shop-floor performance. The simulator is primarily intended for modelling make-to-order batch manufacturing systems and the thesis presents example models created using a working version of WBS/Office that demonstrate the feasibility of using the system to analyse manufacturing system designs in this way.
Resumo:
In analysing manufacturing systems, for either design or operational reasons, failure to account for the potentially significant dynamics could produce invalid results. There are many analysis techniques that can be used, however, simulation is unique in its ability to assess detailed, dynamic behaviour. The use of simulation to analyse manufacturing systems would therefore seem appropriate if not essential. Many simulation software products are available but their ease of use and scope of application vary greatly. This is illustrated at one extreme by simulators which offer rapid but limited application whilst at the other simulation languages which are extremely flexible but tedious to code. Given that a typical manufacturing engineer does not posses in depth programming and simulation skills then the use of simulators over simulation languages would seem a more appropriate choice. Whilst simulators offer ease of use their limited functionality may preclude their use in many applications. The construction of current simulators makes it difficult to amend or extend the functionality of the system to meet new challenges. Some simulators could even become obsolete as users, demand modelling functionality that reflects the latest manufacturing system design and operation concepts. This thesis examines the deficiencies in current simulation tools and considers whether they can be overcome by the application of object-oriented principles. Object-oriented techniques have gained in popularity in recent years and are seen as having the potential to overcome any of the problems traditionally associated with software construction. There are a number of key concepts that are exploited in the work described in this thesis: the use of object-oriented techniques to act as a framework for abstracting engineering concepts into a simulation tool and the ability to reuse and extend object-oriented software. It is argued that current object-oriented simulation tools are deficient and that in designing such tools, object -oriented techniques should be used not just for the creation of individual simulation objects but for the creation of the complete software. This results in the ability to construct an easy to use simulator that is not limited by its initial functionality. The thesis presents the design of an object-oriented data driven simulator which can be freely extended. Discussion and work is focused on discrete parts manufacture. The system developed retains the ease of use typical of data driven simulators. Whilst removing any limitation on its potential range of applications. Reference is given to additions made to the simulator by other developers not involved in the original software development. Particular emphasis is put on the requirements of the manufacturing engineer and the need for Ihe engineer to carrv out dynamic evaluations.
Resumo:
This research explores Bayesian updating as a tool for estimating parameters probabilistically by dynamic analysis of data sequences. Two distinct Bayesian updating methodologies are assessed. The first approach focuses on Bayesian updating of failure rates for primary events in fault trees. A Poisson Exponentially Moving Average (PEWMA) model is implemnented to carry out Bayesian updating of failure rates for individual primary events in the fault tree. To provide a basis for testing of the PEWMA model, a fault tree is developed based on the Texas City Refinery incident which occurred in 2005. A qualitative fault tree analysis is then carried out to obtain a logical expression for the top event. A dynamic Fault Tree analysis is carried out by evaluating the top event probability at each Bayesian updating step by Monte Carlo sampling from posterior failure rate distributions. It is demonstrated that PEWMA modeling is advantageous over conventional conjugate Poisson-Gamma updating techniques when failure data is collected over long time spans. The second approach focuses on Bayesian updating of parameters in non-linear forward models. Specifically, the technique is applied to the hydrocarbon material balance equation. In order to test the accuracy of the implemented Bayesian updating models, a synthetic data set is developed using the Eclipse reservoir simulator. Both structured grid and MCMC sampling based solution techniques are implemented and are shown to model the synthetic data set with good accuracy. Furthermore, a graphical analysis shows that the implemented MCMC model displays good convergence properties. A case study demonstrates that Likelihood variance affects the rate at which the posterior assimilates information from the measured data sequence. Error in the measured data significantly affects the accuracy of the posterior parameter distributions. Increasing the likelihood variance mitigates random measurement errors, but casuses the overall variance of the posterior to increase. Bayesian updating is shown to be advantageous over deterministic regression techniques as it allows for incorporation of prior belief and full modeling uncertainty over the parameter ranges. As such, the Bayesian approach to estimation of parameters in the material balance equation shows utility for incorporation into reservoir engineering workflows.
Resumo:
Many-core systems are emerging from the need of more computational power and power efficiency. However there are many issues which still revolve around the many-core systems. These systems need specialized software before they can be fully utilized and the hardware itself may differ from the conventional computational systems. To gain efficiency from many-core system, programs need to be parallelized. In many-core systems the cores are small and less powerful than cores used in traditional computing, so running a conventional program is not an efficient option. Also in Network-on-Chip based processors the network might get congested and the cores might work at different speeds. In this thesis is, a dynamic load balancing method is proposed and tested on Intel 48-core Single-Chip Cloud Computer by parallelizing a fault simulator. The maximum speedup is difficult to obtain due to severe bottlenecks in the system. In order to exploit all the available parallelism of the Single-Chip Cloud Computer, a runtime approach capable of dynamically balancing the load during the fault simulation process is used. The proposed dynamic fault simulation approach on the Single-Chip Cloud Computer shows up to 45X speedup compared to a serial fault simulation approach. Many-core systems can draw enormous amounts of power, and if this power is not controlled properly, the system might get damaged. One way to manage power is to set power budget for the system. But if this power is drawn by just few cores of the many, these few cores get extremely hot and might get damaged. Due to increase in power density multiple thermal sensors are deployed on the chip area to provide realtime temperature feedback for thermal management techniques. Thermal sensor accuracy is extremely prone to intra-die process variation and aging phenomena. These factors lead to a situation where thermal sensor values drift from the nominal values. This necessitates efficient calibration techniques to be applied before the sensor values are used. In addition, in modern many-core systems cores have support for dynamic voltage and frequency scaling. Thermal sensors located on cores are sensitive to the core's current voltage level, meaning that dedicated calibration is needed for each voltage level. In this thesis a general-purpose software-based auto-calibration approach is also proposed for thermal sensors to calibrate thermal sensors on different range of voltages.
Resumo:
This paper describes two new techniques designed to enhance the performance of fire field modelling software. The two techniques are "group solvers" and automated dynamic control of the solution process, both of which are currently under development within the SMARTFIRE Computational Fluid Dynamics environment. The "group solver" is a derivation of common solver techniques used to obtain numerical solutions to the algebraic equations associated with fire field modelling. The purpose of "group solvers" is to reduce the computational overheads associated with traditional numerical solvers typically used in fire field modelling applications. In an example, discussed in this paper, the group solver is shown to provide a 37% saving in computational time compared with a traditional solver. The second technique is the automated dynamic control of the solution process, which is achieved through the use of artificial intelligence techniques. This is designed to improve the convergence capabilities of the software while further decreasing the computational overheads. The technique automatically controls solver relaxation using an integrated production rule engine with a blackboard to monitor and implement the required control changes during solution processing. Initial results for a two-dimensional fire simulation are presented that demonstrate the potential for considerable savings in simulation run-times when compared with control sets from various sources. Furthermore, the results demonstrate the potential for enhanced solution reliability due to obtaining acceptable convergence within each time step, unlike some of the comparison simulations.
Resumo:
The aim of this thesis is to test the ability of some correlative models such as Alpert correlations on 1972 and re-examined on 2011, the investigation of Heskestad and Delichatsios in 1978, the correlations produced by Cooper in 1982, to define both dynamic and thermal characteristics of a fire induced ceiling-jet flow. The flow occurs when the fire plume impinges the ceiling and develops in the radial direction of the fire axis. Both temperature and velocity predictions are decisive for sprinklers positioning, fire alarms positions, detectors (heat, smoke) positions and activation times and back-layering predictions. These correlative models will be compared with a 3D numerical simulation software CFAST. For the results comparison of temperature and velocity near the ceiling. These results are also compared with a Computational Fluid Dynamics (CFD) analysis, using ANSYS FLUENT.
Resumo:
We present a bidomain threshold model of intracellular calcium (Ca²⁺) dynamics in which, as suggested by recent experiments, the cytosolic threshold for Ca²⁺ liberation is modulated by the Ca²⁺ concentration in the releasing compartment. We explicitly construct stationary fronts and determine their stability using an Evans function approach. Our results show that a biologically motivated choice of a dynamic threshold, as opposed to a constant threshold, can pin stationary fronts that would otherwise be unstable. This illustrates a novel mechanism to stabilise pinned interfaces in continuous excitable systems. Our framework also allows us to compute travelling pulse solutions in closed form and systematically probe the wave speed as a function of physiologically important parameters. We find that the existence of travelling wave solutions depends on the time scale of the threshold dynamics, and that facilitating release by lowering the cytosolic threshold increases the wave speed. The construction of the Evans function for a travelling pulse shows that of the co-existing fast and slow solutions the slow one is always unstable.
Resumo:
Mestrado Mediterranean Forestry and Natural Resources Management - Instituto Superior de Agronomia - UL
Resumo:
The Chihuahua desert is one of the most biologically diverse ecosystems in the world, but suffers serious degradation because of changes in fire regimes resulting in large catastrophic fires. My study was conducted in the Sierra La Mojonera (SLM) natural protected area in Mexico. The purpose of this study was to implement the use of FARSITE fire modeling as a fire management tool to develop an integrated fire management plan at SLM. Firebreaks proved to detain 100% of wildfire outbreaks. The rosetophilous scrub experienced the fastest rate of fire spread and lowland creosote bush scrub experienced the slowest rate of fire spread. March experienced the fastest rate of fire spread, while September experienced the slowest rate of fire spread. The results of my study provide a tool for wildfire management through the use geospatial technologies and, in particular, FARSITE fire modeling in SLM and Mexico.
Resumo:
TiO2 and TiO2/WO3 electrodes, irradiated by a solar simulator in configurations for heterogeneous photocatalysis (HP) and electrochemically-assisted HP (EHP), were used to remediate aqueous solutions containing 10 mg L(-1) (34 μmol L(-1)) of 17-α-ethinylestradiol (EE2), active component of most oral contraceptives. The photocatalysts consisted of 4.5 μm thick porous films of TiO2 and TiO2/WO3 (molar ratio W/Ti of 12%) deposited on transparent electrodes from aqueous suspensions of TiO2 particles and WO3 precursors, followed by thermal treatment at 450 (°)C. First, an energy diagram was organized with photoelectrochemical and UV-Vis absorption spectroscopy data and revealed that EE2 could be directly oxidized by the photogenerated holes at the semiconductor surfaces, considering the relative HOMO level for EE2 and the semiconductor valence band edges. Also, for the irradiated hybrid photocatalyst, electrons in TiO2 should be transferred to WO3 conduction band, while holes move toward TiO2 valence band, improving charge separation. The remediated EE2 solutions were analyzed by fluorescence, HPLC and total organic carbon measurements. As expected from the energy diagram, both photocatalysts promoted the EE2 oxidation in HP configuration; after 4 h, the EE2 concentration decayed to 6.2 mg L(-1) (35% of EE2 removal) with irradiated TiO2 while TiO2/WO3 electrode resulted in 45% EE2 removal. A higher performance was achieved in EHP systems, when a Pt wire was introduced as a counter-electrode and the photoelectrodes were biased at +0.7 V; then, the EE2 removal corresponded to 48 and 54% for the TiO2 and TiO2/WO3, respectively. The hybrid TiO2/WO3, when compared to TiO2 electrode, exhibited enhanced sunlight harvesting and improved separation of photogenerated charge carriers, resulting in higher performance for removing this contaminant of emerging concern from aqueous solution.
Resumo:
Current data indicate that the size of high-density lipoprotein (HDL) may be considered an important marker for cardiovascular disease risk. We established reference values of mean HDL size and volume in an asymptomatic representative Brazilian population sample (n=590) and their associations with metabolic parameters by gender. Size and volume were determined in HDL isolated from plasma by polyethyleneglycol precipitation of apoB-containing lipoproteins and measured using the dynamic light scattering (DLS) technique. Although the gender and age distributions agreed with other studies, the mean HDL size reference value was slightly lower than in some other populations. Both HDL size and volume were influenced by gender and varied according to age. HDL size was associated with age and HDL-C (total population); non- white ethnicity and CETP inversely (females); HDL-C and PLTP mass (males). On the other hand, HDL volume was determined only by HDL-C (total population and in both genders) and by PLTP mass (males). The reference values for mean HDL size and volume using the DLS technique were established in an asymptomatic and representative Brazilian population sample, as well as their related metabolic factors. HDL-C was a major determinant of HDL size and volume, which were differently modulated in females and in males.
Resumo:
The objective of this study is to verify the dynamics between fiscal policy, measured by public debt, and monetary policy, measured by a reaction function of a central bank. Changes in monetary policies due to deviations from their targets always generate fiscal impacts. We examine two policy reaction functions: the first related to inflation targets and the second related to economic growth targets. We find that the condition for stable equilibrium is more restrictive in the first case than in the second. We then apply our simulation model to Brazil and United Kingdom and find that the equilibrium is unstable in the Brazilian case but stable in the UK case.
Resumo:
Cancer is a multistep process that begins with the transformation of normal epithelial cells and continues with tumor growth, stromal invasion and metastasis. The remodeling of the peritumoral environment is decisive for the onset of tumor invasiveness. This event is dependent on epithelial-stromal interactions, degradation of extracellular matrix components and reorganization of fibrillar components. Our research group has studied in a new proposed rodent model the participation of cellular and molecular components in the prostate microenvironment that contributes to cancer progression. Our group adopted the gerbil Meriones unguiculatus as an alternative experimental model for prostate cancer study. This model has presented significant responses to hormonal treatments and to development of spontaneous and induced neoplasias. The data obtained indicate reorganization of type I collagen fibers and reticular fibers, synthesis of new components such as tenascin and proteoglycans, degradation of basement membrane components and elastic fibers and increased expression of metalloproteinases. Fibroblasts that border the region, apparently participate in the stromal reaction. The roles of each of these events, as well as some signaling molecules, participants of neoplastic progression and factors that promote genetic reprogramming during epithelial-stromal transition are also discussed.
Resumo:
Yellow passion fruit pulp is unstable, presenting phase separation that can be avoided by the addition of hydrocolloids. For this purpose, xanthan and guar gum [0.3, 0.7 and 1.0% (w/w)] were added to yellow passion fruit pulp and the changes in the dynamic and steady - shear rheological behavior evaluated. Xanthan dispersions showed a more pronounced pseudoplasticity and the presence of yield stress, which was not observed in the guar gum dispersions. Cross model fitting to flow curves showed that the xanthan suspensions also had higher zero shear viscosity than the guar suspensions, and, for both gums, an increase in temperature led to lower values for this parameter. The gums showed different behavior as a function of temperature in the range of 5 - 35ºC. The activation energy of the apparent viscosity was dependent on the shear rate and gum concentration for guar, whereas for xanthan these values only varied with the concentration. The mechanical spectra were well described by the generalized Maxwell model and the xanthan dispersions showed a more elastic character than the guar dispersions, with higher values for the relaxation time. Xanthan was characterized as a weak gel, while guar presented a concentrated solution behavior. The simultaneous evaluation of temperature and concentration showed a stronger influence of the polysaccharide concentration on the apparent viscosity and the G' and G" moduli than the variation in temperature.
Resumo:
Fire management is a common practice in several reserves in the Cerrado, but its influences on bird reproduction remain unknown. In addition, the nesting biology of the Burrowing Owl (Athene cunicularia) has been studied in numerous environments, but not in tropical grasslands managed by fire. This study examined the effects of fire management on the nesting biology of A. cunicularia in Emas National Park, State of Goias, central Brazilian Cerrado. We compared the number of breeding pairs and their burrows in October and November 2009 at 15 study sites in grasslands managed by fire (firebreaks) and unmanaged grasslands adjacent to and distant from firebreaks. We visited active burrows two-four times and described the burrow entrances and sentinel sites and counted and observed adults and young. A total of 19 burrows were found at firebreaks. One and two burrows were found in grasslands adjacent to and distant from firebreaks, respectively. For all burrows found, one to three young reached the adult size, being able to fly and/or run in early November. The 22 burrows found were in the ground, associated or not with termite and ant nests. Most (86.4%) burrows had only one entrance. Only three burrows had two or three entrances. Structures used as sentinel perches by adults were mounds in front of the burrow entrances, termite nests, shrubs and trees. Most of these sentinel sites were shorter than 2 m high and located less than 10 m from the burrow entrance. At Emas National Park, firebreaks appear to provide more attractive conditions to the nesting of A. cunicularia than unmanaged grasslands, likely because of the short herbaceous stratum due to frequent burning of firebreaks. This study suggests that fire management provides suitable conditions for the successful reproduction of A. cunicularia in firebreaks at Emas National Park.