141 resultados para Simulation Modeling
Resumo:
Within the information systems field, the task of conceptual modeling involves building a representation of selected phenomena in some domain. High-quality conceptual-modeling work is important because it facilitates early detection and correction of system development errors. It also plays an increasingly important role in activities like business process reengineering and documentation of best-practice data and process models in enterprise resource planning systems. Yet little research has been undertaken on many aspects of conceptual modeling. In this paper, we propose a framework to motivate research that addresses the following fundamental question: How can we model the world to better facilitate our developing, implementing, using, and maintaining more valuable information systems? The framework comprises four elements: conceptual-modeling grammars, conceptual-modeling methods, conceptual-modeling scripts, and conceptual-modeling contexts. We provide examples of the types of research that have already been undertaken on each element and illustrate research opportunities that exist.
Resumo:
In modeling expectation formation, economic agents are usually viewed as forming expectations adaptively or in accordance with some rationality postulate. We offer an alternative nonlinear model where agents exchange their opinions and information with each other. Such a model yields multiple equilibria, or attracting distributions, that are persistent but subject to sudden large jumps. Using German Federal Statistical Office economic indicators and German IFO Poll expectational data, we show that this kind of model performs well in simulation experiments. Focusing upon producers' expectations in the consumption goods sector, we also discover evidence that structural change in the interactive process occurred over the period of investigation (1970-1998). Specifically, interactions in expectation formation seem to have become less important over time.
Resumo:
The splitting method is a simulation technique for the estimation of very small probabilities. In this technique, the sample paths are split into multiple copies, at various stages in the simulation. Of vital importance to the efficiency of the method is the Importance Function (IF). This function governs the placement of the thresholds or surfaces at which the paths are split. We derive a characterisation of the optimal IF and show that for multi-dimensional models the natural choice for the IF is usually not optimal. We also show how nearly optimal splitting surfaces can be derived or simulated using reverse time analysis. Our numerical experiments illustrate that by using the optimal IF, one can obtain a significant improvement in simulation efficiency.
Resumo:
We are witnessing an enormous growth in biological nitrogen removal from wastewater. It presents specific challenges beyond traditional COD (carbon) removal. A possibility for optimised process design is the use of biomass-supporting media. In this paper, attached growth processes (AGP) are evaluated using dynamic simulations. The advantages of these systems that were qualitatively described elsewhere, are validated quantitatively based on a simulation benchmark for activated sludge treatment systems. This simulation benchmark is extended with a biofilm model that allows for fast and accurate simulation of the conversion of different substrates in a biofilm. The economic feasibility of this system is evaluated using the data generated with the benchmark simulations. Capital savings due to volume reduction and reduced sludge production are weighed out against increased aeration costs. In this evaluation, effluent quality is integrated as well.
Resumo:
Models for the occurrence of the vibrational instability during rolling known as third octave chatter are presented and discussed. An analysis of rolling mill chatter was performed for the purpose of identifying characteristics of the vibrations and to determine any dependency on the rolling schedule. In particular, a stability criterion for the critical rolling speed is used to predict the maximum rolling speed without chatter instability on schedules from a 5 stand tandem mill rolling thin steel product. The results correlate well with measurements of critical speed occurring on the mill using a vibration monitor: This research provides significant insights into the chatter phenomena and has been used to investigate control methods for suppression of the instability.
Resumo:
The performance of the Oxford University Gun Tunnel has been estimated using a quasi-one-dimensional simulation of the facility gas dynamics. The modelling of the actual facility area variations so as to adequately simulate both shock reflection and flow discharge processes has been considered in some detail. Test gas stagnation pressure and temperature histories are compared with measurements at two different operating conditions - one with nitrogen and the other with carbon dioxide as the test gas. It is demonstrated that both the simulated pressures and temperatures are typically within 3% of the experimental measurements.
Resumo:
What interactions are sufficient to simulate arbitrary quantum dynamics in a composite quantum system? We provide an efficient algorithm to simulate any desired two-body Hamiltonian evolution using any fixed two-body entangling n-qubit Hamiltonian and local unitary operations. It follows that universal quantum computation can be performed using any entangling interaction and local unitary operations.
Resumo:
What interactions are sufficient to simulate arbitrary quantum dynamics in a composite quantum system? Dodd [Phys. Rev. A 65, 040301(R) (2002)] provided a partial solution to this problem in the form of an efficient algorithm to simulate any desired two-body Hamiltonian evolution using any fixed two-body entangling N-qubit Hamiltonian, and local unitaries. We extend this result to the case where the component systems are qudits, that is, have D dimensions. As a consequence we explain how universal quantum computation can be performed with any fixed two-body entangling N-qudit Hamiltonian, and local unitaries.
Resumo:
Developments in computer and three dimensional (3D) digitiser technologies have made it possible to keep track of the broad range of data required to simulate an insect moving around or over the highly heterogeneous habitat of a plant's surface. Properties of plant parts vary within a complex canopy architecture, and insect damage can induce further changes that affect an animal's movements, development and likelihood of survival. Models of plant architectural development based on Lindenmayer systems (L-systems) serve as dynamic platforms for simulation of insect movement, providing ail explicit model of the developing 3D structure of a plant as well as allowing physiological processes associated with plant growth and responses to damage to be described and Simulated. Simple examples of the use of the L-system formalism to model insect movement, operating Lit different spatial scales-from insects foraging on an individual plant to insects flying around plants in a field-are presented. Such models can be used to explore questions about the consequences of changes in environmental architecture and configuration on host finding, exploitation and its population consequences. In effect this model is a 'virtual ecosystem' laboratory to address local as well as landscape-level questions pertinent to plant-insect interactions, taking plant architecture into account. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
Models of plant architecture allow us to explore how genotype environment interactions effect the development of plant phenotypes. Such models generate masses of data organised in complex hierarchies. This paper presents a generic system for creating and automatically populating a relational database from data generated by the widely used L-system approach to modelling plant morphogenesis. Techniques from compiler technology are applied to generate attributes (new fields) in the database, to simplify query development for the recursively-structured branching relationship. Use of biological terminology in an interactive query builder contributes towards making the system biologist-friendly. (C) 2002 Elsevier Science Ireland Ltd. All rights reserved.
Resumo:
CULTURE is an Artificial Life simulation that aims to provide primary school children with opportunities to become actively engaged in the high-order thinking processes of problem solving and critical thinking. A preliminary evaluation of CULTURE has found that it offers the freedom for children to take part in process-oriented learning experiences. Through providing children with opportunities to make inferences, validate results, explain discoveries and analyse situations, CULTURE encourages the development of high-order thinking skills. The evaluation found that CULTURE allows users to autonomously explore the important scientific concepts of life and living, and energy and change within a software environment that children find enjoyable and easy to use.
Resumo:
This paper presents results on the simulation of the solid state sintering of copper wires using Monte Carlo techniques based on elements of lattice theory and cellular automata. The initial structure is superimposed onto a triangular, two-dimensional lattice, where each lattice site corresponds to either an atom or vacancy. The number of vacancies varies with the simulation temperature, while a cluster of vacancies is a pore. To simulate sintering, lattice sites are picked at random and reoriented in terms of an atomistic model governing mass transport. The probability that an atom has sufficient energy to jump to a vacant lattice site is related to the jump frequency, and hence the diffusion coefficient, while the probability that an atomic jump will be accepted is related to the change in energy of the system as a result of the jump, as determined by the change in the number of nearest neighbours. The jump frequency is also used to relate model time, measured in Monte Carlo Steps, to the actual sintering time. The model incorporates bulk, grain boundary and surface diffusion terms and includes vacancy annihilation on the grain boundaries. The predictions of the model were found to be consistent with experimental data, both in terms of the microstructural evolution and in terms of the sintering time. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
The retinoid orphan-related receptor-alpha (RORalpha) is a member of the ROR subfamily of orphan receptors and acts as a constitutive activator of transcription in the absence of exogenous ligands. To understand the basis of this activity, we constructed a homology model of Rill using the closely related TRbeta as a template. Molecular modeling suggested that bulky hydrophobic side chains occupy the RORa ligand cavity leaving a small but distinct cavity that may be involved in receptor stabilization. This model was subject to docking simulation with a receptor-interacting peptide from the steroid receptor coactivator, GR-interacting protein-1, which delineated a coactivator binding surface consisting of the signature motif spanning helices 3-5 and helix 12 [activation function 2 (AF2)]. Probing this surface with scanning alanine mutagenesis showed structural and functional equivalence between homologous residues of RORalpha and TRbeta. This was surprising (given that Rill is a ligand-independent activator, whereas TRbeta has an absolute requirement for ligand) and prompted us to use molecular modeling to identify differences between Rill and TRbeta in the way that the All helix interacts with the rest of the receptor. Modeling highlighted a nonconserved amino acid in helix 11 of RORa (Phe491) and a short-length of 3.10 helix at the N terminus of AF2 which we suggest i) ensures that AF2 is locked permanently in the holoconformation described for other liganded receptors and thus 2) enables ligand-independent recruitment of coactivators. Consistent with this, mutation of RORa Phe491 to either methionine or alanine (methionine is the homologous residue in TRbeta), reduced and ablated transcriptional activation and recruitment of coactivators, respectively. Furthermore, we were able to reconstitute transcriptional activity for both a deletion mutant of Ill lacking All and Phe491 Met, by overexpression of a GAL-AF2 fusion protein, demonstrating ligand-independent recruitment of AF2 and a role for Phe491 in recruiting AF2.
Resumo:
Multi-environment trials (METs) used to evaluate breeding lines vary in the number of years that they sample. We used a cropping systems model to simulate the target population of environments (TPE) for 6 locations over 108 years for 54 'near-isolines' of sorghum in north-eastern Australia. For a single reference genotype, each of 547 trials was clustered into 1 of 3 'drought environment types' (DETs) based on a seasonal water stress index. Within sequential METs of 2 years duration, the frequencies of these drought patterns often differed substantially from those derived for the entire TPE. This was reflected in variation in the mean yield of the reference genotype. For the TPE and for 2-year METs, restricted maximum likelihood methods were used to estimate components of genotypic and genotype by environment variance. These also varied substantially, although not in direct correlation with frequency of occurrence of different DETs over a 2-year period. Combined analysis over different numbers of seasons demonstrated the expected improvement in the correlation between MET estimates of genotype performance and the overall genotype averages as the number of seasons in the MET was increased.
Resumo:
This paper presents a new approach to the LU decomposition method for the simulation of stationary and ergodic random fields. The approach overcomes the size limitations of LU and is suitable for any size simulation. The proposed approach can facilitate fast updating of generated realizations with new data, when appropriate, without repeating the full simulation process. Based on a novel column partitioning of the L matrix, expressed in terms of successive conditional covariance matrices, the approach presented here demonstrates that LU simulation is equivalent to the successive solution of kriging residual estimates plus random terms. Consequently, it can be used for the LU decomposition of matrices of any size. The simulation approach is termed conditional simulation by successive residuals as at each step, a small set (group) of random variables is simulated with a LU decomposition of a matrix of updated conditional covariance of residuals. The simulated group is then used to estimate residuals without the need to solve large systems of equations.