918 resultados para Simulation modelling
Resumo:
Mineralogical analysis is often used to assess the liberation properties of particles. A direct method of estimating liberation is to actually break particles and then directly obtain liberation information from applying mineralogical analysis to each size-class of the product. Another technique is to artificially apply random breakage to the feed particle sections to estimate the resultant distribution of product particle sections. This technique provides a useful alternative estimation method. Because this technique is applied to particle sections, the actual liberation properties for particles can only be estimated by applying stereological correction. A recent stereological technique has been developed that allows the discrepancy between the linear intercept composition distribution and the particle section composition distribution to be used as guide for estimating the particle composition distribution. The paper will show results validating this new technique using numerical simulation. (C) 2004 Elsevier Ltd. All rights reserved.
Resumo:
In recent years, many sorghum producers in the more marginal (
Resumo:
The work reported in this paper is part of a project simulating maintenance operations in an automotive engine production facility. The decisions made by the people in charge of these operations form a crucial element of this simulation. Eliciting this knowledge is problematic. One approach is to use the simulation model as part of the knowledge elicitation process. This paper reports on the experience so far with using a simulation model to support knowledge management in this way. Issues are discussed regarding the data available, the use of the model, and the elicitation process itself. © 2004 Elsevier B.V. All rights reserved.
Resumo:
A discrete event simulation model was developed and used to estimate the storage area required for a proposed overseas textile manufacturing facility. It was found that the simulation was able to achieve this because of its ability to both store attribute values and to show queuing levels at an individual product level. It was also found that the process of undertaking the simulation project initiated useful discussions regarding the operation of the facility. Discrete event simulation is shown to be much more than an exercise in quantitative analysis of results and an important task of the simulation project manager is to initiate a debate among decision makers regarding the assumptions of how the system operates.
Resumo:
The performance of direct workers has a significant impact on the competitiveness of many manufacturing systems. Unfortunately, system designers are ill equipped to assess this impact during the design process. An opportunity exists to assist designers by expanding the capabilities of popular simulation modelling tools, and using them as a vehicle to better consider human factors during the process of system design manufacture. To support this requirement, this paper reports on an extensive review of literature that develops a theoretical framework, which summarizes the principal factors and relationships that such a modelling tool should incorporate.
Resumo:
Computer based discrete event simulation (DES) is one of the most commonly used aids for the design of automotive manufacturing systems. However, DES tools represent machines in extensive detail, while only representing workers as simple resources. This presents a problem when modelling systems with a highly manual work content, such as an assembly line. This paper describes research at Cranfield University, in collaboration with the Ford Motor Company, founded on the assumption that human variation is the cause of a large percentage of the disparity between simulation predictions and real world performance. The research aims to improve the accuracy and reliability of simulation prediction by including models of human factors.
Resumo:
Enabling a Simulation Capability in the Organisation addresses the application of simulation modelling techniques in order to enable better informed decisions in business and industrial organisations. The book’s unique approach treats simulation not just as a technical tool, but within as a support for organisational decision making, showing the results from a survey of current and potential users of simulation to suggest reasons why the technique is not used as much as it should be and what are the barriers to its further use. By incorporating an evaluation of six detailed case studies of the application of simulation in industry by the author, the book will teach readers: •the role of simulation in decision making; •how to introduce simulation as a tool for decision making; and •how to undertake simulation studies that lead to change in the organisation. Enabling a Simulation Capability in the Organisation provides an introduction to the state of the art in simulation modelling for researchers in business studies and engineering, as well a useful guide to practitioners and managers in business and industry.
Resumo:
Climate change is one of the most crucial ecological problems of our age with great influence. Seasonal dynamics of aquatic communities are — among others — regulated by the climate, especially by temperature. In this case study we attempted the simulation modelling of the seasonal dynamics of a copepod species, Cyclops vicinus, which ranks among the zooplankton community, based on a quantitative database containing ten years of data from the Danube’s Göd area. We set up a simulation model predicting the abundance of Cyclops vicinus by considering only temperature as it affects the abundance of population. The model was adapted to eight years of daily temperature data observed between 1981 and 1994 and was tested successfully with the additional data of two further years. The model was run with the data series of climate change scenarios specified for the period around 2070- 2100. On the other hand we looked for the geographically analogous areas with the Göd region which are mostly similar to the future climate of the Göd area. By means of the above-mentioned points we can get a view how the climate of the region will change by the end of the 21st century, and the way the seasonal dynamics of a chosen planktonic crustacean species may follow this change. According to our results the area of Göd will be similar to the northern region of Greece. The maximum abundance of the examined species occurs a month to one and a half months earlier, moreover larger variances are expected between years in respect of the abundance. The deviations are expected in the direction of smaller or significantly larger abundance not observed earlier.
Resumo:
Knowledge on the expected effects of climate change on aquatic ecosystems is defined by three ways. On the one hand, long-term observation in the field serves as a basis for the possible changes; on the other hand, the experimental approach may bring valuable pieces of information to the research field. The expected effects of climate change cannot be studied by empirical approach; rather mathematical models are useful tools for this purpose. Within this study, the main findings of field observations and their implications for future were summarized; moreover, the modelling approaches were discussed in a more detailed way. Some models try to describe the variation of physical parameters in a given aquatic habitat, thus our knowledge on their biota is confined to the findings based on our present observations. Others are destined for answering special issues related to the given water body. Complex ecosystem models are the keys of our better understanding of the possible effects of climate change. Basically, these models were not created for testing the influence of global warming, rather focused on the description of a complex system (e. g. a lake) involving environmental variables, nutrients. However, such models are capable of studying climatic changes as well by taking into consideration a large set of environmental variables. Mostly, the outputs are consistent with the assumptions based on the findings in the field. Since synthetized models are rather difficult to handle and require quite large series of data, the authors proposed a more simple modelling approach, which is capable of examining the effects of global warming. This approach includes weather dependent simulation modelling of the seasonal dynamics of aquatic organisms within a simplified framework.
Resumo:
This paper reports on an attempt to apply Genetic Algorithms to the problem of optimising a complex system, through discrete event simulation (Simulation Optimisation), with a view to reducing the noise associated with such a procedure. We are applying this proposed solution approach to our application test bed, a Crossdocking distribution centre, because it provides a good representative of the random and unpredictable behaviour of complex systems i.e. automated machine random failure and the variability of manual order picker skill. It is known that there is noise in the output of discrete event simulation modelling. However, our interest focuses on the effect of noise on the evaluation of the fitness of candidate solutions within the search space, and the development of techniques to handle this noise. The unique quality of our proposed solution approach is we intend to embed a noise reduction technique in our Genetic Algorithm based optimisation procedure, in order for it to be robust enough to handle noise, efficiently estimate suitable fitness function, and produce good quality solutions with minimal computational effort.
Resumo:
When designing systems that are complex, dynamic and stochastic in nature, simulation is generally recognised as one of the best design support technologies, and a valuable aid in the strategic and tactical decision making process. A simulation model consists of a set of rules that define how a system changes over time, given its current state. Unlike analytical models, a simulation model is not solved but is run and the changes of system states can be observed at any point in time. This provides an insight into system dynamics rather than just predicting the output of a system based on specific inputs. Simulation is not a decision making tool but a decision support tool, allowing better informed decisions to be made. Due to the complexity of the real world, a simulation model can only be an approximation of the target system. The essence of the art of simulation modelling is abstraction and simplification. Only those characteristics that are important for the study and analysis of the target system should be included in the simulation model. The purpose of simulation is either to better understand the operation of a target system, or to make predictions about a target system’s performance. It can be viewed as an artificial white-room which allows one to gain insight but also to test new theories and practices without disrupting the daily routine of the focal organisation. What you can expect to gain from a simulation study is very well summarised by FIRMA (2000). His idea is that if the theory that has been framed about the target system holds, and if this theory has been adequately translated into a computer model this would allow you to answer some of the following questions: · Which kind of behaviour can be expected under arbitrarily given parameter combinations and initial conditions? · Which kind of behaviour will a given target system display in the future? · Which state will the target system reach in the future? The required accuracy of the simulation model very much depends on the type of question one is trying to answer. In order to be able to respond to the first question the simulation model needs to be an explanatory model. This requires less data accuracy. In comparison, the simulation model required to answer the latter two questions has to be predictive in nature and therefore needs highly accurate input data to achieve credible outputs. These predictions involve showing trends, rather than giving precise and absolute predictions of the target system performance. The numerical results of a simulation experiment on their own are most often not very useful and need to be rigorously analysed with statistical methods. These results then need to be considered in the context of the real system and interpreted in a qualitative way to make meaningful recommendations or compile best practice guidelines. One needs a good working knowledge about the behaviour of the real system to be able to fully exploit the understanding gained from simulation experiments. The goal of this chapter is to brace the newcomer to the topic of what we think is a valuable asset to the toolset of analysts and decision makers. We will give you a summary of information we have gathered from the literature and of the experiences that we have made first hand during the last five years, whilst obtaining a better understanding of this exciting technology. We hope that this will help you to avoid some pitfalls that we have unwittingly encountered. Section 2 is an introduction to the different types of simulation used in Operational Research and Management Science with a clear focus on agent-based simulation. In Section 3 we outline the theoretical background of multi-agent systems and their elements to prepare you for Section 4 where we discuss how to develop a multi-agent simulation model. Section 5 outlines a simple example of a multi-agent system. Section 6 provides a collection of resources for further studies and finally in Section 7 we will conclude the chapter with a short summary.
Resumo:
In this paper a bond graph methodology is used to model incompressible fluid flows with viscous and thermal effects. The distinctive characteristic of these flows is the role of pressure, which does not behave as a state variable but as a function that must act in such a way that the resulting velocity field has divergence zero. Velocity and entropy per unit volume are used as independent variables for a single-phase, single-component flow. Time-dependent nodal values and interpolation functions are introduced to represent the flow field, from which nodal vectors of velocity and entropy are defined as state variables. The system for momentum and continuity equations is coincident with the one obtained by using the Galerkin method for the weak formulation of the problem in finite elements. The integral incompressibility constraint is derived based on the integral conservation of mechanical energy. The weak formulation for thermal energy equation is modeled with true bond graph elements in terms of nodal vectors of temperature and entropy rates, resulting a Petrov-Galerkin method. The resulting bond graph shows the coupling between mechanical and thermal energy domains through the viscous dissipation term. All kind of boundary conditions are handled consistently and can be represented as generalized effort or flow sources. A procedure for causality assignment is derived for the resulting graph, satisfying the Second principle of Thermodynamics. (C) 2007 Elsevier B.V. All rights reserved.
Resumo:
This technical appendix details the methods used in an assessment of the potential of snus for tobacco harm reduction using simulation modelling.
Resumo:
Objective: The Assessing Cost-Effectiveness - Mental Health (ACE-MH) study aims to assess from a health sector perspective, whether there are options for change that could improve the effectiveness and efficiency of Australia's current mental health services by directing available resources toward 'best practice' cost-effective services. Method: The use of standardized evaluation methods addresses the reservations expressed by many economists about the simplistic use of League Tables based on economic studies confounded by differences in methods, context and setting. The cost-effectiveness ratio for each intervention is calculated using economic and epidemiological data. This includes systematic reviews and randomised controlled trials for efficacy, the Australian Surveys of Mental Health and Wellbeing for current practice and a combination of trials and longitudinal studies for adherence. The cost-effectiveness ratios are presented as cost (A$) per disability-adjusted life year (DALY) saved with a 95% uncertainty interval based on Monte Carlo simulation modelling. An assessment of interventions on 'second filter' criteria ('equity', 'strength of evidence', 'feasibility' and 'acceptability to stakeholders') allows broader concepts of 'benefit' to be taken into account, as well as factors that might influence policy judgements in addition to cost-effectiveness ratios. Conclusions: The main limitation of the study is in the translation of the effect size from trials into a change in the DALY disability weight, which required the use of newly developed methods. While comparisons within disorders are valid, comparisons across disorders should be made with caution. A series of articles is planned to present the results.