858 resultados para Multi-agent simulation
Resumo:
Modern manufacturing systems should satisfy emerging needs related to sustainable development. The design of sustainable manufacturing systems can be valuably supported by simulation, traditionally employed mainly for time and cost reduction. In this paper, a multi-purpose digital simulation approach is proposed to deal with sustainable manufacturing systems design through Discrete Event Simulation (DES) and 3D digital human modelling. DES models integrated with data on power consumption of the manufacturing equipment are utilized to simulate different scenarios with the aim to improve productivity as well as energy efficiency, avoiding resource and energy waste. 3D simulation based on digital human modelling is employed to assess human factors issues related to ergonomics and safety of manufacturing systems. The approach is implemented for the sustainability enhancement of a real manufacturing cell of the aerospace industry, automated by robotic deburring. Alternative scenarios are proposed and simulated, obtaining a significant improvement in terms of energy efficiency (−87%) for the new deburring cell, and a reduction of energy consumption around −69% for the coordinate measuring machine, with high potential annual energy cost savings and increased energy efficiency. Moreover, the simulation-based ergonomic assessment of human operator postures allows 25% improvement of the workcell ergonomic index.
Resumo:
Models of neutrino-driven core-collapse supernova explosions have matured considerably in recent years. Explosions of low-mass progenitors can routinely be simulated in 1D, 2D, and 3D. Nucleosynthesis calculations indicate that these supernovae could be contributors of some lighter neutron-rich elements beyond iron. The explosion mechanism of more massive stars remains under investigation, although first 3D models of neutrino-driven explosions employing multi-group neutrino transport have become available. Together with earlier 2D models and more simplified 3D simulations, these have elucidated the interplay between neutrino heating and hydrodynamic instabilities in the post-shock region that is essential for shock revival. However, some physical ingredients may still need to be added/improved before simulations can robustly explain supernova explosions over a wide range of progenitors. Solutions recently suggested in the literature include uncertainties in the neutrino rates, rotation, and seed perturbations from convective shell burning. We review the implications of 3D simulations of shell burning in supernova progenitors for the ‘perturbations-aided neutrino-driven mechanism,’ whose efficacy is illustrated by the first successful multi-group neutrino hydrodynamics simulation of an 18 solar mass progenitor with 3D initial conditions. We conclude with speculations about the impact of 3D effects on the structure of massive stars through convective boundary mixing.
Resumo:
the work towards increased energy efficiency. In order to plan and perform effective energy renovation of the buildings, it is necessary to have adequate information on the current status of the buildings in terms of architectural features and energy needs. Unfortunately, the official statistics do not include all of the needed information for the whole building stock. This paper aims to fill the gaps in the statistics by gathering data from studies, projects and national energy agencies, and by calibrating TRNSYS models against the existing data to complete missing energy demand data, for countries with similar climate, through simulation. The survey was limited to residential and office buildings in the EU member states (before July 2013). This work was carried out as part of the EU FP7 project iNSPiRe. The building stock survey revealed over 70% of the residential and office floor area is concentrated in the six most populated countries. The total energy consumption in the residential sector is 14 times that of the office sector. In the residential sector, single family houses represent 60% of the heated floor area, albeit with different share in the different countries, indicating that retrofit solutions cannot be focused only on multi-family houses. The simulation results indicate that residential buildings in central and southern European countries are not always heated to 20 °C, but are kept at a lower temperature during at least part of the day. Improving the energy performance of these houses through renovation could allow the occupants to increase the room temperature and improve their thermal comfort, even though the potential for energy savings would then be reduced.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
Résumé: L’Institut pour l'étude de la neige et des avalanches en Suisse (SLF) a développé SNOWPACK, un modèle thermodynamique multi-couches de neige permettant de simuler les propriétés géophysiques du manteau neigeux (densité, température, taille de grain, teneur en eau, etc.) à partir desquelles un indice de stabilité est calculé. Il a été démontré qu’un ajustement de la microstructure serait nécessaire pour une implantation au Canada. L'objectif principal de la présente étude est de permettre au modèle SNOWPACK de modéliser de manière plus réaliste la taille de grain de neige et ainsi obtenir une prédiction plus précise de la stabilité du manteau neigeux à l’aide de l’indice basé sur la taille de grain, le Structural Stability Index (SSI). Pour ce faire, l’erreur modélisée (biais) par le modèle a été analysée à l’aide de données précises sur le terrain de la taille de grain à l’aide de l’instrument IRIS (InfraRed Integrated Sphere). Les données ont été recueillies durant l’hiver 2014 à deux sites différents au Canada : parc National des Glaciers, en Colombie-Britannique ainsi qu’au parc National de Jasper. Le site de Fidelity était généralement soumis à un métamorphisme à l'équilibre tandis que celui de Jasper à un métamorphisme cinétique plus prononcé. Sur chacun des sites, la stratigraphie des profils de densités ainsi des profils de taille de grain (IRIS) ont été complétés. Les profils de Fidelity ont été complétés avec des mesures de micropénétromètre (SMP). L’analyse des profils de densité a démontré une bonne concordance avec les densités modélisées (R[indice supérieur 2]=0.76) et donc la résistance simulée pour le SSI a été jugée adéquate. Les couches d’instabilités prédites par SNOWPACK ont été identifiées à l’aide de la variation de la résistance dans les mesures de SMP. L’analyse de la taille de grain optique a révélé une surestimation systématique du modèle ce qui est en accord avec la littérature. L’erreur de taille de grain optique dans un environnement à l’équilibre était assez constante tandis que l’erreur en milieux cinétique était plus variable. Finalement, une approche orientée sur le type de climat représenterait le meilleur moyen pour effectuer une correction de la taille de grain pour une évaluation de la stabilité au Canada.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
One of the core tasks of the virtual-manufacturing environment is to characterise the transformation of the state of material during each of the unit processes. This transformation in shape, material properties, etc. can only be reliably achieved through the use of models in a simulation context. Unfortunately, many manufacturing processes involve the material being treated in both the liquid and solid state, the trans-formation of which may be achieved by heat transfer and/or electro-magnetic fields. The computational modelling of such processes, involving the interactions amongst various interacting phenomena, is a consider-able challenge. However, it must be addressed effectively if Virtual Manufacturing Environments are to become a reality! This contribution focuses upon one attempt to develop such a multi-physics computational toolkit. The approach uses a single discretisation procedure and provides for direct interaction amongst the component phenomena. The need to exploit parallel high performance hardware is addressed so that simulation elapsed times can be brought within the realms of practicality. Examples of Multiphysics modelling in relation to shape casting, and solder joint formation reinforce the motivation for this work.
Resumo:
The main drivers for the development and evolution of Cyber Physical Systems (CPS) are the reduction of development costs and time along with the enhancement of the designed products. The aim of this survey paper is to provide an overview of different types of system and the associated transition process from mechatronics to CPS and cloud-based (IoT) systems. It will further consider the requirement that methodologies for CPS-design should be part of a multi-disciplinary development process within which designers should focus not only on the separate physical and computational components, but also on their integration and interaction. Challenges related to CPS-design are therefore considered in the paper from the perspectives of the physical processes, computation and integration respectively. Illustrative case studies are selected from different system levels starting with the description of the overlaying concept of Cyber Physical Production Systems (CPPSs). The analysis and evaluation of the specific properties of a sub-system using a condition monitoring system, important for the maintenance purposes, is then given for a wind turbine.
Resumo:
The major function of this model is to access the UCI Wisconsin Breast Cancer data-set[1] and classify the data items into two categories, which are normal and anomalous. This kind of classification can be referred as anomaly detection, which discriminates anomalous behaviour from normal behaviour in computer systems. One popular solution for anomaly detection is Artificial Immune Systems (AIS). AIS are adaptive systems inspired by theoretical immunology and observed immune functions, principles and models which are applied to problem solving. The Dendritic Cell Algorithm (DCA)[2] is an AIS algorithm that is developed specifically for anomaly detection. It has been successfully applied to intrusion detection in computer security. It is believed that agent-based modelling is an ideal approach for implementing AIS, as intelligent agents could be the perfect representations of immune entities in AIS. This model evaluates the feasibility of re-implementing the DCA in an agent-based simulation environment called AnyLogic, where the immune entities in the DCA are represented by intelligent agents. If this model can be successfully implemented, it makes it possible to implement more complicated and adaptive AIS models in the agent-based simulation environment.
Resumo:
Due to the growth of design size and complexity, design verification is an important aspect of the Logic Circuit development process. The purpose of verification is to validate that the design meets the system requirements and specification. This is done by either functional or formal verification. The most popular approach to functional verification is the use of simulation based techniques. Using models to replicate the behaviour of an actual system is called simulation. In this thesis, a software/data structure architecture without explicit locks is proposed to accelerate logic gate circuit simulation. We call thus system ZSIM. The ZSIM software architecture simulator targets low cost SIMD multi-core machines. Its performance is evaluated on the Intel Xeon Phi and 2 other machines (Intel Xeon and AMD Opteron). The aim of these experiments is to: • Verify that the data structure used allows SIMD acceleration, particularly on machines with gather instructions ( section 5.3.1). • Verify that, on sufficiently large circuits, substantial gains could be made from multicore parallelism ( section 5.3.2 ). • Show that a simulator using this approach out-performs an existing commercial simulator on a standard workstation ( section 5.3.3 ). • Show that the performance on a cheap Xeon Phi card is competitive with results reported elsewhere on much more expensive super-computers ( section 5.3.5 ). To evaluate the ZSIM, two types of test circuits were used: 1. Circuits from the IWLS benchmark suit [1] which allow direct comparison with other published studies of parallel simulators.2. Circuits generated by a parametrised circuit synthesizer. The synthesizer used an algorithm that has been shown to generate circuits that are statistically representative of real logic circuits. The synthesizer allowed testing of a range of very large circuits, larger than the ones for which it was possible to obtain open source files. The experimental results show that with SIMD acceleration and multicore, ZSIM gained a peak parallelisation factor of 300 on Intel Xeon Phi and 11 on Intel Xeon. With only SIMD enabled, ZSIM achieved a maximum parallelistion gain of 10 on Intel Xeon Phi and 4 on Intel Xeon. Furthermore, it was shown that this software architecture simulator running on a SIMD machine is much faster than, and can handle much bigger circuits than a widely used commercial simulator (Xilinx) running on a workstation. The performance achieved by ZSIM was also compared with similar pre-existing work on logic simulation targeting GPUs and supercomputers. It was shown that ZSIM simulator running on a Xeon Phi machine gives comparable simulation performance to the IBM Blue Gene supercomputer at very much lower cost. The experimental results have shown that the Xeon Phi is competitive with simulation on GPUs and allows the handling of much larger circuits than have been reported for GPU simulation. When targeting Xeon Phi architecture, the automatic cache management of the Xeon Phi, handles and manages the on-chip local store without any explicit mention of the local store being made in the architecture of the simulator itself. However, targeting GPUs, explicit cache management in program increases the complexity of the software architecture. Furthermore, one of the strongest points of the ZSIM simulator is its portability. Note that the same code was tested on both AMD and Xeon Phi machines. The same architecture that efficiently performs on Xeon Phi, was ported into a 64 core NUMA AMD Opteron. To conclude, the two main achievements are restated as following: The primary achievement of this work was proving that the ZSIM architecture was faster than previously published logic simulators on low cost platforms. The secondary achievement was the development of a synthetic testing suite that went beyond the scale range that was previously publicly available, based on prior work that showed the synthesis technique is valid.
Resumo:
Background: The emergence of multiple-drug resistance bacteria has become a major threat and thus calls for an urgent need to search for new effective and safe anti-bacterial agents. Objectives: This study aims to evaluate the anticancer and antibacterial activities of secondary metabolites from Penicillium sp. , an endophytic fungus associated with leaves of Garcinia nobilis . Methods: The culture filtrate from the fermentation of Penicillium sp. was extracted and analyzed by liquid chromatography– mass spectrometry, and the major metabolites were isolated and identified by spectroscopic analyses and by comparison with published data. The antibacterial activity of the compounds was assessed by broth microdilution method while the anticancer activity was determined by the 3-(4,5-dimethylthiazol-2-yl)-2,5-diphenyltetrazolium bromide assay. Results: The fractionation of the crude extract afforded penialidin A-C (1-3), citromycetin (4), p-hydroxyphenylglyoxalaldoxime (5) and brefelfin A (6). All of the compounds tested here showed antibacterial activity (MIC = 0.50 – 128 μg/mL) against Gramnegative multi-drug resistance bacteria, Vibrio cholerae (causative agent of dreadful disease cholera) and Shigella flexneri (causative agent of shigellosis), as well as the significant anticancer activity (LC50 = 0.88 – 9.21 μg/mL) against HeLa cells. Conclusion: The results obtained indicate that compounds 1-6 showed good antibacterial and anticancer activities with no toxicity to human red blood cells and normal Vero cells.
Resumo:
Automotive producers are adopting multi-modal fulfillment models in which customers can be fulfilled by products from stock, by allocating as yet unmade products that are in the planning pipeline, or by building a product to order. This study explores how fulfillment is sensitive to several parameters of the system and how they interact with different methods for sequencing products into the production plan.
Resumo:
This paper reports on continuing research into the modelling of an order picking process within a Crossdocking distribution centre using Simulation Optimisation. The aim of this project is to optimise a discrete event simulation model and to understand factors that affect finding its optimal performance. Our initial investigation revealed that the precision of the selected simulation output performance measure and the number of replications required for the evaluation of the optimisation objective function through simulation influences the ability of the optimisation technique. We experimented with Common Random Numbers, in order to improve the precision of our simulation output performance measure, and intended to use the number of replications utilised for this purpose as the initial number of replications for the optimisation of our Crossdocking distribution centre simulation model. Our results demonstrate that we can improve the precision of our selected simulation output performance measure value using Common Random Numbers at various levels of replications. Furthermore, after optimising our Crossdocking distribution centre simulation model, we are able to achieve optimal performance using fewer simulations runs for the simulation model which uses Common Random Numbers as compared to the simulation model which does not use Common Random Numbers.
Resumo:
Part 18: Optimization in Collaborative Networks
Resumo:
Part 13: Virtual Reality and Simulation