88 resultados para Fault Simulation
Resumo:
More than ever, the economic globalization is creating the need to increase business competitiveness. Lean manufacturing is a management philosophy oriented to the elimination of activities that do not create any type of value and are thus considered a waste. One of the main differences from other management philosophies is the shop-floor focus and the operators' involvement. Therefore, the training of all organization levels is crucial for the success of lean manufacturing. Universities should also participate actively in this process by developing students' lean management skills and promoting a better and faster integration of students into their future organizations. This paper proposes a single realistic manufacturing platform, involving production and assembly operations, to learn by playing many of the lean tools such as VSM, 5S, SMED, poke-yoke, line balance, TPM, Mizusumashi, plant layout, and JIT/kanban. This simulation game was built in tight cooperation with experienced lean companies under the international program “Lean Learning Academy,”http://www.leanlearningacademy.eu/ and its main aim is to make bachelor and master courses in applied sciences more attractive by integrating classic lectures with a simulated production environment that could result in more motivated students and higher study yields. The simulation game results show that our approach is efficient in providing a realistic platform for the effective learning of lean principles, tools, and mindset, which can be easily included in course classes of less than two hours.
Resumo:
In an increasingly competitive and globalized world, companies need effective training methodologies and tools for their employees. However, selecting the most suitable ones is not an easy task. It depends on the requirements of the target group (namely time restrictions), on the specificities of the contents, etc. This is typically the case for training in Lean, the waste elimination manufacturing philosophy. This paper presents and compares two different approaches to lean training methodologies and tools: a simulation game based on a single realistic manufacturing platform, involving production and assembly operations that allows learning by playing; and a digital game that helps understand lean tools. This paper shows that both tools have advantages in terms of trainee motivation and knowledge acquisition. Furthermore, they can be used in a complementary way, reinforcing the acquired knowledge.
Resumo:
Dependability is a critical factor in computer systems, requiring high quality validation & verification procedures in the development stage. At the same time, digital devices are getting smaller and access to their internal signals and registers is increasingly complex, requiring innovative debugging methodologies. To address this issue, most recent microprocessors include an on-chip debug (OCD) infrastructure to facilitate common debugging operations. This paper proposes an enhanced OCD infrastructure with the objective of supporting the verification of fault-tolerant mechanisms through fault injection campaigns. This upgraded on-chip debug and fault injection (OCD-FI) infrastructure provides an efficient fault injection mechanism with improved capabilities and dynamic behavior. Preliminary results show that this solution provides flexibility in terms of fault triggering and allows high speed real-time fault injection in memory elements
Resumo:
Fault injection is frequently used for the verification and validation of dependable systems. When targeting real time microprocessor based systems the process becomes significantly more complex. This paper proposes two complementary solutions to improve real time fault injection campaign execution, both in terms of performance and capabilities. The methodology is based on the use of the on-chip debug mechanisms present in modern electronic devices. The main objective is the injection of faults in microprocessor memory elements with minimum delay and intrusiveness. Different configurations were implemented and compared in terms of performance gain and logic overhead.
Resumo:
As electronic devices get smaller and more complex, dependability assurance is becoming fundamental for many mission critical computer based systems. This paper presents a case study on the possibility of using the on-chip debug infrastructures present in most current microprocessors to execute real time fault injection campaigns. The proposed methodology is based on a debugger customized for fault injection and designed for maximum flexibility, and consists of injecting bit-flip type faults on memory elements without modifying or halting the target application. The debugger design is easily portable and applicable to different architectures, providing a flexible and efficient mechanism for verifying and validating fault tolerant components.
Resumo:
The rapid increase in the use of microprocessor-based systems in critical areas, where failures imply risks to human lives, to the environment or to expensive equipment, significantly increased the need for dependable systems, able to detect, tolerate and eventually correct faults. The verification and validation of such systems is frequently performed via fault injection, using various forms and techniques. However, as electronic devices get smaller and more complex, controllability and observability issues, and sometimes real time constraints, make it harder to apply most conventional fault injection techniques. This paper proposes a fault injection environment and a scalable methodology to assist the execution of real-time fault injection campaigns, providing enhanced performance and capabilities. Our proposed solutions are based on the use of common and customized on-chip debug (OCD) mechanisms, present in many modern electronic devices, with the main objective of enabling the insertion of faults in microprocessor memory elements with minimum delay and intrusiveness. Different configurations were implemented starting from basic Components Off-The-Shelf (COTS) microprocessors, equipped with real-time OCD infrastructures, to improved solutions based on modified interfaces, and dedicated OCD circuitry that enhance fault injection capabilities and performance. All methodologies and configurations were evaluated and compared concerning performance gain and silicon overhead.
Resumo:
This paper presents the new package entitled Simulator of Intelligent Transportation Systems (SITS) and a computational oriented analysis of traffic dynamics. The SITS adopts a microscopic simulation approach to reproduce real traffic conditions considering different types of vehicles, drivers and roads. A set of experiments with the SITS reveal the dynamic phenomena exhibited by this kind of system. For this purpose a modelling formalism is developed that embeds the statistics and the Laplace transform. The results make possible the adoption of classical system theory tools and point out that it is possible to study traffic systems taking advantage of the knowledge gathered with automatic control algorithms. A complementary perspective for the analysis of the traffic flow is also quantified through the entropy measure.
Resumo:
To increase the amount of logic available to the users in SRAM-based FPGAs, manufacturers are using nanometric technologies to boost logic density and reduce costs, making its use more attractive. However, these technological improvements also make FPGAs particularly vulnerable to configuration memory bit-flips caused by power fluctuations, strong electromagnetic fields and radiation. This issue is particularly sensitive because of the increasing amount of configuration memory cells needed to define their functionality. A short survey of the most recent publications is presented to support the options assumed during the definition of a framework for implementing circuits immune to bit-flips induction mechanisms in memory cells, based on a customized redundant infrastructure and on a detection-and-fix controller.
Resumo:
Fault injection is frequently used for the verification and validation of the fault tolerant features of microprocessors. This paper proposes the modification of a common on-chip debugging (OCD) infrastructure to add fault injection capabilities and improve performance. The proposed solution imposes a very low logic overhead and provides a flexible and efficient mechanism for the execution of fault injection campaigns, being applicable to different target system architectures.
Resumo:
The main purpose of this work was the development of procedures for the simulation of atmospheric ows over complex terrain, using OpenFOAM. For this aim, tools and procedures were developed apart from this code for the preprocessing and data extraction, which were thereafter applied in the simulation of a real case. For the generation of the computational domain, a systematic method able to translate the terrain elevation model to a native OpenFOAM format (blockMeshDict) was developed. The outcome was a structured mesh, in which the user has the ability to de ne the number of control volumes and its dimensions. With this procedure, the di culties of case set up and the high computation computational e ort reported in literature associated to the use of snappyHexMesh, the OpenFOAM resource explored until then for the accomplishment of this task, were considered to be overwhelmed. Developed procedures for the generation of boundary conditions allowed for the automatic creation of idealized inlet vertical pro les, de nition of wall functions boundary conditions and the calculation of internal eld rst guesses for the iterative solution process, having as input experimental data supplied by the user. The applicability of the generated boundary conditions was limited to the simulation of turbulent, steady-state, incompressible and neutrally strati ed atmospheric ows, always recurring to RaNS (Reynolds-averaged Navier-Stokes) models. For the modelling of terrain roughness, the developed procedure allowed to the user the de nition of idealized conditions, like an uniform aerodynamic roughness length or making its value variable as a function of topography characteristic values, or the using of real site data, and it was complemented by the development of techniques for the visual inspection of generated roughness maps. The absence and the non inclusion of a forest canopy model limited the applicability of this procedure to low aerodynamic roughness lengths. The developed tools and procedures were then applied in the simulation of a neutrally strati ed atmospheric ow over the Askervein hill. In the performed simulations was evaluated the solution sensibility to di erent convection schemes, mesh dimensions, ground roughness and formulations of the k - ε and k - ω models. When compared to experimental data, calculated values showed a good agreement of speed-up in hill top and lee side, with a relative error of less than 10% at a height of 10 m above ground level. Turbulent kinetic energy was considered to be well simulated in the hill windward and hill top, and grossly predicted in the lee side, where a zone of ow separation was also identi ed. Despite the need of more work to evaluate the importance of the downstream recirculation zone in the quality of gathered results, the agreement between the calculated and experimental values and the OpenFOAM sensibility to the tested parameters were considered to be generally in line with the simulations presented in the reviewed bibliographic sources.
Resumo:
Pultrusion is an industrial process used to produce glass fibers reinforced polymers profiles. These materials are worldwide used when performing characteristics, such as great electrical and magnetic insulation, high strength to weight ratio, corrosion and weather resistance, long service life and minimal maintenance are required. In this study, we present the results of the modelling and simulation of heat flow through a pultrusion die by means of Finite Element Analysis (FEA). The numerical simulation was calibrated based on temperature profiles computed from thermographic measurements carried out during pultrusion manufacturing process. Obtained results have shown a maximum deviation of 7%, which is considered to be acceptable for this type of analysis, and is below to the 10% value, previously specified as maximum deviation. © 2011, Advanced Engineering Solutions.
Resumo:
Everyday accounting and management teachers face the challenge of creating learning environments that motivate students. This chapter describes the Business Simulation (BS) experience that has taken place at the Polytechnic Institute of Porto, Institute of Accounting and Administration (IPP/ISCAP). The chapter presents students’ perceptions about the course and the teaching/learning approach. The results show that pedagogical methods used (competency-oriented), generic competencies (cooperation and group work), and interpersonal skills (organisational and communication skills) are relevant for future accounting professionals. In addition, positive remarks and possible constraints based on observation, staff meetings, and past research are reported. The chapter concludes with some recommendations from the project implementation.
Resumo:
Volatile organic compounds are a common source of groundwater contamination that can be easily removed by air stripping in columns with random packing and using a counter-current flow between the phases. This work proposes a new methodology for the column design for any particular type of packing and contaminant avoiding the necessity of a pre-defined diameter used in the classical approach. It also renders unnecessary the employment of the graphical Eckert generalized correlation for pressure drop estimates. The hydraulic features are previously chosen as a project criterion and only afterwards the mass transfer phenomena are incorporated, in opposition to conventional approach. The design procedure was translated into a convenient algorithm using C++ as programming language. A column was built in order to test the models used either in the design or in the simulation of the column performance. The experiments were fulfilled using a solution of chloroform in distilled water. Another model was built to simulate the operational performance of the column, both in steady state and in transient conditions. It consists in a system of two partial non linear differential equations (distributed parameters). Nevertheless, when flows are steady, the system became linear, although there is not an evident solution in analytical terms. In steady state the resulting system of ODE can be solved, allowing for the calculation of the concentration profile in both phases inside the column. In transient state the system of PDE was numerically solved by finite differences, after a previous linearization.
Resumo:
A distributed, agent-based intelligent system models and simulates a smart grid using physical players and computationally simulated agents. The proposed system can assess the impact of demand response programs.