901 resultados para Manufacturing processes -- Computer simulation


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis reports a detailed investigation of the micromechanics of agglomerate behaviour under free-fall impact, double (punch) impact and diametrical compression tests using the simulation software TRUBAL. The software is based on the discrete element method (DEM) which incorporates the Newtonian equations of motion and contact mechanics theory to model the interparticle interactions. Four agglomerates have been used: three dense (differing in interface energy and contact density) and one loose. Although the simulated agglomerates are relatively coarse-grained, the results obtained are in good agreement with laboratory test results reported in the literature. The computer simulation results show that, in all three types of test, the loose agglomerate cannot fracture as it is unable to store sufficient elastic energy. Instead, it becomes flattened for low loading-rates and shattered or crushed at higher loading-rates. In impact tests, the dense agglomerates experience only local damage at low impact velocities. Semi-brittle fracture and fragmentation are produced over a range of higher impact velocities and at very high impact velocities shattering occurs. The dense agglomerates fracture in two or three large fragments in the diametrical compression tests. Local damage at the agglomerate-platen interface always occurs prior to fracture and consists of local bond breakage (microcrack formation) and local dislocations (compaction). The fracture process is dynamic and much more complex than that suggested by continuum fracture mechanics theory. Cracks are always initiated from the contact zones and propagate towards the agglomerate centre. Fracture occurs a short time after the start of unloading when a fracture crack "selection" process takes place. The detailed investigation of the agglomerate damage processes includes an examination of the evolution of the fracture surface. Detailed comparisons of the behaviour of the same agglomerate in all three types of test are presented. The particle size distribution curves of the debris are also examined, for both free-fall and double impact tests.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The computer simulation of manufacturing systems is commonly carried out using discrete event simulation (DES). Indeed, there appears to be a lack of applications of continuous simulation methods, particularly system dynamics (SD), despite evidence that this technique is suitable for industrial modelling. This paper investigates whether this is due to a decline in the general popularity of SD, or whether modelling of manufacturing systems represents a missed opportunity for SD. On this basis, the paper first gives a review of the concept of SD and fully describes the modelling technique. Following on, a survey of the published applications of SD in the 1990s is made by developing and using a structured classification approach. From this review, observations are made about the application of the SD method and opportunities for future research are suggested.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quality, production and technological innovation management rank among the most important matters of concern to modern manufacturing organisations. They can provide companies with the decisive means of gaining a competitive advantage, especially within industries where there is an increasing similarity in product design and manufacturing processes. The papers in this special issue of International Journal of Technology Management have all been selected as examples of how aspects of quality, production and technological innovation can help to improve competitive performance. Most are based on presentations made at the UK Operations Management Association's Sixth International Conference held at Aston University at which the theme was 'Getting Ahead Through Technology and People'. At the conference itself over 80 papers were presented by authors from 15 countries around the world. Among the many topics addressed within the conference theme, technological innovation, quality and production management emerged as attracting the greatest concern and interest of delegates, particularly those from industry. For any new initiative to be implemented successfully, it should be led from the top of the organization. Achieving the desired level of commitment from top management can, however, be a difficulty. In the first paper of this issue, Mackness investigates this question by explaining how systems thinking can help. In the systems approach, properties such as 'emergence', 'hierarchy', 'commnication' and 'control' are used to assist top managers in preparing for change. Mackness's paper is then complemented by Iijima and Hasegawa's contribution in which they investigate the development of Quality Information Management (QIM) in Japan. They present the idea of a Design Review and demonstrate how it can be used to trace and reduce quality-related losses. The next paper on the subject of quality is by Whittle and colleagues. It relates to total quality and the process of culture change within organisations. Using the findings of investigations carried out in a number of case study companies, they describe four generic models which have been identified as characterising methods of implementing total quality within existing organisation cultures. Boaden and Dale's paper also relates to the management of quality, but looks specifically at the construction industry where it has been found there is still some confusion over the role of Quality Assurance (QA) and Total Quality Management (TQM). They describe the results of a questionnaire survey of forty companies in the industry and compare them to similar work carried out in other industries. Szakonyi's contribution then completes this group of papers which all relate specifically to the question of quality. His concern is with the two ways in which R&D or engineering managers can work on improving quality. The first is by improving it in the laboratory, while the second is by working with other functions to improve quality in the company. The next group of papers in this issue all address aspects of production management. Umeda's paper proposes a new manufacturing-oriented simulation package for production management which provides important information for both design and operation of manufacturing systems. A simulation for production strategy in a Computer Integrated Manufacturing (CIM) environment is also discussed. This paper is then followed by a contribution by Tanaka and colleagues in which they consider loading schedules for manufacturing orders in a Material Requirements Planning (MRP) environment. They compare mathematical programming with a knowledge-based approach, and comment on their relative effectiveness for different practical situations. Engstrom and Medbo's paper then looks at a particular aspect of production system design, namely the question of devising group working arrangements for assembly with new product structures. Using the case of a Swedish vehicle assembly plant where long cycle assembly work has been adopted, they advocate the use of a generally applicable product structure which can be adapted to suit individual local conditions. In the last paper of this particular group, Tay considers how automation has affected the production efficiency in Singapore. Using data from ten major industries he identifies several factors which are positively correlated with efficiency, with capital intensity being of greatest interest to policy makers. The two following papers examine the case of electronic data interchange (EDI) as a means of improving the efficiency and quality of trading relationships. Banerjee and Banerjee consider a particular approach to material provisioning for production systems using orderless inventory replenishment. Using the example of a single supplier and multiple buyers they develop an analytical model which is applicable for the exchange of information between trading partners using EDI. They conclude that EDI-based inventory control can be attractive from economic as well as other standpoints and that the approach is consistent with and can be instrumental in moving towards just-in-time (JIT) inventory management. Slacker's complementary viewpoint on EDI is from the perspective of the quality relation-ship between the customer and supplier. Based on the experience of Lucas, a supplier within the automotive industry, he concludes that both banks and trading companies must take responsibility for the development of payment mechanisms which satisfy the requirements of quality trading. The three final papers of this issue relate to technological innovation and are all country based. Berman and Khalil report on a survey of US technological effectiveness in the global economy. The importance of education is supported in their conclusions, although it remains unclear to what extent the US government can play a wider role in promoting technological innovation and new industries. The role of technology in national development is taken up by Martinsons and Valdemars who examine the case of the former Soviet Union. The failure to successfully infuse technology into Soviet enterprises is seen as a factor in that country's demise, and it is anticipated that the newly liberalised economies will be able to encourage greater technological creativity. This point is then taken up in Perminov's concluding paper which looks in detail at Russia. Here a similar analysis is made of the concluding paper which looks in detail at Russia. Here a similar analysis is made of the Soviet Union's technological decline, but a development strategy is also presented within the context of the change from a centralised to a free market economy. The papers included in this special issue of the International Journal of Technology Management each represent a unique and particular contribution to their own specific area of concern. Together, however, they also argue or demonstrate the general improvements in competitive performance that can be achieved through the application of modern principles and practice to the management of quality, production and technological innovation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The last two decades have seen a proliferation of research frameworks that emphasise the importance of understanding adaptive processes that happen at different levels. We contribute to this growing body of literature by exploring how cultural (mal)adaptive dynamics relate to multilevel social-ecological processes occurring at different scales, where the lower levels combine into new units with new organizations, functions, and emergent properties or collective behaviors. After a brief review of the concept of “cultural adaptation” from the perspective of cultural evolutionary theory, the core of the paper is constructed around the exploration of multilevel processes occurring at the temporal, spatial, social, and political scales. We do so by using insights from cultural evolutionary theory and by examining small-scale societies as case studies. In each section, we discuss the importance of the selected scale for understanding cultural adaptation and then present an example that illustrates how multilevel processes in the selected scale help explain observed patterns in the cultural adaptive process. The last section of the paper discusses the potential of modeling and computer simulation for studying multilevel processes in cultural adaptation. We conclude by highlighting how elements from cultural evolutionary theory might enrich the multilevel process discussion in resilience theory.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

One of the core tasks of the virtual-manufacturing environment is to characterise the transformation of the state of material during each of the unit processes. This transformation in shape, material properties, etc. can only be reliably achieved through the use of models in a simulation context. Unfortunately, many manufacturing processes involve the material being treated in both the liquid and solid state, the trans-formation of which may be achieved by heat transfer and/or electro-magnetic fields. The computational modelling of such processes, involving the interactions amongst various interacting phenomena, is a consider-able challenge. However, it must be addressed effectively if Virtual Manufacturing Environments are to become a reality! This contribution focuses upon one attempt to develop such a multi-physics computational toolkit. The approach uses a single discretisation procedure and provides for direct interaction amongst the component phenomena. The need to exploit parallel high performance hardware is addressed so that simulation elapsed times can be brought within the realms of practicality. Examples of Multiphysics modelling in relation to shape casting, and solder joint formation reinforce the motivation for this work.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Computer simulation programs are essential tools for scientists and engineers to understand a particular system of interest. As expected, the complexity of the software increases with the depth of the model used. In addition to the exigent demands of software engineering, verification of simulation programs is especially challenging because the models represented are complex and ridden with unknowns that will be discovered by developers in an iterative process. To manage such complexity, advanced verification techniques for continually matching the intended model to the implemented model are necessary. Therefore, the main goal of this research work is to design a useful verification and validation framework that is able to identify model representation errors and is applicable to generic simulators. The framework that was developed and implemented consists of two parts. The first part is First-Order Logic Constraint Specification Language (FOLCSL) that enables users to specify the invariants of a model under consideration. From the first-order logic specification, the FOLCSL translator automatically synthesizes a verification program that reads the event trace generated by a simulator and signals whether all invariants are respected. The second part consists of mining the temporal flow of events using a newly developed representation called State Flow Temporal Analysis Graph (SFTAG). While the first part seeks an assurance of implementation correctness by checking that the model invariants hold, the second part derives an extended model of the implementation and hence enables a deeper understanding of what was implemented. The main application studied in this work is the validation of the timing behavior of micro-architecture simulators. The study includes SFTAGs generated for a wide set of benchmark programs and their analysis using several artificial intelligence algorithms. This work improves the computer architecture research and verification processes as shown by the case studies and experiments that have been conducted.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A two-dimensional numeric simulator is developed to predict the nonlinear, convective-reactive, oxygen mass exchange in a cross-flow hollow fiber blood oxygenator. The numeric simulator also calculates the carbon dioxide mass exchange, as hemoglobin affinity to oxygen is affected by the local pH value, which depends mostly on the local carbon dioxide content in blood. Blood pH calculation inside the oxygenator is made by the simultaneous solution of an equation that takes into account the blood buffering capacity and the classical Henderson-Hasselbach equation. The modeling of the mass transfer conductance in the blood comprises a global factor, which is a function of the Reynolds number, and a local factor, which takes into account the amount of oxygen reacted to hemoglobin. The simulator is calibrated against experimental data for an in-line fiber bundle. The results are: (i) the calibration process allows the precise determination of the mass transfer conductance for both oxygen and carbon dioxide; (ii) very alkaline pH values occur in the blood path at the gas inlet side of the fiber bundle; (iii) the parametric analysis of the effect of the blood base excess (BE) shows that V(CO2) is similar in the case of blood metabolic alkalosis, metabolic acidosis, or normal BE, for a similar blood inlet P(CO2), although the condition of metabolic alkalosis is the worst case, as the pH in the vicinity of the gas inlet is the most alkaline; (iv) the parametric analysis of the effect of the gas flow to blood flow ratio (Q(G)/Q(B)) shows that V(CO2) variation with the gas flow is almost linear up to Q(G)/Q(B) = 2.0. V(O2) is not affected by the gas flow as it was observed that by increasing the gas flow up to eight times, the V(O2) grows only 1%. The mass exchange of carbon dioxide uses the full length of the hollow-fiber only if Q(G)/Q(B) > 2.0, as it was observed that only in this condition does the local variation of pH and blood P(CO2) comprise the whole fiber bundle.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The purpose of this study was to explore the potential advantages, both theoretical and applied, of preserving low-frequency acoustic hearing in cochlear implant patients. Several hypotheses are presented that predict that residual low-frequency acoustic hearing along with electric stimulation for high frequencies will provide an advantage over traditional long-electrode cochlear implants for the recognition of speech in competing backgrounds. A simulation experiment in normal-hearing subjects demonstrated a clear advantage for preserving low-frequency residual acoustic hearing for speech recognition in a background of other talkers, but not in steady noise. Three subjects with an implanted "short-electrode" cochlear implant and preserved low-frequency acoustic hearing were also tested on speech recognition in the same competing backgrounds and compared to a larger group of traditional cochlear implant users. Each of the three short-electrode subjects performed better than any of the traditional long-electrode implant subjects for speech recognition in a background of other talkers, but not in steady noise, in general agreement with the simulation studies. When compared to a subgroup of traditional implant users matched according to speech recognition ability in quiet, the short-electrode patients showed a 9-dB advantage in the multitalker background. These experiments provide strong preliminary support for retaining residual low-frequency acoustic hearing in cochlear implant patients. The results are consistent with the idea that better perception of voice pitch, which can aid in separating voices in a background of other talkers, was responsible for this advantage.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantifying mass and energy exchanges within tropical forests is essential for understanding their role in the global carbon budget and how they will respond to perturbations in climate. This study reviews ecosystem process models designed to predict the growth and productivity of temperate and tropical forest ecosystems. Temperate forest models were included because of the minimal number of tropical forest models. The review provides a multiscale assessment enabling potential users to select a model suited to the scale and type of information they require in tropical forests. Process models are reviewed in relation to their input and output parameters, minimum spatial and temporal units of operation, maximum spatial extent and time period of application for each organization level of modelling. Organizational levels included leaf-tree, plot-stand, regional and ecosystem levels, with model complexity decreasing as the time-step and spatial extent of model operation increases. All ecosystem models are simplified versions of reality and are typically aspatial. Remotely sensed data sets and derived products may be used to initialize, drive and validate ecosystem process models. At the simplest level, remotely sensed data are used to delimit location, extent and changes over time of vegetation communities. At a more advanced level, remotely sensed data products have been used to estimate key structural and biophysical properties associated with ecosystem processes in tropical and temperate forests. Combining ecological models and image data enables the development of carbon accounting systems that will contribute to understanding greenhouse gas budgets at biome and global scales.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents results on the simulation of the solid state sintering of copper wires using Monte Carlo techniques based on elements of lattice theory and cellular automata. The initial structure is superimposed onto a triangular, two-dimensional lattice, where each lattice site corresponds to either an atom or vacancy. The number of vacancies varies with the simulation temperature, while a cluster of vacancies is a pore. To simulate sintering, lattice sites are picked at random and reoriented in terms of an atomistic model governing mass transport. The probability that an atom has sufficient energy to jump to a vacant lattice site is related to the jump frequency, and hence the diffusion coefficient, while the probability that an atomic jump will be accepted is related to the change in energy of the system as a result of the jump, as determined by the change in the number of nearest neighbours. The jump frequency is also used to relate model time, measured in Monte Carlo Steps, to the actual sintering time. The model incorporates bulk, grain boundary and surface diffusion terms and includes vacancy annihilation on the grain boundaries. The predictions of the model were found to be consistent with experimental data, both in terms of the microstructural evolution and in terms of the sintering time. (C) 2002 Elsevier Science B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Mecânica - Ramo Manutenção e Produção

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pultruded products are being targeted by a growing demand due to its excellent mechanical properties and low chemical reactivity, ensuring a low level of maintenance operations and allowing an easier assembly operation process than equivalent steel bars. In order to improve the mechanical drawing process and solve some acoustic and thermal insulation problems, pultruded pipes of glass fibre reinforced plastics (GFRF) can be filled with special products that increase their performance regarding the issues previously referred. The great challenge of this work was drawing a new equipment able to produce pultruded pipes filled with cork or polymeric pre-shaped bars as a continuous process. The project was carried out successfully and the new equipment was built and integrated in the pultrusion equipment already existing, allowing to obtain news products with higher added-value in the market, covering some needs previously identified in the field of civil construction.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The global warming due to high CO2 emission in the last years has made energy saving a global problem nowadays. However, manufacturing processes such as pultrusion necessarily needs heat for curing the resin. Then, the only option available is to apply all efforts to make the process even more efficient. Different heating systems have been used on pultrusion, however, the most widely used are the planar resistances. The main objective of this study is to develop another heating system and compares it with the former one. Thermography was used in spite of define the temperature profile along the die. FEA (finite element analysis) allows to understand how many energy is spend with the initial heating system. After this first approach, changes were done on the die in order to test the new heating system and to check possible quality problems on the product. Thus, this work allows to conclude that with the new heating system a significant reduction in the setup time is now possible and an energy reduction of about 57% was achieved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Com um mercado automóvel cada vez mais competitivo e com os construtores automóveis à procura de atingir os zero defeitos nos seus produtos, a Bosch Car Multimédia Portugal S.A, fabricante de sistemas multimédia para o mercado automóvel, tem como objetivo a qualidade perfeita dos seus produtos. Tal perfeição exige processos de fabrico cada vez mais evoluídos e com melhores sistemas de auxílio à montagem. Nesse sentido, a incorporação de sistemas de visão artificial para verificação da montagem correta dos componentes em sistemas multimédia tem vindo a crescer largamente. Os sistemas de inspeção visual da Cognex tornaram-se o standard da Bosch para a verifi-cação da montagem de componentes por serem sistemas bastante completos, fáceis de con-figurar e com um suporte técnico bastante completo. Estes sistemas têm vindo a ser inte-grados em diversas máquinas (postos) de montagem e nunca foi desenvolvida uma ferra-menta normalizada para integração destes sistemas com as máquinas. A ideia principal deste projeto passou por desenvolver um sistema (uma aplicação informá-tica) que permita controlar os indicadores de qualidade destes sistemas de visão, garantir o seguimento dos produtos montados e, ao mesmo tempo, efetuar cópias de segurança de todo o sistema para utilização em caso de avaria ou de troca de equipamento. Tal sistema foi desenvolvido recorrendo à programação de uma Dynamic Link Library (DLL), através da linguagem VisualBasic.NET, que permite às aplicações dos equipamen-tos (máquinas) da Bosch Car Multimédia comunicarem de uma forma universal e transpa-rente com os sistemas de inspeção visual da marca Cognex. Os objetivos a que o autor se propôs no desenvolvimento deste sistema foram na sua maioria alcançados e o projeto encontra-se atualmente implementado e em execução nas linhas de produção da Bosch Car Multimédia.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação de mestrado integrado em Engenharia Eletrónica Industrial e de Computadores