893 resultados para 090407 Process Control and Simulation
Resumo:
In the drilling processes and especially deep-hole drilling process, the monitoring system and having control on mechanical parameters (e.g. Force, Torque,Vibration and Acoustic emission) are essential. The main focus of this thesis work is to study the characteristics of deep-hole drilling process, and optimize the monitoring system for controlling the process. The vibration is considered as a major defect area of the deep-hole drilling process which often leads to breakage of the drill, therefore by vibration analysis and optimizing the workpiecefixture, this area is studied by finite element method and the suggestions are explained. By study on a present monitoring system, and searching on the new sensor products, the modifications and recommendations are suggested for optimize the present monitoring system for excellent performance in deep-hole drilling process research and measurements.
Resumo:
Process development will be largely driven by the main equipment suppliers. The reason for this development is their ambition to supply complete plants or process systems instead of single pieces of equipment. The pulp and paper companies' interest lies in product development, as their main goal is to create winning brands and effective brand management. Design engineering companies will find their niche in detail engineering based on approved process solutions. Their development work will focus on increasing the efficiency of engineering work. Process design is a content-producing profession, which requires certain special characteristics: creativity, carefulness, the ability to work as a member of a design team according to time schedules and fluency in oral as well as written presentation. In the future, process engineers will increasingly need knowledge of chemistry as well as information and automation technology. Process engineering tools are developing rapidly. At the moment, these tools are good enough for static sizing and balancing, but dynamic simulation tools are not yet good enough for the complicated chemical reactions of pulp and paper chemistry. Dynamic simulation and virtual mill models are used as tools for training the operators. Computational fluid dynamics will certainlygain ground in process design.
Resumo:
The present dissertation is devoted to the systematic approach to the development of organic toxic and refractory pollutants abatement by chemical decomposition methods in aqueous and gaseous phases. The systematic approach outlines the basic scenario of chemical decomposition process applications with a step-by-step approximation to the most effective result with a predictable outcome for the full-scale application, confirmed by successful experience. The strategy includes the following steps: chemistry studies, reaction kinetic studies in interaction with the mass transfer processes under conditions of different control parameters, contact equipment design and studies, mathematical description of the process for its modelling and simulation, processes integration into treatment technology and its optimisation, and the treatment plant design. The main idea of the systematic approach for oxidation process introduction consists of a search for the most effective combination between the chemical reaction and the treatment device, in which the reaction is supposed to take place. Under this strategy,a knowledge of the reaction pathways, its products, stoichiometry and kinetics is fundamental and, unfortunately, often unavailable from the preliminary knowledge. Therefore, research made in chemistry on novel treatment methods, comprisesnowadays a substantial part of the efforts. Chemical decomposition methods in the aqueous phase include oxidation by ozonation, ozone-associated methods (O3/H2O2, O3/UV, O3/TiO2), Fenton reagent (H2O2/Fe2+/3+) and photocatalytic oxidation (PCO). In the gaseous phase, PCO and catalytic hydrolysis over zero valent ironsare developed. The experimental studies within the described methodology involve aqueous phase oxidation of natural organic matter (NOM) of potable water, phenolic and aromatic amino compounds, ethylene glycol and its derivatives as de-icing agents, and oxygenated motor fuel additives ¿ methyl tert-butyl ether (MTBE) ¿ in leachates and polluted groundwater. Gas-phase chemical decomposition includes PCO of volatile organic compounds and dechlorination of chlorinated methane derivatives. The results of the research summarised here are presented in fifteenattachments (publications and papers submitted for publication and under preparation).
Resumo:
The aim of the study is to developa novel robust controller based on sliding mode control technique for the hydraulic servo system with flexible load and for a flexible manipulator with the lift and jib hydraulic actuators. For the purpose of general control design, a dynamic model is derived describing the principle physical behavior for both the hydraulic servo system and the flexible hydraulic manipulator. The mechanism of hydraulic servo systems is described by basic mathematical equations of fluid powersystems and the dynamics of flexible manipulator is modeled by the assumed modemethod. The controller is constructed so as to track desired trajectories in the presence of model imprecision. Experimental and simulation results demonstratethat sliding mode control has benefits which can be used to guarantee stabilityin uncertain systems and improve the system performance and load tolerance.
Resumo:
The topic of this thesis is the simulation of a combination of several control and data assimilation methods, meant to be used for controlling the quality of paper in a paper machine. Paper making is a very complex process and the information obtained from the web is sparse. A paper web scanner can only measure a zig zag path on the web. An assimilation method is needed to process estimates for Machine Direction (MD) and Cross Direction (CD) profiles of the web. Quality control is based on these measurements. There is an increasing need for intelligent methods to assist in data assimilation. The target of this thesis is to study how such intelligent assimilation methods are affecting paper web quality. This work is based on a paper web simulator, which has been developed in the TEKES funded MASI NoTes project. The simulator is a valuable tool in comparing different assimilation methods. The thesis contains the comparison of four different assimilation methods. These data assimilation methods are a first order Bayesian model estimator, an ARMA model based on a higher order Bayesian estimator, a Fourier transform based Kalman filter estimator and a simple block estimator. The last one can be considered to be close to current operational methods. From these methods Bayesian, ARMA and Kalman all seem to have advantages over the commercial one. The Kalman and ARMA estimators seems to be best in overall performance.
Resumo:
A coupled system simulator, based on analytical circuit equations and a finite element method (FEM) model of the motor has been developed and it is used to analyse a frequency-converterfed industrial squirrel-cage induction motor. Two control systems that emulate the behaviour of commercial direct-torque-controlled (DTC) and vector-controlled industrial frequency converters have been studied, implemented in the simulation software and verified by extensive laboratory tests. Numerous factors that affect the operation of a variable speed drive (VSD) and its energy efficiency have been investigated, and their significance in the simulation of the VSD results has been studied. The dependency of the frequency converter, induction motor and system losses on the switching frequency is investigated by simulations and measurements at different speeds for both the vector control and the DTC. Intensive laboratory measurements have been carried out to verify the simulation results.
Resumo:
The control of coating layer properties is becoming increasingly important as a result of an emerging demand for novel coated paper-based products and an increasing popularity of new coating application methods. The governing mechanisms of microstructure formation dynamics during consolidation and drying are nevertheless, still poorly understood. Some of the difficulties encountered by experimental methods can be overcome by the utilisation of numerical modelling and simulation-based studies of the consolidation process. The objective of this study was to improve the fundamental understanding of pigment coating consolidation and structure formation mechanisms taking place on the microscopic level. Furthermore, it is aimed to relate the impact of process and suspension properties to the microstructure of the coating layer. A mathematical model based on a modified Stokesian dynamics particle simulation technique was developed and applied in several studies of consolidation-related phenomena. The model includes particle-particle and particle-boundary hydrodynamics, colloidal interactions, Born repulsion, and a steric repulsion model. The Brownian motion and a free surface model were incorporated to enable the specific investigation of consolidation and drying. Filter cake stability was simulated in various particle systems, and subjected to a range of base substrate absorption rates and system temperatures. The stability of the filter cake was primarily affected by the absorption rate and size of particles. Temperature was also shown to have an influence. The consolidation of polydisperse systems, with varying wet coating thicknesses, was studied using imposed pilot trial and model-based drying conditions. The results show that drying methods have a clear influence on the microstructure development, on small particle distributions in the coating layer and also on the mobility of particles during consolidation. It is concluded that colloidal properties can significantly impact coating layer shrinkage as well as the internal solids concentration profile. Visualisations of particle system development in time and comparison of systems at different conditions are useful in illustrating coating layer structure formation mechanisms. The results aid in understanding the underlying mechanisms of pigment coating layer consolidation. Guidance is given regarding the relationship between coating process conditions and internal coating slurry properties and their effects on the microstructure of the coating.
Resumo:
Delays in the justice system have been undermining the functioning and performance of the court system all over the world for decades. Despite the widespread concern about delays, the solutions have not kept up with the growth of the problem. The delay problem existing in the justice courts processes is a good example of the growing need and pressure in professional public organizations to start improving their business process performance.This study analyses the possibilities and challenges of process improvement in professional public organizations. The study is based on experiences gained in two longitudinal action research improvement projects conducted in two separate Finnish law instances; in the Helsinki Court of Appeal and in the Insurance Court. The thesis has two objectives. First objective is to study what kinds of factors in court system operations cause delays and unmanageable backlogs and how to reduce and prevent delays. Based on the lessons learned from the case projects the objective is to give new insights on the critical factors of process improvement conducted in professional public organizations. Four main areas and factors behind the delay problem is identified: 1) goal setting and performance measurement practices, 2) the process control system, 3) production and capacity planning procedures, and 4) process roles and responsibilities. The appropriate improvement solutions include tools to enhance project planning and scheduling and monitoring the agreed time-frames for different phases of the handling process and pending inventory. The study introduces the identified critical factors in different phases of process improvement work carried out in professional public organizations, the ways the critical factors can be incorporated to the different stages of the projects, and discusses the role of external facilitator in assisting process improvement work and in enhancing ownership towards the solutions and improvement. The study highlights the need to concentrate on the critical factors aiming to get the employees to challenge their existing ways of conducting work, analyze their own processes, and create procedures for diffusing the process improvement culture instead of merely concentrating of finding tools, techniques, and solutions appropriate for applications from the manufacturing sector
Resumo:
The purpose of this study was to simulate and to optimize integrated gasification for combine cycle (IGCC) for power generation and hydrogen (H2) production by using low grade Thar lignite coal and cotton stalk. Lignite coal is abundant of moisture and ash content, the idea of addition of cotton stalk is to increase the mass of combustible material per mass of feed use for the process, to reduce the consumption of coal and to increase the cotton stalk efficiently for IGCC process. Aspen plus software is used to simulate the process with different mass ratios of coal to cotton stalk and for optimization: process efficiencies, net power generation and H2 production etc. are considered while environmental hazard emissions are optimized to acceptance level. With the addition of cotton stalk in feed, process efficiencies started to decline along with the net power production. But for H2 production, it gave positive result at start but after 40% cotton stalk addition, H2 production also started to decline. It also affects negatively on environmental hazard emissions and mass of emissions/ net power production increases linearly with the addition of cotton stalk in feed mixture. In summation with the addition of cotton stalk, overall affects seemed to negative. But the effect is more negative after 40% cotton stalk addition so it is concluded that to get maximum process efficiencies and high production less amount of cotton stalk addition in feed is preferable and the maximum level of addition is estimated to 40%. Gasification temperature should keep lower around 1140 °C and prefer technique for studied feed in IGCC is fluidized bed (ash in dry form) rather than ash slagging gasifier
Resumo:
Industrial applications demand that robots operate in agreement with the position and orientation of their end effector. It is necessary to solve the kinematics inverse problem. This allows the displacement of the joints of the manipulator to be determined, to accomplish a given objective. Complete studies of dynamical control of joint robotics are also necessary. Initially, this article focuses on the implementation of numerical algorithms for the solution of the kinematics inverse problem and the modeling and simulation of dynamic systems. This is done using real time implementation. The modeling and simulation of dynamic systems are performed emphasizing off-line programming. In sequence, a complete study of the control strategies is carried out through the study of several elements of a robotic joint, such as: DC motor, inertia, and gearbox. Finally a trajectory generator, used as input for a generic group of joints, is developed and a proposal of the controller's implementation of joints, using EPLD development system, is presented.
Resumo:
The assembly and maintenance of the International Thermonuclear Experimental Reactor (ITER) vacuum vessel (VV) is highly challenging since the tasks performed by the robot involve welding, material handling, and machine cutting from inside the VV. The VV is made of stainless steel, which has poor machinability and tends to work harden very rapidly, and all the machining operations need to be carried out from inside of the ITER VV. A general industrial robot cannot be used due to its poor stiffness in the heavy duty machining process, and this will cause many problems, such as poor surface quality, tool damage, low accuracy. Therefore, one of the most suitable options should be a light weight mobile robot which is able to move around inside of the VV and perform different machining tasks by replacing different cutting tools. Reducing the mass of the robot manipulators offers many advantages: reduced material costs, reduced power consumption, the possibility of using smaller actuators, and a higher payload-to-robot weight ratio. Offsetting these advantages, the lighter weight robot is more flexible, which makes it more difficult to control. To achieve good machining surface quality, the tracking of the end effector must be accurate, and an accurate model for a more flexible robot must be constructed. This thesis studies the dynamics and control of a 10 degree-of-freedom (DOF) redundant hybrid robot (4-DOF serial mechanism and 6-DOF 6-UPS hexapod parallel mechanisms) hydraulically driven with flexible rods under the influence of machining forces. Firstly, the flexibility of the bodies is described using the floating frame of reference method (FFRF). A finite element model (FEM) provided the Craig-Bampton (CB) modes needed for the FFRF. A dynamic model of the system of six closed loop mechanisms was assembled using the constrained Lagrange equations and the Lagrange multiplier method. Subsequently, the reaction forces between the parallel and serial parts were used to study the dynamics of the serial robot. A PID control based on position predictions was implemented independently to control the hydraulic cylinders of the robot. Secondly, in machining, to achieve greater end effector trajectory tracking accuracy for surface quality, a robust control of the actuators for the flexible link has to be deduced. This thesis investigates the intelligent control of a hydraulically driven parallel robot part based on the dynamic model and two schemes of intelligent control for a hydraulically driven parallel mechanism based on the dynamic model: (1) a fuzzy-PID self-tuning controller composed of the conventional PID control and with fuzzy logic, and (2) adaptive neuro-fuzzy inference system-PID (ANFIS-PID) self-tuning of the gains of the PID controller, which are implemented independently to control each hydraulic cylinder of the parallel mechanism based on rod length predictions. The serial component of the hybrid robot can be analyzed using the equilibrium of reaction forces at the universal joint connections of the hexa-element. To achieve precise positional control of the end effector for maximum precision machining, the hydraulic cylinder should be controlled to hold the hexa-element. Thirdly, a finite element approach of multibody systems using the Special Euclidean group SE(3) framework is presented for a parallel mechanism with flexible piston rods under the influence of machining forces. The flexibility of the bodies is described using the nonlinear interpolation method with an exponential map. The equations of motion take the form of a differential algebraic equation on a Lie group, which is solved using a Lie group time integration scheme. The method relies on the local description of motions, so that it provides a singularity-free formulation, and no parameterization of the nodal variables needs to be introduced. The flexible slider constraint is formulated using a Lie group and used for modeling a flexible rod sliding inside a cylinder. The dynamic model of the system of six closed loop mechanisms was assembled using Hamilton’s principle and the Lagrange multiplier method. A linearized hydraulic control system based on rod length predictions was implemented independently to control the hydraulic cylinders. Consequently, the results of the simulations demonstrating the behavior of the robot machine are presented for each case study. In conclusion, this thesis studies the dynamic analysis of a special hybrid (serialparallel) robot for the above-mentioned special task involving the ITER and investigates different control algorithms that can significantly improve machining performance. These analyses and results provide valuable insight into the design and control of the parallel robot with flexible rods.
Resumo:
Whereas the role of the anterior cingulate cortex (ACC) in cognitive control has received considerable attention, much less work has been done on the role of the ACC in autonomic regulation. Its connections through the vagus nerve to the sinoatrial node of the heart are thought to exert modulatory control over cardiovascular arousal. Therefore, ACC is not only responsible for the implementation of cognitive control, but also for the dynamic regulation of cardiovascular activity that characterizes healthy heart rate and adaptive behaviour. However, cognitive control and autonomic regulation are rarely examined together. Moreover, those studies that have examined the role of phasic vagal cardiac control in conjunction with cognitive performance have produced mixed results, finding relations for specific age groups and types of tasks but not consistently. So, while autonomic regulatory control appears to support effective cognitive performance under some conditions, it is not presently clear just what factors contribute to these relations. The goal of the present study was, therefore, to examine the relations between autonomic arousal, neural responsivity, and cognitive performance in the context of a task that required ACC support. Participants completed a primary inhibitory control task with a working memory load embedded. Pre-test cardiovascular measures were obtained, and ontask ERPs associated with response control (N2/P3) and error-related processes (ERN/Pe) were analyzed. Results indicated that response inhibition was unrelated to phasic vagal cardiac control, as indexed by respiratory sinus arrhythmia (RSA). However, higher resting RSA was associated with larger ERN ampUtude for the highest working memory load condition. This finding suggests that those individuals with greater autonomic regulatory control exhibited more robust ACC error-related responses on the most challenging task condition. On the other hand, exploratory analyses with rate pressure product (RPP), a measure of sympathetic arousal, indicated that higher pre-test RPP (i.e., more sympathetic influence) was associated with more errors on "catch" NoGo trials, i.e., NoGo trials that simultaneously followed other NoGo trials, and consequently, reqviired enhanced response control. Higher pre-test RPP was also associated with smaller amplitude ERNs for all three working memory loads and smaller ampUtude P3s for the low and medium working memory load conditions. Thus, higher pretest sympathetic arousal was associated with poorer performance on more demanding "catch" NoGo trials and less robust ACC-related electrocortical responses. The findings firom the present study highlight tiie interdependence of electrocortical and cardiovascular processes. While higher pre-test parasympathetic control seemed to relate to more robust ACC error-related responses, higher pre-test sympathetic arousal resulted in poorer inhibitory control performance and smaller ACC-generated electrocortical responses. Furthermore, these results provide a base from which to explore the relation between ACC and neuro/cardiac responses in older adults who may display greater variance due to the vulnerabihty of these systems to the normal aging process.
Resumo:
A wide range of tests for heteroskedasticity have been proposed in the econometric and statistics literature. Although a few exact homoskedasticity tests are available, the commonly employed procedures are quite generally based on asymptotic approximations which may not provide good size control in finite samples. There has been a number of recent studies that seek to improve the reliability of common heteroskedasticity tests using Edgeworth, Bartlett, jackknife and bootstrap methods. Yet the latter remain approximate. In this paper, we describe a solution to the problem of controlling the size of homoskedasticity tests in linear regression contexts. We study procedures based on the standard test statistics [e.g., the Goldfeld-Quandt, Glejser, Bartlett, Cochran, Hartley, Breusch-Pagan-Godfrey, White and Szroeter criteria] as well as tests for autoregressive conditional heteroskedasticity (ARCH-type models). We also suggest several extensions of the existing procedures (sup-type of combined test statistics) to allow for unknown breakpoints in the error variance. We exploit the technique of Monte Carlo tests to obtain provably exact p-values, for both the standard and the new tests suggested. We show that the MC test procedure conveniently solves the intractable null distribution problem, in particular those raised by the sup-type and combined test statistics as well as (when relevant) unidentified nuisance parameter problems under the null hypothesis. The method proposed works in exactly the same way with both Gaussian and non-Gaussian disturbance distributions [such as heavy-tailed or stable distributions]. The performance of the procedures is examined by simulation. The Monte Carlo experiments conducted focus on : (1) ARCH, GARCH, and ARCH-in-mean alternatives; (2) the case where the variance increases monotonically with : (i) one exogenous variable, and (ii) the mean of the dependent variable; (3) grouped heteroskedasticity; (4) breaks in variance at unknown points. We find that the proposed tests achieve perfect size control and have good power.
Resumo:
In many real world contexts individuals find themselves in situations where they have to decide between options of behaviour that serve a collective purpose or behaviours which satisfy one’s private interests, ignoring the collective. In some cases the underlying social dilemma (Dawes, 1980) is solved and we observe collective action (Olson, 1965). In others social mobilisation is unsuccessful. The central topic of social dilemma research is the identification and understanding of mechanisms which yield to the observed cooperation and therefore resolve the social dilemma. It is the purpose of this thesis to contribute this research field for the case of public good dilemmas. To do so, existing work that is relevant to this problem domain is reviewed and a set of mandatory requirements is derived which guide theory and method development of the thesis. In particular, the thesis focusses on dynamic processes of social mobilisation which can foster or inhibit collective action. The basic understanding is that success or failure of the required process of social mobilisation is determined by heterogeneous individual preferences of the members of a providing group, the social structure in which the acting individuals are contained, and the embedding of the individuals in economic, political, biophysical, or other external contexts. To account for these aspects and for the involved dynamics the methodical approach of the thesis is computer simulation, in particular agent-based modelling and simulation of social systems. Particularly conductive are agent models which ground the simulation of human behaviour in suitable psychological theories of action. The thesis develops the action theory HAPPenInGS (Heterogeneous Agents Providing Public Goods) and demonstrates its embedding into different agent-based simulations. The thesis substantiates the particular added value of the methodical approach: Starting out from a theory of individual behaviour, in simulations the emergence of collective patterns of behaviour becomes observable. In addition, the underlying collective dynamics may be scrutinised and assessed by scenario analysis. The results of such experiments reveal insights on processes of social mobilisation which go beyond classical empirical approaches and yield policy recommendations on promising intervention measures in particular.
Resumo:
This paper discusses predictive motion control of a MiRoSoT robot. The dynamic model of the robot is deduced by taking into account the whole process - robot, vision, control and transmission systems. Based on the obtained dynamic model, an integrated predictive control algorithm is proposed to position precisely with either stationary or moving obstacle avoidance. This objective is achieved automatically by introducing distant constraints into the open-loop optimization of control inputs. Simulation results demonstrate the feasibility of such control strategy for the deduced dynamic model