125 resultados para Operadores de flexibilidade
Resumo:
Due to the current need of the industry to integrate data of the beginning of production originating from of several sources and of transforming them in useful information for sockets of decisions, a search exists every time larger for systems of visualization of information that come to collaborate with that functionality. On the other hand, a common practice nowadays, due to the high competitiveness of the market, it is the development of industrial systems that possess characteristics of modularity, distribution, flexibility, scalability, adaptation, interoperability, reusability and access through web. Those characteristics provide an extra agility and a larger easiness in adapting to the frequent changes of demand of the market. Based on the arguments exposed above, this work consists of specifying a component-based architecture, with the respective development of a system based on that architecture, for the visualization of industrial data. The system was conceived to be capable to supply on-line information and, optionally, historical information of variables originating from of the beginning of production. In this work it is shown that the component-based architecture developed possesses the necessary requirements for the obtaining of a system robust, reliable and of easy maintenance, being, like this, in agreement with the industrial needs. The use of that architecture allows although components can be added, removed or updated in time of execution, through a manager of components through web, still activating more the adaptation process and updating of the system
Resumo:
The need to implement a software architecture that promotes the development of a SCADA supervisory system for monitoring industrial processes simulated with the flexibility of adding intelligent modules and devices such as CLP, according to the specifications of the problem, it was the motivation for this work. In the present study, we developed an intelligent supervisory system on a simulation of a distillation column modeled with Unisim. Furthermore, OLE Automation was used as communication between the supervisory and simulation software, which, with the use of the database, promoted an architecture both scalable and easy to maintain. Moreover, intelligent modules have been developed for preprocessing, data characteristics extraction, and variables inference. These modules were fundamentally based on the Encog software
Resumo:
Ensuring the dependability requirements is essential for the industrial applications since faults may cause failures whose consequences result in economic losses, environmental damage or hurting people. Therefore, faced from the relevance of topic, this thesis proposes a methodology for the dependability evaluation of industrial wireless networks (WirelessHART, ISA100.11a, WIA-PA) on early design phase. However, the proposal can be easily adapted to maintenance and expansion stages of network. The proposal uses graph theory and fault tree formalism to create automatically an analytical model from a given wireless industrial network topology, where the dependability can be evaluated. The evaluation metrics supported are the reliability, availability, MTTF (mean time to failure), importance measures of devices, redundancy aspects and common cause failures. It must be emphasized that the proposal is independent of any tool to evaluate quantitatively the target metrics. However, due to validation issues it was used a tool widely accepted on academy for this purpose (SHARPE). In addition, an algorithm to generate the minimal cut sets, originally applied on graph theory, was adapted to fault tree formalism to guarantee the scalability of methodology in wireless industrial network environments (< 100 devices). Finally, the proposed methodology was validate from typical scenarios found in industrial environments, as star, line, cluster and mesh topologies. It was also evaluated scenarios with common cause failures and best practices to guide the design of an industrial wireless network. For guarantee scalability requirements, it was analyzed the performance of methodology in different scenarios where the results shown the applicability of proposal for networks typically found in industrial environments
Resumo:
The exponential growth in the applications of radio frequency (RF) is accompanied by great challenges as more efficient use of spectrum as in the design of new architectures for multi-standard receivers or software defined radio (SDR) . The key challenge in designing architecture of the software defined radio is the implementation of a wide-band receiver, reconfigurable, low cost, low power consumption, higher level of integration and flexibility. As a new solution of SDR design, a direct demodulator architecture, based on fiveport technology, or multi-port demodulator, has been proposed. However, the use of the five-port as a direct-conversion receiver requires an I/Q calibration (or regeneration) procedure in order to generate the in-phase (I) and quadrature (Q) components of the transmitted baseband signal. In this work, we propose to evaluate the performance of a blind calibration technique without additional knowledge about training or pilot sequences of the transmitted signal based on independent component analysis for the regeneration of I/Q five-port downconversion, by exploiting the information on the statistical properties of the three output signals
Resumo:
In this work, the Markov chain will be the tool used in the modeling and analysis of convergence of the genetic algorithm, both the standard version as for the other versions that allows the genetic algorithm. In addition, we intend to compare the performance of the standard version with the fuzzy version, believing that this version gives the genetic algorithm a great ability to find a global optimum, own the global optimization algorithms. The choice of this algorithm is due to the fact that it has become, over the past thirty yares, one of the more importan tool used to find a solution of de optimization problem. This choice is due to its effectiveness in finding a good quality solution to the problem, considering that the knowledge of a good quality solution becomes acceptable given that there may not be another algorithm able to get the optimal solution for many of these problems. However, this algorithm can be set, taking into account, that it is not only dependent on how the problem is represented as but also some of the operators are defined, to the standard version of this, when the parameters are kept fixed, to their versions with variables parameters. Therefore to achieve good performance with the aforementioned algorithm is necessary that it has an adequate criterion in the choice of its parameters, especially the rate of mutation and crossover rate or even the size of the population. It is important to remember that those implementations in which parameters are kept fixed throughout the execution, the modeling algorithm by Markov chain results in a homogeneous chain and when it allows the variation of parameters during the execution, the Markov chain that models becomes be non - homogeneous. Therefore, in an attempt to improve the algorithm performance, few studies have tried to make the setting of the parameters through strategies that capture the intrinsic characteristics of the problem. These characteristics are extracted from the present state of execution, in order to identify and preserve a pattern related to a solution of good quality and at the same time that standard discarding of low quality. Strategies for feature extraction can either use precise techniques as fuzzy techniques, in the latter case being made through a fuzzy controller. A Markov chain is used for modeling and convergence analysis of the algorithm, both in its standard version as for the other. In order to evaluate the performance of a non-homogeneous algorithm tests will be applied to compare the standard fuzzy algorithm with the genetic algorithm, and the rate of change adjusted by a fuzzy controller. To do so, pick up optimization problems whose number of solutions varies exponentially with the number of variables
Resumo:
A 2.5D ray-tracing propagation model is proposed to predict radio loss in indoor environment. Specifically, we opted for the Shooting and Bouncing Rays (SBR) method, together with the Geometrieal Theory of Diffrartion (GTD). Besides the line-of-sight propagation (LOS), we consider that the radio waves may experience reflection, refraction, and diffraction (NLOS). In the Shooting and Bouncing Rays (SBR) method, the transmitter antenna launches a bundle of rays that may or may not reach the receiver. Considering the transmitting antenna as a point, the rays will start to launch from this position and can reach the receiver either directly or after reflections, refractions, diffractions, or even after any combination of the previous effects. To model the environment, a database is built to record geometrical characteristics and information on the constituent materials of the scenario. The database works independently of the simulation program, allowing robustness and flexibility to model other seenarios. Each propagation mechanism is treated separately. In line-of-sight propagation, the main contribution to the received signal comes from the direct ray, while reflected, refracted, and diffracted signal dominate when the line-of-sight is blocked. For this case, the transmitted signal reaches the receiver through more than one path, resulting in a multipath fading. The transmitting channel of a mobile system is simulated by moving either the transmitter or the receiver around the environment. The validity of the method is verified through simulations and measurements. The computed path losses are compared with the measured values at 1.8 GHz ftequency. The results were obtained for the main corridor and room classes adjacent to it. A reasonable agreement is observed. The numerical predictions are also compared with published data at 900 MHz and 2.44 GHz frequencies showing good convergence
Resumo:
ln this work the implementation of the SOM (Self Organizing Maps) algorithm or Kohonen neural network is presented in the form of hierarchical structures, applied to the compression of images. The main objective of this approach is to develop an Hierarchical SOM algorithm with static structure and another one with dynamic structure to generate codebooks (books of codes) in the process of the image Vector Quantization (VQ), reducing the time of processing and obtaining a good rate of compression of images with a minimum degradation of the quality in relation to the original image. Both self-organizing neural networks developed here, were denominated HSOM, for static case, and DHSOM, for the dynamic case. ln the first form, the hierarchical structure is previously defined and in the later this structure grows in an automatic way in agreement with heuristic rules that explore the data of the training group without use of external parameters. For the network, the heuristic mIes determine the dynamics of growth, the pruning of ramifications criteria, the flexibility and the size of children maps. The LBO (Linde-Buzo-Oray) algorithm or K-means, one ofthe more used algorithms to develop codebook for Vector Quantization, was used together with the algorithm of Kohonen in its basic form, that is, not hierarchical, as a reference to compare the performance of the algorithms here proposed. A performance analysis between the two hierarchical structures is also accomplished in this work. The efficiency of the proposed processing is verified by the reduction in the complexity computational compared to the traditional algorithms, as well as, through the quantitative analysis of the images reconstructed in function of the parameters: (PSNR) peak signal-to-noise ratio and (MSE) medium squared error
Resumo:
This paper describes the design, implementation and enforcement of a system for industrial process control based on fuzzy logic and developed using Java, with support for industrial communication protocol through the OPC (Ole for Process Control). Besides the java framework, the software is completely independent from other platforms. It provides friendly and functional tools for modeling, construction and editing of complex fuzzy inference systems, and uses these logical systems in control of a wide variety of industrial processes. The main requirements of the developed system should be flexibility, robustness, reliability and ease of expansion
Resumo:
The Methods for compensation of harmonic currents and voltages have been widely used since these methods allow to reduce to acceptable levels the harmonic distortion in the voltages or currents in a power system, and also compensate reactive. The reduction of harmonics and reactive contributes to the reduction of losses in transmission lines and electrical machinery, increasing the power factor, reduce the occurrence of overvoltage and overcurrent. The active power filter is the most efficient method for compensation of harmonic currents and voltages. The active power filter is necessary to use current and voltage controllers loop. Conventionally, the current and voltage control loop of active filter has been done by proportional controllers integrative. This work, investigated the use of a robust adaptive control technique on the shunt active power filter current and voltage control loop to increase robustness and improve the performance of active filter to compensate for harmonics. The proposed control scheme is based on a combination of techniques for adaptive control pole placement and variable structure. The advantages of the proposed method over conventional ones are: lower total harmonic distortion, more flexibility, adaptability and robustness to the system. Moreover, the proposed control scheme improves the performance and improves the transient of active filter. The validation of the proposed technique was verified initially by a simulation program implemented in C++ language and then experimental results were obtained using a prototype three-phase active filter of 1 kVA
Resumo:
This work treats of an implementation OFDMA baseband processor in hardware for LTE Downlink. The LTE or Long Term Evolution consist the last stage of development of the technology called 3G (Mobile System Third Generation) which offers an increasing in data rate and more efficiency and flexibility in transmission with application of advanced antennas and multiple carriers techniques. This technology applies in your physical layer the OFDMA technical (Orthogonal Frequency Division Multiple Access) for generation of signals and mapping of physical resources in downlink and has as base theoretical to OFDM multiple carriers technique (Orthogonal Frequency Division Multiplexing). With recent completion of LTE specifications, different hardware solutions have been developed, mainly, to the level symbol processing where the implementation of OFDMA processor in base band is commonly considered, because it is also considered a basic architecture of others important applications. For implementation of processor, the reconfigurable hardware offered by devices as FPGA are considered which shares not only to meet the high requirements of flexibility and adaptability of LTE as well as offers possibility of an implementation quick and efficient. The implementation of processor in reconfigurable hardware meets the specifications of LTE physical layer as well as have the flexibility necessary for to meet others standards and application which use OFDMA processor as basic architecture for your systems. The results obtained through of simulation and verification functional system approval the functionality and flexibility of processor implemented
Resumo:
The evolution of automation in recent years made possible the continuous monitoring of the processes of industrial plants. With this advance, the amount of information that automation systems are subjected to increased significantly. The alarms generated by the monitoring equipment are a major contributor to this increase, and the equipments are usually deployed in industrial plants without a formal methodology, which entails an increase in the number of alarms generated, thus overloading the alarm system and therefore the operators of such plants. In this context, the works of alarm management comes up with the objective of defining a formal methodology for installation of new equipment and detect problems in existing settings. This thesis aims to propose a set of metrics for the evaluation of alarm systems already deployed, so that you can identify the health of this system by analyzing the proposed indices and comparing them with parameters defined in the technical norms of alarm management. In addition, the metrics will track the work of alarm management, verifying if it is improving the quality of the alarm system. To validate the proposed metrics, data from actual process plants of the petrochemical industry were used
Resumo:
The method of artificial lift of progressing cavity pump is very efficient in the production of oils with high viscosity and oils that carry a great amount of sand. This characteristic converted this lift method into the second most useful one in oil fields production. As it grows the number of its applications it also increases the necessity to dominate its work in a way to define it the best operational set point. To contribute to the knowledge of the operational method of artificial lift of progressing cavity pump, this work intends to develop a computational simulator for oil wells equipped with an artificial lift system. The computational simulator of the system will be able to represent its dynamic behavior when submitted to the various operational conditions. The system was divided into five subsystems: induction motor, multiphase flows into production tubing, rod string, progressing cavity pump and annular tubing-casing. The modeling and simulation of each subsystem permitted to evaluate the dynamic characteristics that defined the criteria connections. With the connections of the subsystems it was possible to obtain the dynamic characteristics of the most important arrays belonging to the system, such as: pressure discharge, pressure intake, pumping rate, rod string rotation and torque applied to polish string. The shown results added to a friendly graphical interface converted the PCP simulator in a great potential tool with a didactic characteristic in serving the technical capability for the system operators and also permitting the production engineering to achieve a more detail analysis of the dynamic operational oil wells equipped with the progressing cavity pump
Resumo:
This work addresses the dynamic control problem of two-wheeled differentially driven non-holonomic mobile robot. Strategies for robot positioning control and robot orientating control are presented. Such strategies just require information about the robot con¯guration (x, y and teta), which can be collected by an absolute positioning system. The strategies development is related to a change on the controlled variables for such systems, from x, y and teta to s (denoting the robot linear displacement) and teta, and makes use of the polar coordinates representation for the robot kinematic model. Thus, it is possible to obtain a linear representation for the mobile robot dynamic model and to develop such strategies. It is also presented that such strategies allow the use of linear controllers to solve the control problem. It is shown that there is flexibility to choice the linear controller (P, PI, PID, Model Matching techniques, others) to be implemented. This work presents an introduction to mobile robotics and their characteristics followed by the control strategies development and controllers design. Finally, simulated and experimental results are presented and commented
Resumo:
There are two main approaches for using in adaptive controllers. One is the so-called model reference adaptive control (MRAC), and the other is the so-called adaptive pole placement control (APPC). In MRAC, a reference model is chosen to generate the desired trajectory that the plant output has to follow, and it can require cancellation of the plant zeros. Due to its flexibility in choosing the controller design methodology (state feedback, compensator design, linear quadratic, etc.) and the adaptive law (least squares, gradient, etc.), the APPC is the most general type of adaptive control. Traditionally, it has been developed in an indirect approach and, as an advantage, it may be applied to non-minimum phase plants, because do not involve plant zero-pole cancellations. The integration to variable structure systems allows to aggregate fast transient and robustness to parametric uncertainties and disturbances, as well. In this work, a variable structure adaptive pole placement control (VS-APPC) is proposed. Therefore, new switching laws are proposed, instead of using the traditional integral adaptive laws. Additionally, simulation results for an unstable first order system and simulation and practical results for a three-phase induction motor are shown
Resumo:
One of the most important decisions to turn a substation automatic and no attended it relates to the communication media between this substation and Operation Center. Generally energy companies uses radio or optic fiber, depending of distances and infrastructure of each situation. This rule applies to common substations. Mobile substations are a particular case, therefore they are conceived for use at provisional situations, emergencies, preventive or corrective maintenance. Thus the telecommunication solution used at common substations are not applied so easily to mobile substations, due absence of infrastructure (media) or difficulty to insert the mobile substation data in existing automation network not long. The ideal media must supply covering in a great geographic area to satisfy presented requirements. The implantation costs of this big infrastructure are expensive, however a existing operator may be used. Two services that fulfill that requirements are satellite and cellular telephony. This work presents a solution for automation of mobile substations through satellite. It was successfully implanted at a brazilian electric energy concessionaire named COSERN. The operation became transparent to operators. Other gotten benefits had been operational security, quality in the supply of electric energy and costs reduction. The project presented is a new solution, designed to substations and general applications where few data should be transmitted, but there is difficulties in relation to the media. Despite the satellite having been used, the same resulted can be gotten using celullar telephony, through Short Messages or packet networks as GPRS or EDGE.