172 resultados para Algoritmo memético


Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work consists on the study of two important problems arising from the operations of petroleum and natural gas industries. The first problem the pipe dimensioning problem on constrained gas distribution networks consists in finding the least cost combination of diameters from a discrete set of commercially available ones for the pipes of a given gas network, such that it respects minimum pressure requirements at each demand node and upstream pipe conditions. On its turn, the second problem the piston pump unit routing problem comes from the need of defining the piston pump unit routes for visiting a number of non-emergent wells in on-shore fields, i.e., wells which don t have enough pressure to make the oil emerge to surface. The periodic version of this problem takes into account the wells re-filling equation to provide a more accurate planning in the long term. Besides the mathematical formulation of both problems, an exact algorithm and a taboo search were developed for the solution of the first problem and a theoretical limit and a ProtoGene transgenetic algorithm were developed for the solution of the second problem. The main concepts of the metaheuristics are presented along with the details of their application to the cited problems. The obtained results for both applications are promising when compared to theoretical limits and alternate solutions, either relative to the quality of the solutions or to associated running time

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Hiker Dice was a game recently proposed in a software designed by Mara Kuzmich and Leonardo Goldbarg. In the game a dice is responsible for building a trail on an n x m board. As the dice waits upon a cell on the board, it prints the side that touches the surface. The game shows the Hamiltonian Path Problem Simple Maximum Hiker Dice (Hidi-CHS) in trays Compact Nth , this problem is then characterized by looking for a Hamiltonian Path that maximize the sum of marked sides on the board. The research now related, models the problem through Graphs, and proposes two classes of solution algorithms. The first class, belonging to the exact algorithms, is formed by a backtracking algorithm planed with a return through logical rules and limiting the best found solution. The second class of algorithms is composed by metaheuristics type Evolutionary Computing, Local Ramdomized search and GRASP (Greed Randomized Adaptative Search). Three specific operators for the algorithms were created as follows: restructuring, recombination with two solutions and random greedy constructive.The exact algorithm was teste on 4x4 to 8x8 boards exhausting the possibility of higher computational treatment of cases due to the explosion in processing time. The heuristics algorithms were tested on 5x5 to 14x14 boards. According to the applied methodology for evaluation, the results acheived by the heuristics algorithms suggests a better performance for the GRASP algorithm

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Scientific Algorithms are a new metaheuristics inspired in the scientific research process. The new method introduces the idea of theme to search the solution space of hard problems. The inspiration for this class of algorithms comes from the act of researching that comprises thinking, knowledge sharing and disclosing new ideas. The ideas of the new method are illustrated in the Traveling Salesman Problem. A computational experiment applies the proposed approach to a new variant of the Traveling Salesman Problem named Car Renter Salesman Problem. The results are compared to state-of-the-art algorithms for the latter problem

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Image segmentation is the process of subdiving an image into constituent regions or objects that have similar features. In video segmentation, more than subdividing the frames in object that have similar features, there is a consistency requirement among segmentations of successive frames of the video. Fuzzy segmentation is a region growing technique that assigns to each element in an image (which may have been corrupted by noise and/or shading) a grade of membership between 0 and 1 to an object. In this work we present an application that uses a fuzzy segmentation algorithm to identify and select particles in micrographs and an extension of the algorithm to perform video segmentation. Here, we treat a video shot is treated as a three-dimensional volume with different z slices being occupied by different frames of the video shot. The volume is interactively segmented based on selected seed elements, that will determine the affinity functions based on their motion and color properties. The color information can be extracted from a specific color space or from three channels of a set of color models that are selected based on the correlation of the information from all channels. The motion information is provided into the form of dense optical flows maps. Finally, segmentation of real and synthetic videos and their application in a non-photorealistic rendering (NPR) toll are presented

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este trabalho apresenta um algoritmo transgenético híbrido para a solução de um Problema de Configuração de uma Rede de Distribuição de Gás Natural. O problema da configuração dessas redes requer a definição de um traçado por onde os dutos devem ser colocados para atender aos clientes. É estudada neste trabalho uma maneira de conectar os clientes em uma rede com arquitetura em forma de árvore. O objetivo é minimizar o custo de construção da rede, mesmo que para isso alguns clientes que não proporcionam lucros deixem de ser atendidos. Esse problema pode ser formulado computacionalmente através do Problema de Steiner com Prêmios. Este é um problema de otimização combinatória da classe dos NPÁrduos. Este trabalho apresenta um algoritmo heurístico para a solução do problema. A abordagem utilizada é chamada de Algoritmos Transgenéticos, que se enquadram na categoria dos algoritmos evolucionários. Para a geração de soluções inicias é utilizado um algoritmo primaldual, e pathrelinking é usado como intensificador

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Non-Photorealisitc Rendering (NPR) is a class of techniques that aims to reproduce artistic techniques, trying to express feelings and moods on the rendered scenes, giving an aspect of that they had been made "by hand". Another way of defining NPR is that it is the processing of scenes, images or videos into artwork, generating scenes, images or videos that can have the visual appeal of pieces of art, expressing the visual and emotional characteristics of artistic styles. This dissertation presents a new method of NPR for stylization of images and videos, based on a typical artistic expression of the Northeast region of Brazil, that uses colored sand to compose landscape images on the inner surface of glass bottles. This method is comprised by one technique for generating 2D procedural textures of sand, and two techniques that mimic effects created by the artists using their tools. It also presents a method for generating 21 2D animations in sandbox from the stylized video. The temporal coherence within these stylized videos can be enforced on individual objects with the aid of a video segmentation algorithm. The present techniques in this work were used on stylization of synthetic and real videos, something close to impossible to be produced by artist in real life

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Curimataú estuary is located in the oriental coast of Rio Grande do Norte State in Brazil. Its importance resides in the fact that this region possesses one of the last portions of preserved mangrove in the Rio Grande do Norte State. Nevertheless, it has been severely affected by many anthropogenic activities, as sugarcane monoculture and shrimp farming. Former works demonstrated that an accumulation of heavy metals is occurring in oysters in this estuary, and perhaps it could be explained by the input of metals in this ecosystem deriving from the shrimp farming. To better understanding the origin of these metals, bottom sediment samples, cores and suspended particulate matter were collected for a characterization of metal concentrations (Al, Ba, Cd, Cu, Cr, Fe, Mn, Ni, Pb, Zn) and to determine the potentially bioavailable metals. Additionally, the enrichment ratio for each element analyzed was calculated. The mineralogical composition of sediment samples and cores were obtained by X-ray diffraction. Moreover, data of orbital remote sensing were used in order to detect and quantify suspended matter by applying a logarithmic algorithm. Geochemical data of bottom sediments and cores revealed that, excepting Ba and Pb, the elements analyzed presented concentrations characteristic of an unpolluted ecosystem (Al: 0,25 - 8,76 %; Ba: 3,03 - 870 µg.g-1; Cd: < 0,25 µg.g-1; Cr: 1,72 - 82,4 µg.g-1; Cu: 0,12 -25,3 µg.g-1; Pb: 0,38 - 23,7 µg.g-1; Fe: 0,10 - 5,82 %; Mn: 15,1 - 815 µg.g-1; Ni: 0,14 - 36,1 µg.g-1; Zn: 1,37 - 113 µg.g-1). During the dry season a distribution pattern was observed, with higher metal concentrations in the margins, decreasing toward the central portion of the channel. These metal concentrations were well correlated with mineralogical compositions, with clay minerals prevailing at the margins, and quartz and feldspar in the center. However, this pattern was not observed during the wet season, probably because of the high water flux that disturbed bottom sediments. But, as observed for the dry season, a good correlation between metal concentrations and mineralogical composition was also observed for the wet season, with high metal concentrations where there were high quantities of clay minerals. Low enrichment ratios were obtained for the majority of elements analyzed, excepting for Mn, Ba and Pb. Manganese presented the higher ratios downstream for both seasons, and it can be an evidence of anthropogenic impact by shrimp farming. As barium and lead concentrations in sediment samples presented analytical problems during the total sample digestion, one cannot be sure that the ratios obtained correspond to the reality. The highest metal concentrations in particulate matter were obtained in the portion dominated by fluvial transport for all metals analyzed, excepting for copper. Barium and zinc were the only elements that presented elevated concentrations that are not common of unpolluted ecosystems (Ba: 5730 - 8355 µg.g-1; Zn: 3899 - 4348 µg.g-1). However, these high concentrations could not be related to the shrimp farming and waste waters from the town of Canguaretama, once they were obtained from the fluvial particulate matter, that is upstream from the activities above mentioned. The application of the logarithmic algorithm to the processed LANDSAT image was well succeeded, although the acquired image does not correspond exactly to the field campaigns. The IKONOS image provided very detailed views of the suspended sediment concentration at the estuary, as the mixture of distinct water flows at the confluence of Cunhaú and Curimataú rivers, with more turbid waters from Cunhaú river, that is directly affected by effluents from shrimp farming and urban waste waters deriving from the town of Canguaretama

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work an algorithm for fault location is proposed. It contains the following functions: fault detection, fault classification and fault location. Mathematical Morphology is used to process currents obtained in the monitored terminals. Unlike Fourier and Wavelet transforms that are usually applied to fault location, the Mathematical Morphology is a non-linear operation that uses only basic operation (sum, subtraction, maximum and minimum). Thus, Mathematical Morphology is computationally very efficient. For detection and classification functions, the Morphological Wavelet was used. On fault location module the Multiresolution Morphological Gradient was used to detect the traveling waves and their polarities. Hence, recorded the arrival in the two first traveling waves incident at the measured terminal and knowing the velocity of propagation, pinpoint the fault location can be estimated. The algorithm was applied in a 440 kV power transmission system, simulated on ATP. Several fault conditions where studied and the following parameters were evaluated: fault location, fault type, fault resistance, fault inception angle, noise level and sampling rate. The results show that the application of Mathematical Morphology in faults location is very promising

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The robustness and performance of the Variable Structure Adaptive Pole Placement Controller are evaluated in this work, where this controller is applied to control a synchronous generator connected to an infinite bus. The evaluation of the robustness of this controller will be accomplished through simulations, where the control algorithm was subjected to adverse conditions, such as: disturbances, parametric variations and unmodeled dynamic. It was also made a comparison of this control strategy with another one, using classic controllers. In the simulations, it is used a coupled model of the synchronous generator which variables have a high degree of coupling, in other words, if there is a change in the input variables of the generator, it will change all outputs simultaneously. The simulation results show which control strategy performs better and is more robust to disturbances, parametric variations and unmodeled dynamics for the control of Synchronous Generator

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The separation methods are reduced applications as a result of the operational costs, the low output and the long time to separate the uids. But, these treatment methods are important because of the need for extraction of unwanted contaminants in the oil production. The water and the concentration of oil in water should be minimal (around 40 to 20 ppm) in order to take it to the sea. Because of the need of primary treatment, the objective of this project is to study and implement algorithms for identification of polynomial NARX (Nonlinear Auto-Regressive with Exogenous Input) models in closed loop, implement a structural identification, and compare strategies using PI control and updated on-line NARX predictive models on a combination of three-phase separator in series with three hydro cyclones batteries. The main goal of this project is to: obtain an optimized process of phase separation that will regulate the system, even in the presence of oil gushes; Show that it is possible to get optimized tunings for controllers analyzing the mesh as a whole, and evaluate and compare the strategies of PI and predictive control applied to the process. To accomplish these goals a simulator was used to represent the three phase separator and hydro cyclones. Algorithms were developed for system identification (NARX) using RLS(Recursive Least Square), along with methods for structure models detection. Predictive Control Algorithms were also implemented with NARX model updated on-line, and optimization algorithms using PSO (Particle Swarm Optimization). This project ends with a comparison of results obtained from the use of PI and predictive controllers (both with optimal state through the algorithm of cloud particles) in the simulated system. Thus, concluding that the performed optimizations make the system less sensitive to external perturbations and when optimized, the two controllers show similar results with the assessment of predictive control somewhat less sensitive to disturbances

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This research aims at developing a variable structure adaptive backstepping controller (VS-ABC) by using state observers for SISO (Single Input Single Output), linear and time invariant systems with relative degree one. Therefore, the lters were replaced by a Luenberger Adaptive Observer and the control algorithm uses switching laws. The presented simulations compare the controller performance, considering when the state variables are estimated by an observer, with the case that the variables are available for measurement. Even with numerous performance advantages, adaptive backstepping controllers still have very complex algorithms, especially when the system state variables are not measured, since the use of lters on the plant input and output is not something trivial. As an attempt to make the controller design more intuitive, an adaptive observer as an alternative to commonly used K lters can be used. Furthermore, since the states variables are considered known, the controller has a reduction on the dependence of the unknown plant parameters on the design. Also, switching laws could be used in the controller instead of the traditional integral adaptive laws because they improve the system transient performance and increase the robustness against external disturbances in the plant input

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The use of Multiple Input Multiple Output (MIMO) systems has permitted the recent evolution of wireless communication standards. The Spatial Multiplexing MIMO technique, in particular, provides a linear gain at the transmission capacity with the minimum between the numbers of transmit and receive antennas. To obtain a near capacity performance in SM-MIMO systems a soft decision Maximum A Posteriori Probability MIMO detector is necessary. However, such detector is too complex for practical solutions. Hence, the goal of a MIMO detector algorithm aimed for implementation is to get a good approximation of the ideal detector while keeping an acceptable complexity. Moreover, the algorithm needs to be mapped to a VLSI architecture with small area and high data rate. Since Spatial Multiplexing is a recent technique, it is argued that there is still much room for development of related algorithms and architectures. Therefore, this thesis focused on the study of sub optimum algorithms and VLSI architectures for broadband MIMO detector with soft decision. As a result, novel algorithms have been developed starting from proposals of optimizations for already established algorithms. Based on these results, new MIMO detector architectures with configurable modulation and competitive area, performance and data rate parameters are here proposed. The developed algorithms have been extensively simulated and the architectures were synthesized so that the results can serve as a reference for other works in the area

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this work, we propose a two-stage algorithm for real-time fault detection and identification of industrial plants. Our proposal is based on the analysis of selected features using recursive density estimation and a new evolving classifier algorithm. More specifically, the proposed approach for the detection stage is based on the concept of density in the data space, which is not the same as probability density function, but is a very useful measure for abnormality/outliers detection. This density can be expressed by a Cauchy function and can be calculated recursively, which makes it memory and computational power efficient and, therefore, suitable for on-line applications. The identification/diagnosis stage is based on a self-developing (evolving) fuzzy rule-based classifier system proposed in this work, called AutoClass. An important property of AutoClass is that it can start learning from scratch". Not only do the fuzzy rules not need to be prespecified, but neither do the number of classes for AutoClass (the number may grow, with new class labels being added by the on-line learning process), in a fully unsupervised manner. In the event that an initial rule base exists, AutoClass can evolve/develop it further based on the newly arrived faulty state data. In order to validate our proposal, we present experimental results from a level control didactic process, where control and error signals are used as features for the fault detection and identification systems, but the approach is generic and the number of features can be significant due to the computationally lean methodology, since covariance or more complex calculations, as well as storage of old data, are not required. The obtained results are significantly better than the traditional approaches used for comparison

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We studied the Ising model ferromagnetic as spin-1/2 and the Blume-Capel model as spin-1, > 0 on small world network, using computer simulation through the Metropolis algorithm. We calculated macroscopic quantities of the system, such as internal energy, magnetization, specific heat, magnetic susceptibility and Binder cumulant. We found for the Ising model the same result obtained by Koreans H. Hong, Beom Jun Kim and M. Y. Choi [6] and critical behavior similar Blume-Capel model

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this work, the study of some complex systems is done with use of two distinct procedures. In the first part, we have studied the usage of Wavelet transform on analysis and characterization of (multi)fractal time series. We have test the reliability of Wavelet Transform Modulus Maxima method (WTMM) in respect to the multifractal formalism, trough the calculation of the singularity spectrum of time series whose fractality is well known a priori. Next, we have use the Wavelet Transform Modulus Maxima method to study the fractality of lungs crackles sounds, a biological time series. Since the crackles sounds are due to the opening of a pulmonary airway bronchi, bronchioles and alveoli which was initially closed, we can get information on the phenomenon of the airway opening cascade of the whole lung. Once this phenomenon is associated with the pulmonar tree architecture, which displays fractal geometry, the analysis and fractal characterization of this noise may provide us with important parameters for comparison between healthy lungs and those affected by disorders that affect the geometry of the tree lung, such as the obstructive and parenchymal degenerative diseases, which occurs, for example, in pulmonary emphysema. In the second part, we study a site percolation model for square lattices, where the percolating cluster grows governed by a control rule, corresponding to a method of automatic search. In this model of percolation, which have characteristics of self-organized criticality, the method does not use the automated search on Leaths algorithm. It uses the following control rule: pt+1 = pt + k(Rc − Rt), where p is the probability of percolation, k is a kinetic parameter where 0 < k < 1 and R is the fraction of percolating finite square lattices with side L, LxL. This rule provides a time series corresponding to the dynamical evolution of the system, in particular the likelihood of percolation p. We proceed an analysis of scaling of the signal obtained in this way. The model used here enables the study of the automatic search method used for site percolation in square lattices, evaluating the dynamics of their parameters when the system goes to the critical point. It shows that the scaling of , the time elapsed until the system reaches the critical point, and tcor, the time required for the system loses its correlations, are both inversely proportional to k, the kinetic parameter of the control rule. We verify yet that the system has two different time scales after: one in which the system shows noise of type 1 f , indicating to be strongly correlated. Another in which it shows white noise, indicating that the correlation is lost. For large intervals of time the dynamics of the system shows ergodicity