877 resultados para Multi-objective evolutionary algorithm


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Devido ao auge do crescimento industrial na Região Norte e, em especial, o Pólo Industrial de Manaus (PIM), são necessários obter ferramentas matemáticas que facilitem ao especialista tomar decisões sobre a seleção e dimensionamento dos filtros harmônicos que proporcionam neutralizar os efeitos prejudiciais dos harmônicos gerados pelas cargas não lineares da indústria e alcançar conformidade com os padrões das normas de qualidade de energia correspondentes. Além disso, como os filtros harmônicos passivos têm a capacidade de gerar potência reativa à rede, estes meios são eficazes compensadores de potência reativa e, portanto, podem conseguir uma economia significativa no faturamento de energia elétrica consumida por essas instalações industriais. Esta tese tem como objetivo geral desenvolver um método matemático e uma ferramenta computacional para a seleção da configuração e parâmetros do projeto de um conjunto de filtros harmônicos passivos para sistemas elétricos industriais. Nesta ótica, o problema de otimização da compensação de harmônicos por meio de filtros passivos foi formulado como um problema multiobjetivo que considera tanto os objetivos da redução da distorção harmônica como da efetividade econômica do projeto considerando as características das tarifas brasileiras. Todavia, a formulação apresentada considera as restrições relevantes impostas pelas normas brasileiras e estrangeiras. A solução computacional para este problema foi conseguida, usando o algoritmo genético NSGA-II que determina um conjunto de soluções ótimas de Pareto (Fronteira) que permitem ao projetista escolher as soluções mais adequadas para o problema. Por conseguinte, a ferramenta computacional desenvolvida tem várias novidades como: não só calcula os parâmetros que caracterizam os filtros, como também seleciona o tipo de configuração e o número de ramos do filtro em cada barra candidata de acordo com um conjunto de configurações pré-estabelecidas; têm implementada duas normas para a avaliação das restrições de qualidade de energia (Prodist-Módulo 8 e IEEE 519-92) que podem ser selecionadas pelo usuário; determina soluções com bons indicadores de desempenho para vários cenários característicos e não característicos do sistema que permitem a representação das as variações diárias da carga; das variações dos parâmetros do sistema e dos filtros; avalia o custo das contas de energia numa rede elétrica industrial que tem diferentes condições de operação (cenários característicos); e avalia o efeito econômico de filtros de harmônicos como compensadores de potência reativa. Para desenvolver a ferramenta computacional adequada desta tese, foi empregado um modelo trifásico em coordenadas de fase para redes de energia elétrica industriais e de serviços onde foram feitos vários programas utilizando várias ferramentas computacionais adicionais. Estas ferramentas compreendem um programa de varredura de freqüência, um programa do fluxo de harmônicos por injeção de correntes e um programa de fluxo de potência à freqüência fundamental. Os resultados positivos desta tese, a partir da análise de vários exemplos práticos, mostram as vantagens do método desenvolvido.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Esta dissertação apresenta uma metodologia baseada em algoritmo genético (AG) para determinar modelos dinâmicos equivalentes de parques eólicos com geradores de indução em gaiola de esquilo ( GIGE) e geradores de indução duplamente alimentados ( GIDA), apresentando parâmetros elétricos e mecânicos distintos. A técnica se baseia em uma formulação multiobjetiva solucionada por um AG para minimizar os erros quadráticos das potências ativa e reativa entre modelo de um único gerador equivalente e o modelo do parque eólico investigado. A influência do modelo equivalente do parque eólico no comportamento dinâmico dos geradores síncronos é também investigada por meio do método proposto. A abordagem é testada em um parque eólico de 10MW composto por quatro turbinas eólicas ( 2x2MW e 2x3MW), consistindo alternadamente de geradores GIGE e GIDA interligados a uma barra infinita e posteriormente a rede elétrica do IEEE 14 barras. Os resultados obtidos pelo uso do modelo dinâmico detalhado para a representação do parque eólico são comparados aos do modelo equivalente proposto para avaliar a precisão e o custo computacional do modelo proposto.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

O trabalho em pauta tem como objetivo o modelamento da crosta, através da inversão de dados de refração sísmica profunda, segundo camadas planas horizontais lateralmente homogêneas, sobre um semi-espaço. O modelo direto é dado pela expressão analítica da curva tempo-distância como uma função que depende da distância fonte-estação e do vetor de parâmetros velocidades e espessuras de cada camada, calculado segundo as trajetórias do raio sísmico, regidas pela Lei de Snell. O cálculo dos tempos de chegada por este procedimento, exige a utilização de um modelo cujas velocidades sejam crescentes com a profundidade, de modo que a ocorrência das camadas de baixa velocidade (CBV) é contornada pela reparametrização do modelo, levando-se em conta o fato de que o topo da CBV funciona apenas como um refletor do raio sísmico, e não como refrator. A metodologia de inversão utilizada tem em vista não só a determinação das soluções possíveis, mas também a realização de uma análise sobre as causas responsáveis pela ambiguidade do problema. A região de pesquisa das prováveis soluções é vinculada segundo limites superiores e inferiores para cada parâmetro procurado, e pelo estabelecimento de limites superiores para os valores de distâncias críticas, calculadas a partir do vetor de parâmetros. O processo de inversão é feito utilizando-se uma técnica de otimização do ajuste de curvas através da busca direta no espaço dos parâmetros, denominado COMPLEX. Esta técnica apresenta a vantagem de poder ser utilizada com qualquer função objeto, e ser bastante prática na obtenção de múltiplas soluções do problema. Devido a curva tempo-distância corresponder ao caso de uma multi-função, o algoritmo foi adaptado de modo a minimizar simultaneamente várias funções objetos, com vínculos nos parâmetros. A inversão é feita de modo a se obter um conjunto de soluções representativas do universo existente. Por sua vez, a análise da ambiguidade é realizada pela análise fatorial modo-Q, através da qual é possível se caracterizar as propriedades comuns existentes no elenco das soluções analisadas. Os testes com dados sintéticos e reais foram feitos tendo como aproximação inicial ao processo de inversão, os valores de velocidades e espessuras calculados diretamente da interpretação visual do sismograma. Para a realização dos primeiros, utilizou-se sismogramas calculados pelo método da refletividade, segundo diferentes modelos. Por sua vez, os testes com dados reais foram realizados utilizando-se dados extraídos de um dos sismogramas coletados pelo projeto Lithospheric Seismic Profile in Britain (LISPB), na região norte da Grã-Bretanha. Em todos os testes foi verificado que a geometria do modelo possui um maior peso na ambiguidade do problema, enquanto os parâmetros físicos apresentam apenas suaves variações, no conjunto das soluções obtidas.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pós-graduação em Engenharia Elétrica - FEIS

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pós-graduação em Engenharia Elétrica - FEB

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a technique for performing analog design synthesis at circuit level providing feedback to the designer through the exploration of the Pareto frontier. A modified simulated annealing which is able to perform crossover with past anchor points when a local minimum is found which is used as the optimization algorithm on the initial synthesis procedure. After all specifications are met, the algorithm searches for the extreme points of the Pareto frontier in order to obtain a non-exhaustive exploration of the Pareto front. Finally, multi-objective particle swarm optimization is used to spread the results and to find a more accurate frontier. Piecewise linear functions are used as single-objective cost functions to produce a smooth and equal convergence of all measurements to the desired specifications during the composition of the aggregate objective function. To verify the presented technique two circuits were designed, which are: a Miller amplifier with 96 dB Voltage gain, 15.48 MHz unity gain frequency, slew rate of 19.2 V/mu s with a current supply of 385.15 mu A, and a complementary folded cascode with 104.25 dB Voltage gain, 18.15 MHz of unity gain frequency and a slew rate of 13.370 MV/mu s. These circuits were synthesized using a 0.35 mu m technology. The results show that the method provides a fast approach for good solutions using the modified SA and further good Pareto front exploration through its connection to the particle swarm optimization algorithm.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Predição de estruturas de proteínas (PSP) é um problema computacionalmente complexo. Modelos simplificados da molécula proteica (como o Modelo HP) e o uso de Algoritmos Evolutivos (AEs) estão entre as principais técnicas investigadas para PSP. Entretanto, a avaliação de uma estrutura representada pelo Modelo HP considera apenas o número de contatos hidrofóbicos, não possibilitando distinguir entre estruturas com o mesmo número de contatos hidrofóbicos. Neste trabalho, é apresentada uma nova formulação multiobjetivo para PSP em Modelo HP. Duas métricas são avaliadas: o número de contatos hidrofóbicos e a distância entre os aminoácidos hidrofóbicos, as quais são tratados pelo AE Multiobjetivo em Tabelas (AEMT). O algoritmo mostrou-se rápido e robusto.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Small scale fluid flow systems have been studied for various applications, such as chemical reagent dosages and cooling devices of compact electronic components. This work proposes to present the complete cycle development of an optimized heat sink designed by using Topology Optimization Method (TOM) for best performance, including minimization of pressure drop in fluid flow and maximization of heat dissipation effects, aiming small scale applications. The TOM is applied to a domain, to obtain an optimized channel topology, according to a given multi-objective function that combines pressure drop minimization and heat transfer maximization. Stokes flow hypothesis is adopted. Moreover, both conduction and forced convection effects are included in the steady-state heat transfer model. The topology optimization procedure combines the Finite Element Method (to carry out the physical analysis) with Sequential Linear Programming (as the optimization algorithm). Two-dimensional topology optimization results of channel layouts obtained for a heat sink design are presented as example to illustrate the design methodology. 3D computational simulations and prototype manufacturing have been carried out to validate the proposed design methodology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Peer-to-Peer network paradigm is drawing the attention of both final users and researchers for its features. P2P networks shift from the classic client-server approach to a high level of decentralization where there is no central control and all the nodes should be able not only to require services, but to provide them to other peers as well. While on one hand such high level of decentralization might lead to interesting properties like scalability and fault tolerance, on the other hand it implies many new problems to deal with. A key feature of many P2P systems is openness, meaning that everybody is potentially able to join a network with no need for subscription or payment systems. The combination of openness and lack of central control makes it feasible for a user to free-ride, that is to increase its own benefit by using services without allocating resources to satisfy other peers’ requests. One of the main goals when designing a P2P system is therefore to achieve cooperation between users. Given the nature of P2P systems based on simple local interactions of many peers having partial knowledge of the whole system, an interesting way to achieve desired properties on a system scale might consist in obtaining them as emergent properties of the many interactions occurring at local node level. Two methods are typically used to face the problem of cooperation in P2P networks: 1) engineering emergent properties when designing the protocol; 2) study the system as a game and apply Game Theory techniques, especially to find Nash Equilibria in the game and to reach them making the system stable against possible deviant behaviors. In this work we present an evolutionary framework to enforce cooperative behaviour in P2P networks that is alternative to both the methods mentioned above. Our approach is based on an evolutionary algorithm inspired by computational sociology and evolutionary game theory, consisting in having each peer periodically trying to copy another peer which is performing better. The proposed algorithms, called SLAC and SLACER, draw inspiration from tag systems originated in computational sociology, the main idea behind the algorithm consists in having low performance nodes copying high performance ones. The algorithm is run locally by every node and leads to an evolution of the network both from the topology and from the nodes’ strategy point of view. Initial tests with a simple Prisoners’ Dilemma application show how SLAC is able to bring the network to a state of high cooperation independently from the initial network conditions. Interesting results are obtained when studying the effect of cheating nodes on SLAC algorithm. In fact in some cases selfish nodes rationally exploiting the system for their own benefit can actually improve system performance from the cooperation formation point of view. The final step is to apply our results to more realistic scenarios. We put our efforts in studying and improving the BitTorrent protocol. BitTorrent was chosen not only for its popularity but because it has many points in common with SLAC and SLACER algorithms, ranging from the game theoretical inspiration (tit-for-tat-like mechanism) to the swarms topology. We discovered fairness, meant as ratio between uploaded and downloaded data, to be a weakness of the original BitTorrent protocol and we drew inspiration from the knowledge of cooperation formation and maintenance mechanism derived from the development and analysis of SLAC and SLACER, to improve fairness and tackle freeriding and cheating in BitTorrent. We produced an extension of BitTorrent called BitFair that has been evaluated through simulation and has shown the abilities of enforcing fairness and tackling free-riding and cheating nodes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this Doctoral Thesis is to develop a genetic algorithm based optimization methods to find the best conceptual design architecture of an aero-piston-engine, for given design specifications. Nowadays, the conceptual design of turbine airplanes starts with the aircraft specifications, then the most suited turbofan or turbo propeller for the specific application is chosen. In the aeronautical piston engines field, which has been dormant for several decades, as interest shifted towards turboaircraft, new materials with increased performance and properties have opened new possibilities for development. Moreover, the engine’s modularity given by the cylinder unit, makes it possible to design a specific engine for a given application. In many real engineering problems the amount of design variables may be very high, characterized by several non-linearities needed to describe the behaviour of the phenomena. In this case the objective function has many local extremes, but the designer is usually interested in the global one. The stochastic and the evolutionary optimization techniques, such as the genetic algorithms method, may offer reliable solutions to the design problems, within acceptable computational time. The optimization algorithm developed here can be employed in the first phase of the preliminary project of an aeronautical piston engine design. It’s a mono-objective genetic algorithm, which, starting from the given design specifications, finds the engine propulsive system configuration which possesses minimum mass while satisfying the geometrical, structural and performance constraints. The algorithm reads the project specifications as input data, namely the maximum values of crankshaft and propeller shaft speed and the maximal pressure value in the combustion chamber. The design variables bounds, that describe the solution domain from the geometrical point of view, are introduced too. In the Matlab® Optimization environment the objective function to be minimized is defined as the sum of the masses of the engine propulsive components. Each individual that is generated by the genetic algorithm is the assembly of the flywheel, the vibration damper and so many pistons, connecting rods, cranks, as the number of the cylinders. The fitness is evaluated for each individual of the population, then the rules of the genetic operators are applied, such as reproduction, mutation, selection, crossover. In the reproduction step the elitist method is applied, in order to save the fittest individuals from a contingent mutation and recombination disruption, making it undamaged survive until the next generation. Finally, as the best individual is found, the optimal dimensions values of the components are saved to an Excel® file, in order to build a CAD-automatic-3D-model for each component of the propulsive system, having a direct pre-visualization of the final product, still in the engine’s preliminary project design phase. With the purpose of showing the performance of the algorithm and validating this optimization method, an actual engine is taken, as a case study: it’s the 1900 JTD Fiat Avio, 4 cylinders, 4T, Diesel. Many verifications are made on the mechanical components of the engine, in order to test their feasibility and to decide their survival through generations. A system of inequalities is used to describe the non-linear relations between the design variables, and is used for components checking for static and dynamic loads configurations. The design variables geometrical boundaries are taken from actual engines data and similar design cases. Among the many simulations run for algorithm testing, twelve of them have been chosen as representative of the distribution of the individuals. Then, as an example, for each simulation, the corresponding 3D models of the crankshaft and the connecting rod, have been automatically built. In spite of morphological differences among the component the mass is almost the same. The results show a significant mass reduction (almost 20% for the crankshaft) in comparison to the original configuration, and an acceptable robustness of the method have been shown. The algorithm here developed is shown to be a valid method for an aeronautical-piston-engine preliminary project design optimization. In particular the procedure is able to analyze quite a wide range of design solutions, rejecting the ones that cannot fulfill the feasibility design specifications. This optimization algorithm could increase the aeronautical-piston-engine development, speeding up the production rate and joining modern computation performances and technological awareness to the long lasting traditional design experiences.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Optical coherence tomography (OCT) is a well-established image modality in ophthalmology and used daily in the clinic. Automatic evaluation of such datasets requires an accurate segmentation of the retinal cell layers. However, due to the naturally low signal to noise ratio and the resulting bad image quality, this task remains challenging. We propose an automatic graph-based multi-surface segmentation algorithm that internally uses soft constraints to add prior information from a learned model. This improves the accuracy of the segmentation and increase the robustness to noise. Furthermore, we show that the graph size can be greatly reduced by applying a smart segmentation scheme. This allows the segmentation to be computed in seconds instead of minutes, without deteriorating the segmentation accuracy, making it ideal for a clinical setup. An extensive evaluation on 20 OCT datasets of healthy eyes was performed and showed a mean unsigned segmentation error of 3.05 ±0.54 μm over all datasets when compared to the average observer, which is lower than the inter-observer variability. Similar performance was measured for the task of drusen segmentation, demonstrating the usefulness of using soft constraints as a tool to deal with pathologies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Cloud Computing enables provisioning and distribution of highly scalable services in a reliable, on-demand and sustainable manner. However, objectives of managing enterprise distributed applications in cloud environments under Service Level Agreement (SLA) constraints lead to challenges for maintaining optimal resource control. Furthermore, conflicting objectives in management of cloud infrastructure and distributed applications might lead to violations of SLAs and inefficient use of hardware and software resources. This dissertation focusses on how SLAs can be used as an input to the cloud management system, increasing the efficiency of allocating resources, as well as that of infrastructure scaling. First, we present an extended SLA semantic model for modelling complex service-dependencies in distributed applications, and for enabling automated cloud infrastructure management operations. Second, we describe a multi-objective VM allocation algorithm for optimised resource allocation in infrastructure clouds. Third, we describe a method of discovering relations between the performance indicators of services belonging to distributed applications and then using these relations for building scaling rules that a CMS can use for automated management of VMs. Fourth, we introduce two novel VM-scaling algorithms, which optimally scale systems composed of VMs, based on given SLA performance constraints. All presented research works were implemented and tested using enterprise distributed applications.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Advancements in cloud computing have enabled the proliferation of distributed applications, which require management and control of multiple services. However, without an efficient mechanism for scaling services in response to changing workload conditions, such as number of connected users, application performance might suffer, leading to violations of Service Level Agreements (SLA) and possible inefficient use of hardware resources. Combining dynamic application requirements with the increased use of virtualised computing resources creates a challenging resource Management context for application and cloud-infrastructure owners. In such complex environments, business entities use SLAs as a means for specifying quantitative and qualitative requirements of services. There are several challenges in running distributed enterprise applications in cloud environments, ranging from the instantiation of service VMs in the correct order using an adequate quantity of computing resources, to adapting the number of running services in response to varying external loads, such as number of users. The application owner is interested in finding the optimum amount of computing and network resources to use for ensuring that the performance requirements of all her/his applications are met. She/he is also interested in appropriately scaling the distributed services so that application performance guarantees are maintained even under dynamic workload conditions. Similarly, the infrastructure Providers are interested in optimally provisioning the virtual resources onto the available physical infrastructure so that her/his operational costs are minimized, while maximizing the performance of tenants’ applications. Motivated by the complexities associated with the management and scaling of distributed applications, while satisfying multiple objectives (related to both consumers and providers of cloud resources), this thesis proposes a cloud resource management platform able to dynamically provision and coordinate the various lifecycle actions on both virtual and physical cloud resources using semantically enriched SLAs. The system focuses on dynamic sizing (scaling) of virtual infrastructures composed of virtual machines (VM) bounded application services. We describe several algorithms for adapting the number of VMs allocated to the distributed application in response to changing workload conditions, based on SLA-defined performance guarantees. We also present a framework for dynamic composition of scaling rules for distributed service, which used benchmark-generated application Monitoring traces. We show how these scaling rules can be combined and included into semantic SLAs for controlling allocation of services. We also provide a detailed description of the multi-objective infrastructure resource allocation problem and various approaches to satisfying this problem. We present a resource management system based on a genetic algorithm, which performs allocation of virtual resources, while considering the optimization of multiple criteria. We prove that our approach significantly outperforms reactive VM-scaling algorithms as well as heuristic-based VM-allocation approaches.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Metamodels have proven be very useful when it comes to reducing the computational requirements of Evolutionary Algorithm-based optimization by acting as quick-solving surrogates for slow-solving fitness functions. The relationship between metamodel scope and objective function varies between applications, that is, in some cases the metamodel acts as a surrogate for the whole fitness function, whereas in other cases it replaces only a component of the fitness function. This paper presents a formalized qualitative process to evaluate a fitness function to determine the most suitable metamodel scope so as to increase the likelihood of calibrating a high-fidelity metamodel and hence obtain good optimization results in a reasonable amount of time. The process is applied to the risk-based optimization of water distribution systems; a very computationally-intensive problem for real-world systems. The process is validated with a simple case study (modified New York Tunnels) and the power of metamodelling is demonstrated on a real-world case study (Pacific City) with a computational speed-up of several orders of magnitude.