893 resultados para Multi objective evolutionary algorithms


Relevância:

100.00% 100.00%

Publicador:

Resumo:

El objetivo principal de esta tesis doctoral es profundizar en el análisis y diseño de un sistema inteligente para la predicción y control del acabado superficial en un proceso de fresado a alta velocidad, basado fundamentalmente en clasificadores Bayesianos, con el prop´osito de desarrollar una metodolog´ıa que facilite el diseño de este tipo de sistemas. El sistema, cuyo propósito es posibilitar la predicción y control de la rugosidad superficial, se compone de un modelo aprendido a partir de datos experimentales con redes Bayesianas, que ayudar´a a comprender los procesos dinámicos involucrados en el mecanizado y las interacciones entre las variables relevantes. Dado que las redes neuronales artificiales son modelos ampliamente utilizados en procesos de corte de materiales, también se incluye un modelo para fresado usándolas, donde se introdujo la geometría y la dureza del material como variables novedosas hasta ahora no estudiadas en este contexto. Por lo tanto, una importante contribución en esta tesis son estos dos modelos para la predicción de la rugosidad superficial, que se comparan con respecto a diferentes aspectos: la influencia de las nuevas variables, los indicadores de evaluación del desempeño, interpretabilidad. Uno de los principales problemas en la modelización con clasificadores Bayesianos es la comprensión de las enormes tablas de probabilidad a posteriori producidas. Introducimos un m´etodo de explicación que genera un conjunto de reglas obtenidas de árboles de decisión. Estos árboles son inducidos a partir de un conjunto de datos simulados generados de las probabilidades a posteriori de la variable clase, calculadas con la red Bayesiana aprendida a partir de un conjunto de datos de entrenamiento. Por último, contribuimos en el campo multiobjetivo en el caso de que algunos de los objetivos no se puedan cuantificar en números reales, sino como funciones en intervalo de valores. Esto ocurre a menudo en aplicaciones de aprendizaje automático, especialmente las basadas en clasificación supervisada. En concreto, se extienden las ideas de dominancia y frontera de Pareto a esta situación. Su aplicación a los estudios de predicción de la rugosidad superficial en el caso de maximizar al mismo tiempo la sensibilidad y la especificidad del clasificador inducido de la red Bayesiana, y no solo maximizar la tasa de clasificación correcta. Los intervalos de estos dos objetivos provienen de un m´etodo de estimación honesta de ambos objetivos, como e.g. validación cruzada en k rodajas o bootstrap.---ABSTRACT---The main objective of this PhD Thesis is to go more deeply into the analysis and design of an intelligent system for surface roughness prediction and control in the end-milling machining process, based fundamentally on Bayesian network classifiers, with the aim of developing a methodology that makes easier the design of this type of systems. The system, whose purpose is to make possible the surface roughness prediction and control, consists of a model learnt from experimental data with the aid of Bayesian networks, that will help to understand the dynamic processes involved in the machining and the interactions among the relevant variables. Since artificial neural networks are models widely used in material cutting proceses, we include also an end-milling model using them, where the geometry and hardness of the piecework are introduced as novel variables not studied so far within this context. Thus, an important contribution in this thesis is these two models for surface roughness prediction, that are then compared with respecto to different aspects: influence of the new variables, performance evaluation metrics, interpretability. One of the main problems with Bayesian classifier-based modelling is the understanding of the enormous posterior probabilitiy tables produced. We introduce an explanation method that generates a set of rules obtained from decision trees. Such trees are induced from a simulated data set generated from the posterior probabilities of the class variable, calculated with the Bayesian network learned from a training data set. Finally, we contribute in the multi-objective field in the case that some of the objectives cannot be quantified as real numbers but as interval-valued functions. This often occurs in machine learning applications, especially those based on supervised classification. Specifically, the dominance and Pareto front ideas are extended to this setting. Its application to the surface roughness prediction studies the case of maximizing simultaneously the sensitivity and specificity of the induced Bayesian network classifier, rather than only maximizing the correct classification rate. Intervals in these two objectives come from a honest estimation method of both objectives, like e.g. k-fold cross-validation or bootstrap.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Hoy en día, el refuerzo y reparación de estructuras de hormigón armado mediante el pegado de bandas de polímeros reforzados con fibras (FRP) se emplea cada vez con más frecuencia a causa de sus numerosas ventajas. Sin embargo, las vigas reforzadas con esta técnica pueden experimentar un modo de fallo frágil a causa del despegue repentino de la banda de FRP a partir de una fisura intermedia. A pesar de su importancia, el número de trabajos que abordan el estudio de este mecanismo de fallo y su monitorización es muy limitado. Por ello, el desarrollo de metodologías capaces de monitorizar a largo plazo la adherencia de este refuerzo a las estructuras de hormigón e identificar cuándo se inicia el despegue de la banda constituyen un importante desafío a abordar. El principal objetivo de esta tesis es la implementación de una metodología fiable y efectiva, capaz de detectar el despegue de una banda de FRP en una viga de hormigón armado a partir de una fisura intermedia. Para alcanzar este objetivo se ha implementado un procedimiento de calibración numérica a partir de ensayos experimentales. Para ello, en primer lugar, se ha desarrollado un modelo numérico unidimensional simple y no costoso representativo del comportamiento de este tipo vigas de hormigón reforzadas con FRP, basado en un modelo de fisura discreta para el hormigón y el método de elementos espectrales. La formación progresiva de fisuras a flexion y el consiguiente despegue en la interface entre el hormigón y el FRP se formulan mediante la introducción de un nuevo elemento capaz de representar ambos fenómenos simultáneamente sin afectar al procedimiento numérico. Además, con el modelo propuesto, se puede obtener de una forma sencilla la respuesta dinámica en altas frecuencias de este tipo de estructuras, lo cual puede hacer muy útil su uso como herramienta de diagnosis y detección del despegue en su fase inicial mediante una monitorización de la variación de las características dinámicas locales de la estructura. Un método de evaluación no destructivo muy prometedor para la monitorización local de las estructuras es el método de la impedancia usando sensores-actuadores piezoeléctricos (PZT). La impedancia eléctrica de los sensores PZT se puede relacionar con la impedancia mecánica de las estructuras donde se encuentran adheridos Ya que la impedancia mecánica de una estructura se verá afectada por su deterioro, se pueden implementar indicadores de daño mediante una comparación del espectro de admitancia (inversa de la impedancia) a lo largo de distintas etapas durante el periodo de servicio de una estructura. Cualquier cambio en el espectro se podría interpretar como una variación en la integridad de la estructura. La impedancia eléctrica se mide a altas frecuencias con lo cual esta metodología debería ser muy sensible a la detección de estados de daño incipiente local, tal como se desea en la aplicación de este trabajo. Se ha implementado un elemento espectral PZT-FRP como extensión del modelo previamente desarrollado, con el objetivo de poder calcular numéricamente la impedancia eléctrica de sensores PZT adheridos a bandas de FRP sobre una viga de hormigón armado. El modelo, combinado con medidas experimentales captadas mediante sensores PZT, se implementa en el marco de una metodología de calibración de modelos para detectar cuantitativamente el despegue en la interfase entre una banda de FRP y una viga de hormigón. El procedimiento de optimización se resuelve empleando el método del enjambre cooperativo con un algoritmo bagging. Los resultados muestran una gran aproximación en la estimación del daño para el problema propuesto. Adicionalmente, se ha desarrollado también un método adaptativo para el mallado de elementos espectrales con el objetivo de localizar las zonas dañadas a partir de los resultados experimentales, el cual contribuye a aumentar la robustez y efectividad del método propuesto a la hora de identificar daños incipientes en su aparición inicial. Finalmente, se ha llevado a cabo un procedimiento de optimización multi-objetivo para detectar el despegue inicial en una viga de hormigón a escala real reforzada con FRP a partir de las impedancias captadas con una red de sensores PZT instrumentada a lo largo de la longitud de la viga. Cada sensor aporta los datos para definir cada una de las funciones objetivo que definen el procedimiento. Combinando el modelo previo de elementos espectrales con un algoritmo PSO multi-objetivo el procedimiento de detección de daño resultante proporciona resultados satisfactorios considerando la escala de la estructura y todas las incertidumbres características ligadas a este proceso. Los resultados obtenidos prueban la viabilidad y capacidad de los métodos antes mencionados y también su potencial en aplicaciones reales. Abstract Nowadays, the external bonding of fibre reinforced polymer (FRP) plates or sheets is increasingly used for the strengthening and retrofitting of reinforced concrete (RC) structures due to its numerous advantages. However, this kind of strengthening often leads to brittle failure modes being the most dominant failure mode the debonding induced by an intermediate crack (IC). In spite of its importance, the number of studies regarding the IC debonding mechanism and bond health monitoring is very limited. Methodologies able to monitor the long-term efficiency of bonding and successfully identify the initiation of FRP debonding constitute a challenge to be met. The main purpose of this thesisis the implementation of a reliable and effective methodology of damage identification able to detect intermediate crack debonding in FRP-strengthened RC beams. To achieve this goal, a model updating procedure based on numerical simulations and experimental tests has been implemented. For it, firstly, a simple and non-expensive one-dimensional model based on the discrete crack approach for concrete and the spectral element method has been developed. The progressive formation of flexural cracks and subsequent concrete-FRP interfacial debonding is formulated by the introduction of a new element able to represent both phenomena simultaneously without perturbing the numerical procedure. Furthermore, with the proposed model, high frequency dynamic response for these kinds of structures can also be obtained in a very simple and non-expensive way, which makes this procedure very useful as a tool for diagnoses and detection of debonding in its initial stage by monitoring the change in local dynamic characteristics. One very promising active non-destructive evaluation method for local monitoring is impedance-based structural health monitoring(SHM)using piezoelectric ceramic (PZT) sensor-actuators. The electrical impedance of the PZT can be directly related to the mechanical impedance of the host structural component where the PZT transducers are attached. Since the structural mechanical impedance will be affected by the presence of structural damage, comparisons of admittance (inverse of impedance) spectra at various times during the service period of the structure can be used as damage indicator. Any change in the spectra might be an indication of a change in the structural integrity. The electrical impedance is measured at high frequencies with which this methodology appears to be very sensitive to incipient damage in structural systems as desired for our application. Abonded-PZT-FRP spectral beam element approach based on an extension of the previous discrete crack approach is implemented in the calculation of the electrical impedance of the PZT transducer bonded to the FRP plates of a RC beam. This approach in conjunction with the experimental measurements of PZT actuator-sensors mounted on the structure is used to present an updating methodology to quantitatively detect interfacial debonding between a FRP strip and the host RC structure. The updating procedure is solved by using an ensemble particle swarm optimization approach with abagging algorithm, and the results demonstrate a big improvement for the performance and accuracy of the damage detection in the proposed problem. Additionally, an adaptive strategy of spectral element mesh has been also developed to detect damage location with experimental results, which shows the robustness and effectiveness of the proposed method to identify initial and incipient damages at its early stage. Lastly, multi-objective optimization has been carried out to detect debonding damage in a real scale FRP-strengthened RC beam by using impedance signatures. A net of PZT sensors is distributed along the beam to construct impedance-based multiple objectives under gradually induced damage scenario. By combining the spectral element model presented previously and an ensemble multi-objective PSO algorithm, the implemented damage detection process yields satisfactory predictions considering the scale and uncertainties of the structure. The obtained results prove the feasibility and capability of the aforementioned methods and also their potentials in real engineering applications.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El principal objetivo de esta tesis es el desarrollo de métodos de síntesis de diagramas de radiación de agrupaciones de antenas, en donde se realiza una caracterización electromagnética rigurosa de los elementos radiantes y de los acoplos mutuos existentes. Esta caracterización no se realiza habitualmente en la gran mayoría de métodos de síntesis encontrados en la literatura, debido fundamentalmente a dos razones. Por un lado, se considera que el diagrama de radiación de un array de antenas se puede aproximar con el factor de array que únicamente tiene en cuenta la posición de los elementos y las excitaciones aplicadas a los mismos. Sin embargo, como se mostrará en esta tesis, en múltiples ocasiones un riguroso análisis de los elementos radiantes y del acoplo mutuo entre ellos es importante ya que los resultados obtenidos pueden ser notablemente diferentes. Por otro lado, no es sencillo combinar un método de análisis electromagnético con un proceso de síntesis de diagramas de radiación. Los métodos de análisis de agrupaciones de antenas suelen ser costosos computacionalmente, ya que son estructuras grandes en términos de longitudes de onda. Generalmente, un diseño de un problema electromagnético suele comprender varios análisis de la estructura, dependiendo de las variaciones de las características, lo que hace este proceso muy costoso. Dos métodos se utilizan en esta tesis para el análisis de los arrays acoplados. Ambos están basados en el método de los elementos finitos, la descomposición de dominio y el análisis modal para analizar la estructura radiante y han sido desarrollados en el grupo de investigación donde se engloba esta tesis. El primero de ellos es una técnica de análisis de arrays finitos basado en la aproximación de array infinito. Su uso es indicado para arrays planos de grandes dimensiones con elementos equiespaciados. El segundo caracteriza el array y el acoplo mutuo entre elementos a partir de una expansión en modos esféricos del campo radiado por cada uno de los elementos. Este método calcula los acoplos entre los diferentes elementos del array usando las propiedades de traslación y rotación de los modos esféricos. Es capaz de analizar agrupaciones de elementos distribuidos de forma arbitraria. Ambas técnicas utilizan una formulación matricial que caracteriza de forma rigurosa el campo radiado por el array. Esto las hace muy apropiadas para su posterior uso en una herramienta de diseño, como los métodos de síntesis desarrollados en esta tesis. Los resultados obtenidos por estas técnicas de síntesis, que incluyen métodos rigurosos de análisis, son consecuentemente más precisos. La síntesis de arrays consiste en modificar uno o varios parámetros de las agrupaciones de antenas buscando unas determinadas especificaciones de las características de radiación. Los parámetros utilizados como variables de optimización pueden ser varios. Los más utilizados son las excitaciones aplicadas a los elementos, pero también es posible modificar otros parámetros de diseño como son las posiciones de los elementos o las rotaciones de estos. Los objetivos de las síntesis pueden ser dirigir el haz o haces en una determinada dirección o conformar el haz con formas arbitrarias. Además, es posible minimizar el nivel de los lóbulos secundarios o del rizado en las regiones deseadas, imponer nulos que evitan posibles interferencias o reducir el nivel de la componente contrapolar. El método para el análisis de arrays finitos basado en la aproximación de array infinito considera un array finito como un array infinito con un número finito de elementos excitados. Los elementos no excitados están físicamente presentes y pueden presentar tres diferentes terminaciones, corto-circuito, circuito abierto y adaptados. Cada una de estas terminaciones simulará mejor el entorno real en el que el array se encuentre. Este método de análisis se integra en la tesis con dos métodos diferentes de síntesis de diagramas de radiación. En el primero de ellos se presenta un método basado en programación lineal en donde es posible dirigir el haz o haces, en la dirección deseada, además de ejercer un control sobre los lóbulos secundarios o imponer nulos. Este método es muy eficiente y obtiene soluciones óptimas. El mismo método de análisis es también aplicado a un método de conformación de haz, en donde un problema originalmente no convexo (y de difícil solución) es transformado en un problema convexo imponiendo restricciones de simetría, resolviendo de este modo eficientemente un problema complejo. Con este método es posible diseñar diagramas de radiación con haces de forma arbitraria, ejerciendo un control en el rizado del lóbulo principal, así como en el nivel de los lóbulos secundarios. El método de análisis de arrays basado en la expansión en modos esféricos se integra en la tesis con tres técnicas de síntesis de diagramas de radiación. Se propone inicialmente una síntesis de conformación del haz basado en el método de la recuperación de fase resuelta de forma iterativa mediante métodos convexos, en donde relajando las restricciones del problema original se consiguen unas soluciones cercanas a las óptimas de manera eficiente. Dos métodos de síntesis se han propuesto, donde las variables de optimización son las posiciones y las rotaciones de los elementos respectivamente. Se define una función de coste basada en la intensidad de radiación, la cual es minimizada de forma iterativa con el método del gradiente. Ambos métodos reducen el nivel de los lóbulos secundarios minimizando una función de coste. El gradiente de la función de coste es obtenido en términos de la variable de optimización en cada método. Esta función de coste está formada por la expresión rigurosa de la intensidad de radiación y por una función de peso definida por el usuario para imponer prioridades sobre las diferentes regiones de radiación, si así se desea. Por último, se presenta un método en el cual, mediante técnicas de programación entera, se buscan las fases discretas que generan un diagrama de radiación lo más cercano posible al deseado. Con este método se obtienen diseños que minimizan el coste de fabricación. En cada uno de las diferentes técnicas propuestas en la tesis, se presentan resultados con elementos reales que muestran las capacidades y posibilidades que los métodos ofrecen. Se comparan los resultados con otros métodos disponibles en la literatura. Se muestra la importancia de tener en cuenta los diagramas de los elementos reales y los acoplos mutuos en el proceso de síntesis y se comparan los resultados obtenidos con herramientas de software comerciales. ABSTRACT The main objective of this thesis is the development of optimization methods for the radiation pattern synthesis of array antennas in which a rigorous electromagnetic characterization of the radiators and the mutual coupling between them is performed. The electromagnetic characterization is usually overlooked in most of the available synthesis methods in the literature, this is mainly due to two reasons. On the one hand, it is argued that the radiation pattern of an array is mainly influenced by the array factor and that the mutual coupling plays a minor role. As it is shown in this thesis, the mutual coupling and the rigorous characterization of the array antenna influences significantly in the array performance and its computation leads to differences in the results obtained. On the other hand, it is difficult to introduce an analysis procedure into a synthesis technique. The analysis of array antennas is generally expensive computationally as the structure to analyze is large in terms of wavelengths. A synthesis method requires to carry out a large number of analysis, this makes the synthesis problem very expensive computationally or intractable in some cases. Two methods have been used in this thesis for the analysis of coupled antenna arrays, both of them have been developed in the research group in which this thesis is involved. They are based on the finite element method (FEM), the domain decomposition and the modal analysis. The first one obtains a finite array characterization with the results obtained from the infinite array approach. It is specially indicated for the analysis of large arrays with equispaced elements. The second one characterizes the array elements and the mutual coupling between them with a spherical wave expansion of the radiated field by each element. The mutual coupling is computed using the properties of translation and rotation of spherical waves. This method is able to analyze arrays with elements placed on an arbitrary distribution. Both techniques provide a matrix formulation that makes them very suitable for being integrated in synthesis techniques, the results obtained from these synthesis methods will be very accurate. The array synthesis stands for the modification of one or several array parameters looking for some desired specifications of the radiation pattern. The array parameters used as optimization variables are usually the excitation weights applied to the array elements, but some other array characteristics can be used as well, such as the array elements positions or rotations. The desired specifications may be to steer the beam towards any specific direction or to generate shaped beams with arbitrary geometry. Further characteristics can be handled as well, such as minimize the side lobe level in some other radiating regions, to minimize the ripple of the shaped beam, to take control over the cross-polar component or to impose nulls on the radiation pattern to avoid possible interferences from specific directions. The analysis method based on the infinite array approach considers an infinite array with a finite number of excited elements. The infinite non-excited elements are physically present and may have three different terminations, short-circuit, open circuit and match terminated. Each of this terminations is a better simulation for the real environment of the array. This method is used in this thesis for the development of two synthesis methods. In the first one, a multi-objective radiation pattern synthesis is presented, in which it is possible to steer the beam or beams in desired directions, minimizing the side lobe level and with the possibility of imposing nulls in the radiation pattern. This method is very efficient and obtains optimal solutions as it is based on convex programming. The same analysis method is used in a shaped beam technique in which an originally non-convex problem is transformed into a convex one applying symmetry restrictions, thus solving a complex problem in an efficient way. This method allows the synthesis of shaped beam radiation patterns controlling the ripple in the mainlobe and the side lobe level. The analysis method based on the spherical wave expansion is applied for different synthesis techniques of the radiation pattern of coupled arrays. A shaped beam synthesis is presented, in which a convex formulation is proposed based on the phase retrieval method. In this technique, an originally non-convex problem is solved using a relaxation and solving a convex problems iteratively. Two methods are proposed based on the gradient method. A cost function is defined involving the radiation intensity of the coupled array and a weighting function that provides more degrees of freedom to the designer. The gradient of the cost function is computed with respect to the positions in one of them and the rotations of the elements in the second one. The elements are moved or rotated iteratively following the results of the gradient. A highly non-convex problem is solved very efficiently, obtaining very good results that are dependent on the starting point. Finally, an optimization method is presented where discrete digital phases are synthesized providing a radiation pattern as close as possible to the desired one. The problem is solved using linear integer programming procedures obtaining array designs that greatly reduce the fabrication costs. Results are provided for every method showing the capabilities that the above mentioned methods offer. The results obtained are compared with available methods in the literature. The importance of introducing a rigorous analysis into the synthesis method is emphasized and the results obtained are compared with a commercial software, showing good agreement.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Los regímenes fiscales que se aplican a los contratos de exploración y desarrollo de petróleo y gas, entre los propietarios del recurso natural (generalmente el país soberano representado por su gobierno) y las compañías operadoras internacionales (COI) que aportan capital, experiencia y tecnología, no han sabido responder a la reciente escalada de los precios del crudo y han dado lugar a que los países productores no estén recibiendo la parte de renta correspondiente al incremento de precios. Esto ha provocado una ola de renegociaciones llegándose incluso a la imposición unilateral de nuevos términos por parte de algunos gobiernos entre los que destacan el caso de Venezuela y Argentina, por ser los más radicales. El objetivo del presente trabajo es el estudio y diseño de un régimen fiscal que, en las actuales condiciones del mercado, consiga que los gobiernos optimicen sus ingresos incentivando la inversión. Para ello se simulan los efectos de siete tipos diferentes de fiscalidades aplicadas a dos yacimientos de características muy distintas y se valoran los resultados. El modelo utilizado para la simulación es el modelo de escenarios, ampliamente utilizado tanto por la comunidad académica como por la industria para comparar el comportamiento de diferentes regímenes fiscales. Para decidir cuál de las fiscalidades estudiadas es la mejor se emplea un método optimización multicriterio. Los criterios que se han aplicado para valorar los resultados recogen la opinión de expertos de la industria sobre qué factores se consideran deseables en un contrato a la hora invertir. El resultado permite delinear las características de un marco fiscal ideal del tipo acuerdo de producción compartida, sin royalties, con un límite alto de recuperación de crudo coste que permita recobrar todos los costes operativos y una parte de los de capital en cualquier escenario de precios, un reparto de los beneficios en función de un indicador de rentabilidad como es la TIR, con un mecanismo de recuperación de costes adicional (uplift) que incentive la inversión y con disposiciones que premien la exploración y más la de alto riesgo como la amortización acelerada de los gastos de capital o una ampliación de la cláusula de ringfence. Un contrato con estas características permitirá al gobierno optimizar los ingresos obtenidos de sus reservas de petróleo y gas maximizando la producción al atraer inversión para la exploración y mejorar la recuperación alargando la vida del yacimiento. Además al reducir el riesgo percibido por el inversor que recupera sus costes, menor será la rentabilidad exigida al capital invertido y por tanto mayor la parte de esos ingresos que irá a parar al gobierno del país productor. ABSTRACT Fiscal systems used in petroleum arrangements between the owners of the resource (usually a sovereign country represented by its government) and the international operating company (IOC) that provides capital, knowhow and technology, have failed to allocate profits from the recent escalation of oil prices and have resulted in producing countries not receiving the right share of that increase. This has caused a wave of renegotiations and even in some cases, like Venezuela and Argentina, government unilaterally imposed new terms. This paper aims to outline desirable features of a petroleum fiscal system, under current market conditions, for governments to maximize their revenues while encouraging investment. Firstly the impact of seven different types of fiscal regimes is studied with a simulation for two separate oil fields using the scenario approach. The scenario approach has been frequently employed by academic and business researchers to compare the performance of diverse fiscal regimes. In order to decide which of the fiscal regimes’ performance is best we used a multi-objective optimization decision making approach to assess the results. The criteria applied gather the preferences of a panel of industry experts about the desirable features of a contract when making investment decisions. The results show the characteristics of an ideal fiscal framework that closely resembles a production sharing contract, with no royalty payment and a high cost recovery limit that allows the IOC to recover all operating expenses and a share of its capital costs under any price scenario, a profit oil sharing mechanism based on a profitability indicator such as the ROR, with an uplift that allows to recover an additional percentage of capital costs and provisions that promote exploration investment, specially high-risk exploration, such as accelerated depreciation for capital costs and a wide definition of the ringfence clause. A contract with these features will allow governments to optimize overall revenues from its petroleum resources maximizing production by promoting investment on exploration and extending oil fields life. Also by reducing the investor’s perception of risk it will reduce the minimum return to capital required by the IOC and therefore it will increase the government share of those revenues.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Este trabalho é referente ao desenvolvimento de um calibrador multiobjetivo automático do modelo SWMM (Storm Water Management Model), e avaliação de algumas fontes de incertezas presentes no processo de calibração, visando à representação satisfatória da transformação chuva-vazão. O código foi escrito em linguagem C, e aplica os conceitos do método de otimização multiobjetivo NSGAII (Non Dominated Sorting Genetic Algorithm) com elitismo controlado, além de utilizar o código fonte do modelo SWMM para a determinação das vazões simuladas. Paralelamente, também foi criada uma interface visual, para melhorar a facilidade de utilização do calibrador. Os testes do calibrador foram aplicados a três sistemas diferentes: um sistema hipotético disponibilizado no pacote de instalação do SWMM; um sistema real de pequenas dimensões, denominado La Terraza, localizado no município de Sierra Vista, Arizona (EUA); e um sistema de maiores dimensões, a bacia hidrográfica do Córrego do Gregório, localizada no município de São Carlos (SP). Os resultados indicam que o calibrador construído apresenta, em geral, eficiência satisfatória, porém é bastante dependente da qualidade dos dados observados em campo e dos parâmetros de entrada escolhidos pelo usuário. Foi demonstrada a importância da escolha dos eventos utilizados na calibração, do estabelecimento de limites adequados nos valores das variáveis de decisão, da escolha das funções objetivo e, principalmente, da qualidade e representatividade dos dados de monitoramento pluvio e fluviométrico. Conclui-se que estes testes desenvolvidos contribuem para o entendimento mais aprofundado dos processos envolvidos na modelagem e calibração, possibilitando avanços na confiabilidade dos resultados da modelagem.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This work addresses the optimization of ammonia–water absorption cycles for cooling and refrigeration applications with economic and environmental concerns. Our approach combines the capabilities of process simulation, multi-objective optimization (MOO), cost analysis and life cycle assessment (LCA). The optimization task is posed in mathematical terms as a multi-objective mixed-integer nonlinear program (moMINLP) that seeks to minimize the total annualized cost and environmental impact of the cycle. This moMINLP is solved by an outer-approximation strategy that iterates between primal nonlinear programming (NLP) subproblems with fixed binaries and a tailored mixed-integer linear programming (MILP) model. The capabilities of our approach are illustrated through its application to an ammonia–water absorption cycle used in cooling and refrigeration applications.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The delineation of functional economic areas, or market areas, is a problem of high practical relevance, since the delineation of functional sets such as economic areas in the US, Travel-to-Work Areas in the United Kingdom, and their counterparts in other OECD countries are the basis of many statistical operations and policy making decisions at local level. This is a combinatorial optimisation problem defined as the partition of a given set of indivisible spatial units (covering a territory) into regions characterised by being (a) self-contained and (b) cohesive, in terms of spatial interaction data (flows, relationships). Usually, each region must reach a minimum size and self-containment level, and must be continuous. Although these optimisation problems have been typically solved through greedy methods, a recent strand of the literature in this field has been concerned with the use of evolutionary algorithms with ad hoc operators. Although these algorithms have proved to be successful in improving the results of some of the more widely applied official procedures, they are so time consuming that cannot be applied directly to solve real-world problems. In this paper we propose a new set of group-based mutation operators, featuring general operations over disjoint groups, tailored to ensure that all the constraints are respected during the operation to improve efficiency. A comparative analysis of our results with those from previous approaches shows that the proposed algorithm systematically improves them in terms of both quality and processing time, something of crucial relevance since it allows dealing with most large, real-world problems in reasonable time.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Poster presented in the 24th European Symposium on Computer Aided Process Engineering (ESCAPE 24), Budapest, Hungary, June 15-18, 2014.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work, we analyze the effect of incorporating life cycle inventory (LCI) uncertainty on the multi-objective optimization of chemical supply chains (SC) considering simultaneously their economic and environmental performance. To this end, we present a stochastic multi-scenario mixed-integer linear programming (MILP) coupled with a two-step transformation scenario generation algorithm with the unique feature of providing scenarios where the LCI random variables are correlated and each one of them has the desired lognormal marginal distribution. The environmental performance is quantified following life cycle assessment (LCA) principles, which are represented in the model formulation through standard algebraic equations. The capabilities of our approach are illustrated through a case study of a petrochemical supply chain. We show that the stochastic solution improves the economic performance of the SC in comparison with the deterministic one at any level of the environmental impact, and moreover the correlation among environmental burdens provides more realistic scenarios for the decision making process.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Convex vector (or multi-objective) semi-infinite optimization deals with the simultaneous minimization of finitely many convex scalar functions subject to infinitely many convex constraints. This paper provides characterizations of the weakly efficient, efficient and properly efficient points in terms of cones involving the data and Karush–Kuhn–Tucker conditions. The latter characterizations rely on different local and global constraint qualifications. The results in this paper generalize those obtained by the same authors on linear vector semi-infinite optimization problems.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

There are many models in the literature that have been proposed in the last decades aimed at assessing the reliability, availability and maintainability (RAM) of safety equipment, many of them with a focus on their use to assess the risk level of a technological system or to search for appropriate design and/or surveillance and maintenance policies in order to assure that an optimum level of RAM of safety systems is kept during all the plant operational life. This paper proposes a new approach for RAM modelling that accounts for equipment ageing and maintenance and testing effectiveness of equipment consisting of multiple items in an integrated manner. This model is then used to perform the simultaneous optimization of testing and maintenance for ageing equipment consisting of multiple items. An example of application is provided, which considers a simplified High Pressure Injection System (HPIS) of a typical Power Water Reactor (PWR). Basically, this system consists of motor driven pumps (MDP) and motor operated valves (MOV), where both types of components consists of two items each. These components present different failure and cause modes and behaviours, and they also undertake complex test and maintenance activities depending on the item involved. The results of the example of application demonstrate that the optimization algorithm provide the best solutions when the optimization problem is formulated and solved considering full flexibility in the implementation of testing and maintenance activities taking part of such an integrated RAM model.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The research literature on metalieuristic and evolutionary computation has proposed a large number of algorithms for the solution of challenging real-world optimization problems. It is often not possible to study theoretically the performance of these algorithms unless significant assumptions are made on either the algorithm itself or the problems to which it is applied, or both. As a consequence, metalieuristics are typically evaluated empirically using a set of test problems. Unfortunately, relatively little attention has been given to the development of methodologies and tools for the large-scale empirical evaluation and/or comparison of metaheuristics. In this paper, we propose a landscape (test-problem) generator that can be used to generate optimization problem instances for continuous, bound-constrained optimization problems. The landscape generator is parameterized by a small number of parameters, and the values of these parameters have a direct and intuitive interpretation in terms of the geometric features of the landscapes that they produce. An experimental space is defined over algorithms and problems, via a tuple of parameters for any specified algorithm and problem class (here determined by the landscape generator). An experiment is then clearly specified as a point in this space, in a way that is analogous to other areas of experimental algorithmics, and more generally in experimental design. Experimental results are presented, demonstrating the use of the landscape generator. In particular, we analyze some simple, continuous estimation of distribution algorithms, and gain new insights into the behavior of these algorithms using the landscape generator.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

As an alternative to traditional evolutionary algorithms (EAs), population-based incremental learning (PBIL) maintains a probabilistic model of the best individual(s). Originally, PBIL was applied in binary search spaces. Recently, some work has been done to extend it to continuous spaces. In this paper, we review two such extensions of PBIL. An improved version of the PBIL based on Gaussian model is proposed that combines two main features: a new updating rule that takes into account all the individuals and their fitness values and a self-adaptive learning rate parameter. Furthermore, a new continuous PBIL employing a histogram probabilistic model is proposed. Some experiments results are presented that highlight the features of the new algorithms.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Group decision making is the study of identifying and selecting alternatives based on the values and preferences of the decision maker. Making a decision implies that there are several alternative choices to be considered. This paper uses the concept of Data Envelopment Analysis to introduce a new mathematical method for selecting the best alternative in a group decision making environment. The introduced model is a multi-objective function which is converted into a multi-objective linear programming model from which the optimal solution is obtained. A numerical example shows how the new model can be applied to rank the alternatives or to choose a subset of the most promising alternatives.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper introduces a new mathematical method for improving the discrimination power of data envelopment analysis and to completely rank the efficient decision-making units (DMUs). Fuzzy concept is utilised. For this purpose, first all DMUs are evaluated with the CCR model. Thereafter, the resulted weights for each output are considered as fuzzy sets and are then converted to fuzzy numbers. The introduced model is a multi-objective linear model, endpoints of which are the highest and lowest of the weighted values. An added advantage of the model is its ability to handle the infeasibility situation sometimes faced by previously introduced models.