880 resultados para Fuzzy Multi-Objective Linear Programming


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Large-grain synchronous dataflow graphs or multi-rate graphs have the distinct feature that the nodes of the dataflow graph fire at different rates. Such multi-rate large-grain dataflow graphs have been widely regarded as a powerful programming model for DSP applications. In this paper we propose a method to minimize buffer storage requirement in constructing rate-optimal compile-time (MBRO) schedules for multi-rate dataflow graphs. We demonstrate that the constraints to minimize buffer storage while executing at the optimal computation rate (i.e. the maximum possible computation rate without storage constraints) can be formulated as a unified linear programming problem in our framework. A novel feature of our method is that in constructing the rate-optimal schedule, it directly minimizes the memory requirement by choosing the schedule time of nodes appropriately. Lastly, a new circular-arc interval graph coloring algorithm has been proposed to further reduce the memory requirement by allowing buffer sharing among the arcs of the multi-rate dataflow graph. We have constructed an experimental testbed which implements our MBRO scheduling algorithm as well as (i) the widely used periodic admissible parallel schedules (also known as block schedules) proposed by Lee and Messerschmitt (IEEE Transactions on Computers, vol. 36, no. 1, 1987, pp. 24-35), (ii) the optimal scheduling buffer allocation (OSBA) algorithm of Ning and Gao (Conference Record of the Twentieth Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages, Charleston, SC, Jan. 10-13, 1993, pp. 29-42), and (iii) the multi-rate software pipelining (MRSP) algorithm (Govindarajan and Gao, in Proceedings of the 1993 International Conference on Application Specific Array Processors, Venice, Italy, Oct. 25-27, 1993, pp. 77-88). Schedules generated for a number of random dataflow graphs and for a set of DSP application programs using the different scheduling methods are compared. The experimental results have demonstrated a significant improvement (10-20%) in buffer requirements for the MBRO schedules compared to the schedules generated by the other three methods, without sacrificing the computation rate. The MBRO method also gives a 20% average improvement in computation rate compared to Lee's Block scheduling method.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We know, from the classical work of Tarski on real closed fields, that elimination is, in principle, a fundamental engine for mechanized deduction. But, in practice, the high complexity of elimination algorithms has limited their use in the realization of mechanical theorem proving. We advocate qualitative theorem proving, where elimination is attractive since most processes of reasoning take place through the elimination of middle terms, and because the computational complexity of the proof is not an issue. Indeed what we need is the existence of the proof and not its mechanization. In this paper, we treat the linear case and illustrate the power of this paradigm by giving extremely simple proofs of two central theorems in the complexity and geometry of linear programming.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Bid optimization is now becoming quite popular in sponsored search auctions on the Web. Given a keyword and the maximum willingness to pay of each advertiser interested in the keyword, the bid optimizer generates a profile of bids for the advertisers with the objective of maximizing customer retention without compromising the revenue of the search engine. In this paper, we present a bid optimization algorithm that is based on a Nash bargaining model where the first player is the search engine and the second player is a virtual agent representing all the bidders. We make the realistic assumption that each bidder specifies a maximum willingness to pay values and a discrete, finite set of bid values. We show that the Nash bargaining solution for this problem always lies on a certain edge of the convex hull such that one end point of the edge is the vector of maximum willingness to pay of all the bidders. We show that the other endpoint of this edge can be computed as a solution of a linear programming problem. We also show how the solution can be transformed to a bid profile of the advertisers.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper obtains a new accurate model for sensitivity in power systems and uses it in conjunction with linear programming for the solution of load-shedding problems with a minimum loss of loads. For cases where the error in the sensitivity model increases, other linear programming and quadratic programming models have been developed, assuming currents at load buses as variables and not load powers. A weighted error criterion has been used to take priority schedule into account; it can be either a linear or a quadratic function of the errors, and depending upon the function appropriate programming techniques are to be employed.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Three algorithms for reactive power optimization are proposed in this paper with three different objective functions. The objectives in the proposed algorithm are to minimize the sum of the squares of the voltage deviations of the load buses, minimization of sum of squares of voltage stability L-indices of load buses (:3L2) algorithm, and also the objective of system real power loss (Ploss) minimization. The approach adopted is an iterative scheme with successive power flow analysis using decoupled technique and solution of the linear programming problem using upper bound optimization technique. Results obtained with all these objectives are compared. The analysis of these objective functions are presented to illustrate their advantages. It is observed comparing different objective functions it is possible to identify critical On Load Tap Changers (OLTCs) that should be made manual to avoid possible voltage instability due to their operation based on voltage improvement criteria under heavy load conditions. These algorithms have been tested under simulated conditions on few test systems. The results obtained on practical systems of 24-node equivalent EHV Indian power network, and for a 205 bus EHV system are presented for illustration purposes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Package-board co-design plays a crucial role in determining the performance of high-speed systems. Although there exist several commercial solutions for electromagnetic analysis and verification, lack of Computer Aided Design (CAD) tools for SI aware design and synthesis lead to longer design cycles and non-optimal package-board interconnect geometries. In this work, the functional similarities between package-board design and radio-frequency (RF) imaging are explored. Consequently, qualitative methods common to the imaging community, like Tikhonov Regularization (TR) and Landweber method are applied to solve multi-objective, multi-variable package design problems. In addition, a new hierarchical iterative piecewise linear algorithm is developed as a wrapper over LBP for an efficient solution in the design space.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Production of high tip deflection in a piezoelectric bimorph laminar actuator by applying high voltage is limited by many physical constraints. Therefore, piezoelectric bimorph actuator with a rigid extension of non-piezoelectric material at its tip is used to increase the tip deflection of such an actuator. Research on this type of piezoelectric bending actuator is either limited to first order constitutive relations, which do not include non-linear behavior of piezoelectric element at high electric field, or limited to curve fitting techniques. Therefore, this paper considers high electric field, and analytically models tapered piezoelectric bimorph actuator with a rigid extension of non-piezoelectric material at its tip. The stiffness, capacitance, effective tip deflection, block force, output strain energy, output energy density, input electrical energy and energy efficiency of the actuator are calculated analytically. The paper also discusses the multi-objective optimization of this type of actuator subjected to the mechanical and electrical constraints.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

An economic air pollution control model, which determines the least cost of reaching various air quality levels, is formulated. The model takes the form of a general, nonlinear, mathematical programming problem. Primary contaminant emission levels are the independent variables. The objective function is the cost of attaining various emission levels and is to be minimized subject to constraints that given air quality levels be attained.

The model is applied to a simplified statement of the photochemical smog problem in Los Angeles County in 1975 with emissions specified by a two-dimensional vector, total reactive hydrocarbon, (RHC), and nitrogen oxide, (NOx), emissions. Air quality, also two-dimensional, is measured by the expected number of days per year that nitrogen dioxide, (NO2), and mid-day ozone, (O3), exceed standards in Central Los Angeles.

The minimum cost of reaching various emission levels is found by a linear programming model. The base or "uncontrolled" emission levels are those that will exist in 1975 with the present new car control program and with the degree of stationary source control existing in 1971. Controls, basically "add-on devices", are considered here for used cars, aircraft, and existing stationary sources. It is found that with these added controls, Los Angeles County emission levels [(1300 tons/day RHC, 1000 tons /day NOx) in 1969] and [(670 tons/day RHC, 790 tons/day NOx) at the base 1975 level], can be reduced to 260 tons/day RHC (minimum RHC program) and 460 tons/day NOx (minimum NOx program).

"Phenomenological" or statistical air quality models provide the relationship between air quality and emissions. These models estimate the relationship by using atmospheric monitoring data taken at one (yearly) emission level and by using certain simple physical assumptions, (e. g., that emissions are reduced proportionately at all points in space and time). For NO2, (concentrations assumed proportional to NOx emissions), it is found that standard violations in Central Los Angeles, (55 in 1969), can be reduced to 25, 5, and 0 days per year by controlling emissions to 800, 550, and 300 tons /day, respectively. A probabilistic model reveals that RHC control is much more effective than NOx control in reducing Central Los Angeles ozone. The 150 days per year ozone violations in 1969 can be reduced to 75, 30, 10, and 0 days per year by abating RHC emissions to 700, 450, 300, and 150 tons/day, respectively, (at the 1969 NOx emission level).

The control cost-emission level and air quality-emission level relationships are combined in a graphical solution of the complete model to find the cost of various air quality levels. Best possible air quality levels with the controls considered here are 8 O3 and 10 NO2 violations per year (minimum ozone program) or 25 O3 and 3 NO2 violations per year (minimum NO2 program) with an annualized cost of $230,000,000 (above the estimated $150,000,000 per year for the new car control program for Los Angeles County motor vehicles in 1975).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantum Computing is a relatively modern field which simulates quantum computation conditions. Moreover, it can be used to estimate which quasiparticles would endure better in a quantum environment. Topological Quantum Computing (TQC) is an approximation for reducing the quantum decoherence problem1, which is responsible for error appearance in the representation of information. This project tackles specific instances of TQC problems using MOEAs (Multi-objective Optimization Evolutionary Algorithms). A MOEA is a type of algorithm which will optimize two or more objectives of a problem simultaneously, using a population based approach. We have implemented MOEAs that use probabilistic procedures found in EDAs (Estimation of Distribution Algorithms), since in general, EDAs have found better solutions than ordinary EAs (Evolutionary Algorithms), even though they are more costly. Both, EDAs and MOEAs are population-based algorithms. The objective of this project was to use a multi-objective approach in order to find good solutions for several instances of a TQC problem. In particular, the objectives considered in the project were the error approximation and the length of a solution. The tool we used to solve the instances of the problem was the multi-objective framework PISA. Because PISA has not too much documentation available, we had to go through a process of reverse-engineering of the framework to understand its modules and the way they communicate with each other. Once its functioning was understood, we began working on a module dedicated to the braid problem. Finally, we submitted this module to an exhaustive experimentation phase and collected results.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[ES] La necesidad de gestionar y repartir eficazmente los recursos escasos entre las diferentes operaciones de las empresas, hacen que éstas recurran a aplicar técnicas de la Investigación de Operaciones. Éste es el caso de los centros de llamadas, un sector emergente y dinámico que se encuentra en constante desarrollo. En este sector, la administración del trabajo requiere de técnicas predictivas para determinar el número de trabajadores adecuado y así evitar en la medida de lo posible tanto el exceso como la escasez del mismo. Este trabajo se centrará en el estudio del centro de llamadas de emergencias 112 de Andalucía. Partiendo de los datos estadísticos del número medio de llamadas que se realiza en cada franja horaria, facilitados por la Junta de esta Comunidad Autónoma, formularemos y modelizaremos el problema aplicando la Programación Lineal. Posteriormente, lo resolveremos con dos programas de software, con la finalidad de obtener una distribución óptima de agentes que minimice el coste salarial, ya que supone un 65% del gasto de explotación total. Finalmente, mediante la teoría de colas, observaremos los tiempos de espera en cola y calcularemos el número objetivo de agentes que permita no sólo minimizar el coste salarial sino mejorar la calidad de servicio teniendo unos tiempos de espera razonables.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Redes de trocadores de calor são bastante utilizadas na indústria química para promover a integração energética do processo, recuperando calor de correntes quentes para aquecer correntes frias. Estas redes estão sujeitas à deposição, o que causa um aumento na resistência à transferência de calor, prejudicando-a. Uma das principais formas de diminuir o prejuízo causado por este fenômeno é a realização periódica de limpezas nos trocadores de calor. O presente trabalho tem como objetivo desenvolver um novo método para encontrar a programação ótima das limpezas em uma rede de trocadores de calor. O método desenvolvido utiliza o conceito de horizonte deslizante associado a um problema de programação linear inteira mista (MILP). Este problema MILP é capaz de definir o conjunto ótimo de trocadores de calor a serem limpos em um determinado instante de tempo (primeiro instante do horizonte deslizante), levando em conta sua influência nos instantes futuros (restante do horizonte deslizante). O problema MILP utiliza restrições referentes aos balanços de energia, equações de trocadores de calor e número máximo de limpezas simultâneas, com o objetivo de minimizar o consumo de energia da planta. A programação ótima das limpezas é composta pela combinação dos resultados obtidos em cada um dos instantes de tempo.O desempenho desta abordagem foi analisado através de sua aplicação em diversos exemplos típicos apresentados na literatura, inclusive um exemplo de grande porte de uma refinaria brasileira. Os resultados mostraram que a abordagem aplicada foi capaz de prover ganhos semelhantes e, algumas vezes, superiores aos da literatura, indicando que o método desenvolvido é capaz de fornecer bons resultados com um baixo esforço computacional

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dutos de transmissão são tubulações especialmente desenvolvidas para transportar produtos diversos a longas distâncias e representam a forma mais segura e econômica de transporte para grandes quantidades de fluidos. Os dutos de gás natural, denominados gasodutos, são usados para transportar o gás desde os campos de produção até os centros consumidores, onde o gás é inserido em redes de distribuição para entrega aos consumidores finais. Os gasodutos de transporte apresentam diversas características de monopólio natural, que são o principal argumento econômico para sua regulação. A regulação visa garantir que esta atividade seja explorada de maneira eficiente, refletindo em tarifas de transporte justas para os consumidores e que proporcionem o retorno adequado aos investidores, levando-se em consideração a quantidade de gás transportado. Neste contexto, o presente trabalho tem como objetivo propor metodologias de otimização multi-objetivo de projetos de redes de gasodutos de transporte, envolvendo métodos a posteriori. O problema de otimização formulado contempla restrições associadas ao escoamento do gás e o comportamento das estações de compressão. A solução do problema fornece um conjunto de projetos ótimos de redes de transporte em função da maximização da quantidade de gás natural transportado e da minimização da tarifa associada a esse serviço. A ferramenta foi aplicada a diversos estudos de caso com configurações típicas da indústria de transporte de gás natural. Os resultados mostraram que as metodologias propostas são capazes de fornecer subsídios que permitem ao tomador de decisão do ponto de vista regulatório realizar uma análise de trade-off entre a quantidade de gás transportado e a tarifa, buscando assim atender ao interesse da sociedade em relação à exploração do serviço de transporte

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Multi-objective Genetic Algorithms have become a popular choice to aid in optimising the size of the whole hybrid power train. Within these optimisation processes, other optimisation techniques for the control strategy are implemented. This optimisation within an optimisation requires many simulations to be run, so reducing the computational cost is highly desired. This paper presents an optimisation framework consisting of a series hybrid optimisation algorithm, in which a global search optimizes a submarine propulsion system using low-fidelity models and, in order to refine the results, a local search is used with high-fidelity models. The effectiveness of the Hybrid optimisation algorithm is demonstrated with the optimisation of a submarine propulsion system. © 2011 EPE Association - European Power Electr.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A sensitivity study has been conducted to assess the robustness of the conclusions presented in the MIT Fuel Cycle Study. The Once Through Cycle (OTC) is considered as the base-line case, while advanced technologies with fuel recycling characterize the alternative fuel cycles. The options include limited recycling in LWRs and full recycling in fast reactors and in high conversion LWRs. Fast reactor technologies studied include both oxide and metal fueled reactors. The analysis allowed optimization of the fast reactor conversion ratio with respect to desired fuel cycle performance characteristics. The following parameters were found to significantly affect the performance of recycling technologies and their penetration over time: Capacity Factors of the fuel cycle facilities, Spent Fuel Cooling Time, Thermal Reprocessing Introduction Date, and incore and Out-of-core TRU Inventory Requirements for recycling technology. An optimization scheme of the nuclear fuel cycle is proposed. Optimization criteria and metrics of interest for different stakeholders in the fuel cycle (economics, waste management, environmental impact, etc.) are utilized for two different optimization techniques (linear and stochastic). Preliminary results covering single and multi-variable and single and multi-objective optimization demonstrate the viability of the optimization scheme.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A linear accelerator as a new injector for the SSC (Separated Sector Cyclotron) of the HIRFL (Heavy ton Research Facility Lanzhou) is being designed. The DTL (Drift-Tube-Linac) has been designed to accelerate U-238(34+) from 0.140 MeV/u to 0.97 MeV/u. To the first accelerating tank which accelerates U-238(34+) to 0.54 MeV/u, the approach of Alternating-Phase-Focusing (APF) is applied. The phase array is obtained by coupling optimization software Dakota and beam optics code LINREV. With the hybrid of Multi-objective Genetic Algorithm (MOGA) and a pattern search method, an optimum array of asynchronous phases is determined. The final growth, both transversely and longitudinally, can meet the design requirements. In this paper, the deign optimization of the APF DTL is presented.