978 resultados para Robust Stochastic Optimization


Relevância:

40.00% 40.00%

Publicador:

Resumo:

Abstract : Wastepaper sludge ash (WSA) is generated by a cogeneration station by burning wastepaper sludge. It mainly consists of amorphous aluminosilicate phase, anhydrite, gehlenite, calcite, lime, C2S, C3A, quartz, anorthite, traces of mayenite. Because of its free lime content (~10%), WSA suspension has a high pH (13). Previous researchers have found that the WSA composition has poor robustness and the variations lead to some unsoundness for Portland cement (PC) blended WSA concrete. This thesis focused on the use of WSA in different types of concrete mixes to avoid the deleterious effect of the expansion due to the WSA hydration. As a result, WSA were used in making alkali-activated materials (AAMs) as a precursor source and as a potential activator in consideration of its amorphous content and the high alkaline nature. Moreover, the autogenous shrinkage behavior of PC concrete at low w/b ratio was used in order to compensate the expansion effect due to WSA. The concrete properties as well as the volume change were investigated for the modified WSA blended concrete. The reaction mechanism and microstructure of newly formed binder were evaluated by X-ray diffraction (XRD), calorimetry, thermogravimetric analysis (TGA), scanning electron microscopy (SEM) and energy dispersive X-ray spectroscopy (EDX). When WSA was used as precursor, the results showed incompatible reaction between WSA and alkaline solution. The mixtures were not workable and provided very low compressive strength no matter what kinds of chemical activators were used. This was due to the metallic aluminum in WSA, which releases abundant hydrogen gas when WSA reacts with strong alkaline solution. Besides, the results of this thesis showed that WSA can activate the glassy phase contained in slag, glass powder (GP) and class F fly ash (FFA) with an optimum blended ratio of 50:50. The WSA/slag (mass ratio of 50:50) mortar (w/b of 0.47) attained 46 MPa at 28 days without heat curing assistance. A significant fast setting was noticed for the WSA-activated binder due to the C3A phase, free lime and metallic aluminum contained in the WSA. Adding 5% of gypsum can delay the fast setting, but this greatly increased the potential risk of intern sulfate attack. The XRD, TGA and calorimetry analyses demonstrated the formation of ettringite, C-S-H, portlandite, hydrogarnet and calcium carboaluminate in the hydrated binder. The mechanical performance of different binder was closely related to the microstructure of corresponding binder which was proved by the SEM observation. The hydrated WSA/slag and WSA/FFA binder formed a C-A-S-H type of gel with lower Ca/Si ratio (0.47~1.6). A hybrid gel (i.e. C-N-A-S-H) was observed for the WSA/GP binder with a very low Ca/Si ratio (0.26) and Na/Si ratio (0.03). The SEM/EDX analyses displayed the formation of expansive gel (ettringite and thaumasite) in the gypsum added WSA/slag concrete. The gradual emission of hydrogen gas due to the reaction of WSA with alkaline environment significantly increased the porosity and degraded the microstructure of hydrated matrix after the setting. In the last phase of this research WSA-PC blended binder was tailored to form a high autogenous shrinkage concrete in order to compensate the initial expansion. Different binders were proportioned with PC, WSA, silica fume or slag. The microstructure and mechanical properties of concrete can be improved by decreasing w/b ratios and by incorporating silica fume or slag. The 28-day compressive strength of WSA-blended concrete was above 22 MPa and reached 45 MPa when silica fume was added. The PC concrete incorporating silica fume or slag tended to develop higher autogenous shrinkage at low w/b ratios, and thus the ternary binder with the addition of WSA inhibited the long term shrinkage due to the initial expansion property to WSA. In the restrained shrinkage test, the concrete ring incorporating the ternary binder (PC/WSA/slag) revealed negligible potential to cracking up to 96 days as a result of the offset effect by WSA expansion. The WSA blended regular concrete could be produced for potential applications with reduced expansion, good mechanical property and lower permeability.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Considerable interest in renewable energy has increased in recent years due to the concerns raised over the environmental impact of conventional energy sources and their price volatility. In particular, wind power has enjoyed a dramatic global growth in installed capacity over the past few decades. Nowadays, the advancement of wind turbine industry represents a challenge for several engineering areas, including materials science, computer science, aerodynamics, analytical design and analysis methods, testing and monitoring, and power electronics. In particular, the technological improvement of wind turbines is currently tied to the use of advanced design methodologies, allowing the designers to develop new and more efficient design concepts. Integrating mathematical optimization techniques into the multidisciplinary design of wind turbines constitutes a promising way to enhance the profitability of these devices. In the literature, wind turbine design optimization is typically performed deterministically. Deterministic optimizations do not consider any degree of randomness affecting the inputs of the system under consideration, and result, therefore, in an unique set of outputs. However, given the stochastic nature of the wind and the uncertainties associated, for instance, with wind turbine operating conditions or geometric tolerances, deterministically optimized designs may be inefficient. Therefore, one of the ways to further improve the design of modern wind turbines is to take into account the aforementioned sources of uncertainty in the optimization process, achieving robust configurations with minimal performance sensitivity to factors causing variability. The research work presented in this thesis deals with the development of a novel integrated multidisciplinary design framework for the robust aeroservoelastic design optimization of multi-megawatt horizontal axis wind turbine (HAWT) rotors, accounting for the stochastic variability related to the input variables. The design system is based on a multidisciplinary analysis module integrating several simulations tools needed to characterize the aeroservoelastic behavior of wind turbines, and determine their economical performance by means of the levelized cost of energy (LCOE). The reported design framework is portable and modular in that any of its analysis modules can be replaced with counterparts of user-selected fidelity. The presented technology is applied to the design of a 5-MW HAWT rotor to be used at sites of wind power density class from 3 to 7, where the mean wind speed at 50 m above the ground ranges from 6.4 to 11.9 m/s. Assuming the mean wind speed to vary stochastically in such range, the rotor design is optimized by minimizing the mean and standard deviation of the LCOE. Airfoil shapes, spanwise distributions of blade chord and twist, internal structural layup and rotor speed are optimized concurrently, subject to an extensive set of structural and aeroelastic constraints. The effectiveness of the multidisciplinary and robust design framework is demonstrated by showing that the probabilistically designed turbine achieves more favorable probabilistic performance than those of the initial baseline turbine and a turbine designed deterministically.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Production companies use raw materials to compose end-products. They often make different products with the same raw materials. In this research, the focus lies on the production of two end-products consisting of (partly) the same raw materials as cheap as possible. Each of the products has its own demand and quality requirements consisting of quadratic constraints. The minimization of the costs, given the quadratic constraints is a global optimization problem, which can be difficult because of possible local optima. Therefore, the multi modal character of the (bi-) blend problem is investigated. Standard optimization packages (solvers) in Matlab and GAMS were tested on their ability to solve the problem. In total 20 test cases were generated and taken from literature to test solvers on their effectiveness and efficiency to solve the problem. The research also gives insight in adjusting the quadratic constraints of the problem in order to make a robust problem formulation of the bi-blend problem.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Une approche classique pour traiter les problèmes d’optimisation avec incertitude à deux- et multi-étapes est d’utiliser l’analyse par scénario. Pour ce faire, l’incertitude de certaines données du problème est modélisée par vecteurs aléatoires avec des supports finis spécifiques aux étapes. Chacune de ces réalisations représente un scénario. En utilisant des scénarios, il est possible d’étudier des versions plus simples (sous-problèmes) du problème original. Comme technique de décomposition par scénario, l’algorithme de recouvrement progressif est une des méthodes les plus populaires pour résoudre les problèmes de programmation stochastique multi-étapes. Malgré la décomposition complète par scénario, l’efficacité de la méthode du recouvrement progressif est très sensible à certains aspects pratiques, tels que le choix du paramètre de pénalisation et la manipulation du terme quadratique dans la fonction objectif du lagrangien augmenté. Pour le choix du paramètre de pénalisation, nous examinons quelques-unes des méthodes populaires, et nous proposons une nouvelle stratégie adaptive qui vise à mieux suivre le processus de l’algorithme. Des expériences numériques sur des exemples de problèmes stochastiques linéaires multi-étapes suggèrent que la plupart des techniques existantes peuvent présenter une convergence prématurée à une solution sous-optimale ou converger vers la solution optimale, mais avec un taux très lent. En revanche, la nouvelle stratégie paraît robuste et efficace. Elle a convergé vers l’optimalité dans toutes nos expériences et a été la plus rapide dans la plupart des cas. Pour la question de la manipulation du terme quadratique, nous faisons une revue des techniques existantes et nous proposons l’idée de remplacer le terme quadratique par un terme linéaire. Bien que qu’il nous reste encore à tester notre méthode, nous avons l’intuition qu’elle réduira certaines difficultés numériques et théoriques de la méthode de recouvrement progressif.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Une approche classique pour traiter les problèmes d’optimisation avec incertitude à deux- et multi-étapes est d’utiliser l’analyse par scénario. Pour ce faire, l’incertitude de certaines données du problème est modélisée par vecteurs aléatoires avec des supports finis spécifiques aux étapes. Chacune de ces réalisations représente un scénario. En utilisant des scénarios, il est possible d’étudier des versions plus simples (sous-problèmes) du problème original. Comme technique de décomposition par scénario, l’algorithme de recouvrement progressif est une des méthodes les plus populaires pour résoudre les problèmes de programmation stochastique multi-étapes. Malgré la décomposition complète par scénario, l’efficacité de la méthode du recouvrement progressif est très sensible à certains aspects pratiques, tels que le choix du paramètre de pénalisation et la manipulation du terme quadratique dans la fonction objectif du lagrangien augmenté. Pour le choix du paramètre de pénalisation, nous examinons quelques-unes des méthodes populaires, et nous proposons une nouvelle stratégie adaptive qui vise à mieux suivre le processus de l’algorithme. Des expériences numériques sur des exemples de problèmes stochastiques linéaires multi-étapes suggèrent que la plupart des techniques existantes peuvent présenter une convergence prématurée à une solution sous-optimale ou converger vers la solution optimale, mais avec un taux très lent. En revanche, la nouvelle stratégie paraît robuste et efficace. Elle a convergé vers l’optimalité dans toutes nos expériences et a été la plus rapide dans la plupart des cas. Pour la question de la manipulation du terme quadratique, nous faisons une revue des techniques existantes et nous proposons l’idée de remplacer le terme quadratique par un terme linéaire. Bien que qu’il nous reste encore à tester notre méthode, nous avons l’intuition qu’elle réduira certaines difficultés numériques et théoriques de la méthode de recouvrement progressif.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Over the last century, mathematical optimization has become a prominent tool for decision making. Its systematic application in practical fields such as economics, logistics or defense led to the development of algorithmic methods with ever increasing efficiency. Indeed, for a variety of real-world problems, finding an optimal decision among a set of (implicitly or explicitly) predefined alternatives has become conceivable in reasonable time. In the last decades, however, the research community raised more and more attention to the role of uncertainty in the optimization process. In particular, one may question the notion of optimality, and even feasibility, when studying decision problems with unknown or imprecise input parameters. This concern is even more critical in a world becoming more and more complex —by which we intend, interconnected —where each individual variation inside a system inevitably causes other variations in the system itself. In this dissertation, we study a class of optimization problems which suffer from imprecise input data and feature a two-stage decision process, i.e., where decisions are made in a sequential order —called stages —and where unknown parameters are revealed throughout the stages. The applications of such problems are plethora in practical fields such as, e.g., facility location problems with uncertain demands, transportation problems with uncertain costs or scheduling under uncertain processing times. The uncertainty is dealt with a robust optimization (RO) viewpoint (also known as "worst-case perspective") and we present original contributions to the RO literature on both the theoretical and practical side.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper deals with the problem of state prediction for descriptor systems subject to bounded uncertainties. The problem is stated in terms of the optimization of an appropriate quadratic functional. This functional is well suited to derive not only the robust predictor for descriptor systems but also that for usual state-space systems. Numerical examples are included in order to demonstrate the performance of this new filter. (C) 2008 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We derive an easy-to-compute approximate bound for the range of step-sizes for which the constant-modulus algorithm (CMA) will remain stable if initialized close to a minimum of the CM cost function. Our model highlights the influence, of the signal constellation used in the transmission system: for smaller variation in the modulus of the transmitted symbols, the algorithm will be more robust, and the steady-state misadjustment will be smaller. The theoretical results are validated through several simulations, for long and short filters and channels.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we deal with a generalized multi-period mean-variance portfolio selection problem with market parameters Subject to Markov random regime switchings. Problems of this kind have been recently considered in the literature for control over bankruptcy, for cases in which there are no jumps in market parameters (see [Zhu, S. S., Li, D., & Wang, S. Y. (2004). Risk control over bankruptcy in dynamic portfolio selection: A generalized mean variance formulation. IEEE Transactions on Automatic Control, 49, 447-457]). We present necessary and Sufficient conditions for obtaining an optimal control policy for this Markovian generalized multi-period meal-variance problem, based on a set of interconnected Riccati difference equations, and oil a set of other recursive equations. Some closed formulas are also derived for two special cases, extending some previous results in the literature. We apply the results to a numerical example with real data for Fisk control over bankruptcy Ill a dynamic portfolio selection problem with Markov jumps selection problem. (C) 2008 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

1. Establishing biological control agents in the field is a major step in any classical biocontrol programme, yet there are few general guidelines to help the practitioner decide what factors might enhance the establishment of such agents. 2. A stochastic dynamic programming (SDP) approach, linked to a metapopulation model, was used to find optimal release strategies (number and size of releases), given constraints on time and the number of biocontrol agents available. By modelling within a decision-making framework we derived rules of thumb that will enable biocontrol workers to choose between management options, depending on the current state of the system. 3. When there are few well-established sites, making a few large releases is the optimal strategy. For other states of the system, the optimal strategy ranges from a few large releases, through a mixed strategy (a variety of release sizes), to many small releases, as the probability of establishment of smaller inocula increases. 4. Given that the probability of establishment is rarely a known entity, we also strongly recommend a mixed strategy in the early stages of a release programme, to accelerate learning and improve the chances of finding the optimal approach.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Numerical optimisation methods are being more commonly applied to agricultural systems models, to identify the most profitable management strategies. The available optimisation algorithms are reviewed and compared, with literature and our studies identifying evolutionary algorithms (including genetic algorithms) as superior in this regard to simulated annealing, tabu search, hill-climbing, and direct-search methods. Results of a complex beef property optimisation, using a real-value genetic algorithm, are presented. The relative contributions of the range of operational options and parameters of this method are discussed, and general recommendations listed to assist practitioners applying evolutionary algorithms to the solution of agricultural systems. (C) 2001 Elsevier Science Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

New designs for force-minimized compact high-field clinical MRI magnets are described. The design method is a modified simulated annealing (SA) procedure which includes Maxwell forces in the error function to be minimized. This permits an automated force reduction in the magnet designs while controlling the overall dimensions of the system. As SA optimization requires many iterations to achieve a final design, it is important that each iteration in the procedure is rapid. We have therefore developed a rapid force calculation algorithm. Novel designs for short 3- and 4-T clinical MRI systems are presented in which force reduction has been invoked. The final designs provide large homogeneous regions and reduced stray fields in remarkable short magnets. A shielded 4-T design that is approximately 30% shorter than current designs is presented. This novel magnet generates a full 50-cm diameter homogeneous region.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, a stochastic programming approach is proposed for trading wind energy in a market environment under uncertainty. Uncertainty in the energy market prices is the main cause of high volatility of profits achieved by power producers. The volatile and intermittent nature of wind energy represents another source of uncertainty. Hence, each uncertain parameter is modeled by scenarios, where each scenario represents a plausible realization of the uncertain parameters with an associated occurrence probability. Also, an appropriate risk measurement is considered. The proposed approach is applied on a realistic case study, based on a wind farm in Portugal. Finally, conclusions are duly drawn. (C) 2011 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The current regulatory framework for maintenance outage scheduling in distribution systems needs revision to face the challenges of future smart grids. In the smart grid context, generation units and the system operator perform new roles with different objectives, and an efficient coordination between them becomes necessary. In this paper, the distribution system operator (DSO) of a microgrid receives the proposals for shortterm (ST) planned outages from the generation and transmission side, and has to decide the final outage plans, which is mandatory for the members to follow. The framework is based on a coordination procedure between the DSO and other market players. This paper undertakes the challenge of optimization problem in a smart grid where the operator faces with uncertainty. The results show the effectiveness and applicability of the proposed regulatory framework in the modified IEEE 34- bus test system.