942 resultados para Multi-objective optimization techniques


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pós-graduação em Engenharia Elétrica - FEIS

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

There are strong uncertainties regarding LAI dynamics in forest ecosystems in response to climate change. While empirical growth & yield models (G&YMs) provide good estimations of tree growth at the stand level on a yearly to decennial scale, process-based models (PBMs) use LAI dynamics as a key variable for enabling the accurate prediction of tree growth over short time scales. Bridging the gap between PBMs and G&YMs could improve the prediction of forest growth and, therefore, carbon, water and nutrient fluxes by combining modeling approaches at the stand level.Our study aimed to estimate monthly changes of leaf area in response to climate variations from sparse measurements of foliage area and biomass. A leaf population probabilistic model (SLCD) was designed to simulate foliage renewal. The leaf population was distributed in monthly cohorts, and the total population size was limited depending on forest age and productivity. Foliage dynamics were driven by a foliation function and the probabilities ruling leaf aging or fall. Their formulation depends on the forest environment.The model was applied to three tree species growing under contrasting climates and soil types. In tropical Brazilian evergreen broadleaf eucalypt plantations, the phenology was described using 8 parameters. A multi-objective evolutionary algorithm method (MOEA) was used to fit the model parameters on litterfall and LAI data over an entire stand rotation. Field measurements from a second eucalypt stand were used to validate the model. Seasonal LAI changes were accurately rendered for both sites (R-2 = 0.898 adjustment, R-2 = 0.698 validation). Litterfall production was correctly simulated (R-2 = 0.562, R-2 = 0.4018 validation) and may be improved by using additional validation data in future work. In two French temperate deciduous forests (beech and oak), we adapted phenological sub-modules of the CASTANEA model to simulate canopy dynamics, and SLCD was validated using LAI measurements. The phenological patterns were simulated with good accuracy in the two cases studied. However, IA/max was not accurately simulated in the beech forest, and further improvement is required.Our probabilistic approach is expected to contribute to improving predictions of LAI dynamics. The model formalism is general and suitable to broadleaf forests for a large range of ecological conditions. (C) 2014 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Medical Physics has been developing very fast due to the progress of the technologies and to the increase of the concerns with cure of diseases. One of the Medical Physics main performances at the present time is the use of ionizing radiations for cancer treatment, especially, services as Radiotherapy. The radiotherapy technique uses ionizing radiation with therapeutic end of cancer controls, avoiding your proliferation and it worsens of the patient. For the treatment a radiation bunch is used, with rectangular form, that it passes through the different types of tissues of the patient's body, and depending on the attenuation and of the depth of the fabrics, a great amount of energy is deposited inside in different points of the body. Like this, to plan this treatment type it should be obtained the dimension of the distribution and dose absorption along the volume. For this, it is necessary in the planning of the treatment of the cancer for radiotherapy to build isodose curves, which are lines that represent points of same amount of dose to be deposited in the area to be treated. To aid the construction of the curves of form isodose to reach the best result in the planning of the treatment, in other words, a great planning, providing the maximum of dose in the tumor and saving the healthy and critical organs, it has been using mathematical tools and computational. A plan of cancer treatment for radiotherapy is considered great when all the parameters that involve the treatment, be them physical or biological, they were investigated and adapted individually for the patient. For that, is considered the type and the location of the tumor, worrying about the elimination of the cancer without damaging the healthy tissue of the treated area, mainly the risk organs, which are in general very sensitive to the radiations. This way, the optimization techniques... (Complete abstract click electronic access below)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a mixed-integer convex-optimization-based approach for optimum investment reactive power sources in transmission systems. Unlike some convex-optimization techniques for the reactive power planning solution, in the proposed approach the taps settings of under-load tap-changing of transformers are modeled as a mixed-integer linear set equations. Are also considered the continuous and discrete variables for the existing and new capacitive and reactive power sources. The problem is solved for three significant demand scenarios (low demand, average demand and peak demand). Numerical results are presented for the CIGRE-32 electric power system.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work, nanometric displacement amplitudes of a Piezoelectric Flextensional Actuator (PFA) designed using the topology optimization technique and operating in its linear range are measured by using a homodyne Michelson interferometer. A new improved version of the J1...J4 method for optical phase measurements, named J1...J5 method, is presented, which is of easier implementation than the original one. This is a passive phase detection scheme, unaffected by signal fading, source instabilities and changes in visibility. Experimental results using this improvement were compared with those obtained by using the J1... J4, J1...J6(pos) and J1...J 6(neg) methods, concluding that the dynamic range is increased while maintaining the sensitivity. Analysis based on the 1/f voltage noise and random fading show the new method is more stable to phase drift than all those methods. © 2012 IEEE.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recently, researches have shown that the performance of metaheuristics can be affected by population initialization. Opposition-based Differential Evolution (ODE), Quasi-Oppositional Differential Evolution (QODE), and Uniform-Quasi-Opposition Differential Evolution (UQODE) are three state-of-the-art methods that improve the performance of the Differential Evolution algorithm based on population initialization and different search strategies. In a different approach to achieve similar results, this paper presents a technique to discover promising regions in a continuous search-space of an optimization problem. Using machine-learning techniques, the algorithm named Smart Sampling (SS) finds regions with high possibility of containing a global optimum. Next, a metaheuristic can be initialized inside each region to find that optimum. SS and DE were combined (originating the SSDE algorithm) to evaluate our approach, and experiments were conducted in the same set of benchmark functions used by ODE, QODE and UQODE authors. Results have shown that the total number of function evaluations required by DE to reach the global optimum can be significantly reduced and that the success rate improves if SS is employed first. Such results are also in consonance with results from the literature, stating the importance of an adequate starting population. Moreover, SS presents better efficacy to find initial populations of superior quality when compared to the other three algorithms that employ oppositional learning. Finally and most important, the SS performance in finding promising regions is independent of the employed metaheuristic with which SS is combined, making SS suitable to improve the performance of a large variety of optimization techniques. (C) 2012 Elsevier Inc. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper studies relational goods as immaterial assets creating real effects in society. The work starts answering to this question: what kind of effects do relational goods produce? After an accurate literature examination we suppose relational goods are social relations of second order. In the hypotesis they come from the emergence of two distinct social relations: interpersonal and reflexive relations. We describe empirical evidences of these emergent assets in social life and we test the effects they produce with a model. In the work we focus on four targets. First of all we describe the emergence of relational goods through a mathematical model. Then we individualize social realities where relational goods show evident effects and we outline our scientific hypotesis. The following step consists in the formulation of empirical tests. At last we explain final results. Our aim is to set apart the constitutive structure of relational goods into a checkable model coherently with the empirical evidences shown in the research. In the study we use multi-variate analysis techniques to see relational goods in a new way and we use qualitative and quantitative strategies. Relational goods are analysed both as dependent and independent variable in order to consider causative factors acting in a black-box model. Moreover we analyse effects of relational goods inside social spheres, especially in third sector and capitalistic economy. Finally we attain to effective indexes of relational goods in order to compare them with some performance indexes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Die vorliegende Arbeit beschäftigt sich mit dem Einfluß von Kettenverzweigungen unterschiedlicher Topologien auf die statischen Eigenschaften von Polymeren. Diese Untersuchungen werden mit Hilfe von Monte-Carlo- und Molekular-Dynamik-Simulationen durchgeführt.Zunächst werden einige theoretische Konzepte und Modelle eingeführt, welche die Beschreibung von Polymerketten auf mesoskopischen Längenskalen gestatten. Es werden wichtige Bestimmungsgrößen eingeführt und erläutert, welche zur quantitativen Charakterisierung von Verzweigungsstrukturen bei Polymeren geeignet sind. Es wird ebenso auf die verwendeten Optimierungstechniken eingegangen, die bei der Implementierung des Computerprogrammes Verwendung fanden. Untersucht werden neben linearen Polymerketten unterschiedliche Topolgien -Sternpolymere mit variabler Armzahl, Übergang von Sternpolymeren zu linearen Polymeren, Ketten mit variabler Zahl von Seitenketten, reguläre Dendrimere und hyperverzweigte Strukturen - in Abhängigkeit von der Lösungsmittelqualität. Es wird zunächst eine gründliche Analyse des verwendeten Simulationsmodells an sehr langen linearen Einzelketten vorgenommen. Die Skalierungseigenschaften der linearen Ketten werden untersucht in dem gesamten Lösungsmittelbereich vom guten Lösungsmittel bis hin zu weitgehend kollabierten Ketten im schlechten Lösungsmittel. Ein wichtiges Ergebnis dieser Arbeit ist die Bestätigung der Korrekturen zum Skalenverhalten des hydrodynamischen Radius Rh. Dieses Ergebnis war möglich aufgrund der großen gewählten Kettenlängen und der hohen Qualität der erhaltenen Daten in dieser Arbeit, insbesondere bei den linearen ketten, und es steht im Widerspruch zu vielen bisherigen Simulations-Studien und experimentellen Arbeiten. Diese Korrekturen zum Skalenverhalten wurden nicht nur für die linearen Ketten, sondern auch für Sternpolymere mit unterchiedlicher Armzahl gezeigt. Für lineare Ketten wird der Einfluß von Polydispersität untersucht.Es wird gezeigt, daß eine eindeutige Abbildung von Längenskalen zwischen Simulationsmodell und Experiment nicht möglich ist, da die zu diesem Zweck verwendete dimensionslose Größe eine zu schwache Abhängigkeit von der Polymerisation der Ketten besitzt. Ein Vergleich von Simulationsdaten mit industriellem Low-Density-Polyäthylen(LDPE) zeigt, daß LDPE in Form von stark verzweigten Ketten vorliegt.Für reguläre Dendrimere konnte ein hochgradiges Zurückfalten der Arme in die innere Kernregion nachgewiesen werden.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Materials are inherently multi-scale in nature consisting of distinct characteristics at various length scales from atoms to bulk material. There are no widely accepted predictive multi-scale modeling techniques that span from atomic level to bulk relating the effects of the structure at the nanometer (10-9 meter) on macro-scale properties. Traditional engineering deals with treating matter as continuous with no internal structure. In contrast to engineers, physicists have dealt with matter in its discrete structure at small length scales to understand fundamental behavior of materials. Multiscale modeling is of great scientific and technical importance as it can aid in designing novel materials that will enable us to tailor properties specific to an application like multi-functional materials. Polymer nanocomposite materials have the potential to provide significant increases in mechanical properties relative to current polymers used for structural applications. The nanoscale reinforcements have the potential to increase the effective interface between the reinforcement and the matrix by orders of magnitude for a given reinforcement volume fraction as relative to traditional micro- or macro-scale reinforcements. To facilitate the development of polymer nanocomposite materials, constitutive relationships must be established that predict the bulk mechanical properties of the materials as a function of the molecular structure. A computational hierarchical multiscale modeling technique is developed to study the bulk-level constitutive behavior of polymeric materials as a function of its molecular chemistry. Various parameters and modeling techniques from computational chemistry to continuum mechanics are utilized for the current modeling method. The cause and effect relationship of the parameters are studied to establish an efficient modeling framework. The proposed methodology is applied to three different polymers and validated using experimental data available in literature.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Particulate matter (PM) emissions standards set by the US Environmental Protection Agency (EPA) have become increasingly stringent over the years. The EPA regulation for PM in heavy duty diesel engines has been reduced to 0.01 g/bhp-hr for the year 2010. Heavy duty diesel engines make use of an aftertreatment filtration device, the Diesel Particulate Filter (DPF). DPFs are highly efficient in filtering PM (known as soot) and are an integral part of 2010 heavy duty diesel aftertreatment system. PM is accumulated in the DPF as the exhaust gas flows through it. This PM needs to be removed by oxidation periodically for the efficient functioning of the filter. This oxidation process is also known as regeneration. There are 2 types of regeneration processes, namely active regeneration (oxidation of PM by external means) and passive oxidation (oxidation of PM by internal means). Active regeneration occurs typically in high temperature regions, about 500 - 600 °C, which is much higher than normal diesel exhaust temperatures. Thus, the exhaust temperature has to be raised with the help of external devices like a Diesel Oxidation Catalyst (DOC) or a fuel burner. The O2 oxidizes PM producing CO2 as oxidation product. In passive oxidation, one way of regeneration is by the use of NO2. NO2 oxidizes the PM producing NO and CO2 as oxidation products. The passive oxidation process occurs at lower temperatures (200 - 400 °C) in comparison to the active regeneration temperatures. Generally, DPF substrate walls are washcoated with catalyst material to speed up the rate of PM oxidation. The catalyst washcoat is observed to increase the rate of PM oxidation. The goal of this research is to develop a simple mathematical model to simulate the PM depletion during the active regeneration process in a DPF (catalyzed and non-catalyzed). A simple, zero-dimensional kinetic model was developed in MATLAB. Experimental data required for calibration was obtained by active regeneration experiments performed on PM loaded mini DPFs in an automated flow reactor. The DPFs were loaded with PM from the exhaust of a commercial heavy duty diesel engine. The model was calibrated to the data obtained from active regeneration experiments. Numerical gradient based optimization techniques were used to estimate the kinetic parameters of the model.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

With proper application of Best Management Practices (BMPs), the impact from the sediment to the water bodies could be minimized. However, finding the optimal allocation of BMP can be difficult, since there are numerous possible options. Also, economics plays an important role in BMP affordability and, therefore, the number of BMPs able to be placed in a given budget year. In this study, two methodologies are presented to determine the optimal cost-effective BMP allocation, by coupling a watershed-level model, Soil and Water Assessment Tool (SWAT), with two different methods, targeting and a multi-objective genetic algorithm (Non-dominated Sorting Genetic Algorithm II, NSGA-II). For demonstration, these two methodologies were applied to an agriculture-dominant watershed located in Lower Michigan to find the optimal allocation of filter strips and grassed waterways. For targeting, three different criteria were investigated for sediment yield minimization, during the process of which it was found that the grassed waterways near the watershed outlet reduced the watershed outlet sediment yield the most under this study condition, and cost minimization was also included as a second objective during the cost-effective BMP allocation selection. NSGA-II was used to find the optimal BMP allocation for both sediment yield reduction and cost minimization. By comparing the results and computational time of both methodologies, targeting was determined to be a better method for finding optimal cost-effective BMP allocation under this study condition, since it provided more than 13 times the amount of solutions with better fitness for the objective functions while using less than one eighth of the SWAT computational time than the NSGA-II with 150 generations did.