947 resultados para Bio-inspired optimization techniques


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a new methodology for the adjustment of fuzzy inference systems. A novel approach, which uses unconstrained optimization techniques, is developed in order to adjust the free parameters of the fuzzy inference system, such as its intrinsic parameters of the membership function and the weights of the inference rules. This methodology is interesting, not only for the results presented and obtained through computer simulations, but also for its generality concerning to the kind of fuzzy inference system used. Therefore, this methodology is expandable either to the Mandani architecture or also to that suggested by Takagi-Sugeno. The validation of the presented methodology is accomplished through an estimation of time series. More specifically, the Mackey-Glass chaotic time series estimation is used for the validation of the proposed methodology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a metaheuristic algorithm inspired in evolutionary computation and swarm intelligence concepts and fundamentals of echolocation of micro bats. The aim is to optimize the mono and multiobjective optimization problems related to the brushless DC wheel motor problems, which has 5 design parameters and 6 constraints for the mono-objective problem and 2 objectives, 5 design parameters, and 5 constraints for multiobjective version. Furthermore, results are compared with other optimization approaches proposed in the recent literature, showing the feasibility of this newly introduced technique to high nonlinear problems in electromagnetics.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Die biologische Stickstofffixierung durch Molybdän-haltige Nitrogenasen sowie die Erforschung des zugrundeliegenden komplexen Mechanismus (N2-Aktivierung an Metall-Zentren, 6-fache Protonierung und Reduktion, N–N Bindungsspaltung unter Bildung von Ammoniak) ist von erheblichem Interesse. Insbesondere Molybdän-Komplexe wurden bereits erfolgreich als Modellverbindungen für die Untersuchung elementarer Einzelschritte der N2-Aktivierung eingesetzt. Durch die Verwendung von Triamidoamin-Liganden ist es Schrock et al. sogar gelungen mehrere Katalysezyklen zu durchlaufen und einen Mechanismus zu formulieren. Trotz der sterisch anspruchsvollen Substituenten in den Schrock-Komplexen ist die Umsatzrate dieses homogenen Katalysators, aufgrund Komplex-Deaktivierung infolge intermolekularer Reaktionen wie Dimerisierung und Disproportionierung, limitiert. In der vorliegenden Arbeit wurden einige dieser Herausforderungen angegangen und die aktiven Spezies auf einer Festphase immobilisiert, um intermolekulare Reaktionen durch räumliche Isolierung der Komplexe zu unterdrücken.rnEin Polymer-verankertes Analogon des Schrock Nitrido-Molybdän(VI)-Komplexes wurde auf einem neuen Reaktionsweg synthetisiert. Dieser beinhaltet nur einen einzigen Reaktionsschritt, um die funktionelle Gruppe „MoN“ einzuführen. Protonierung des immobilisierten Nitrido-Molybdän(VI)-Komplexes LMoVIN (L = Polymer-verankerter Triamidoamin-Ligand) mit 2,6-Lutidinium liefert den entsprechenden Imido-Molybdän(VI)-Komplex. Durch anschließende Ein-Elektronen-Reduktion mit Cobaltocen wird der Polymer-angebundene Imido-Molybdän(V)-Komplex erhalten, bewiesen durch EPR-Spektroskopie (g1,2,3 = 1.989, 1.929, 1.902). Durch die Immobilisierung und die effektive räumliche Separation der Reaktionszentren auf der Festphase werden bimolekulare Nebenreaktionen, die oft in homogenen Systemen auftreten, unterdrückt. Dies ermöglicht zum ersten Mal die Darstellung des Imido-Molybdän(V)-Intermediates des Schrock-Zyklus.rnEPR-Spektren des als Spin-Label eingeführten immobilisierten Nitrato-Kupfer(II)-Komplexes wurden unter verschiedenen Bedingungen (Lösungsmittel, Temperatur) aufgenommen, wobei sich eine starke Abhängigkeit zwischen der Zugänglichkeit und Reaktivität der immobilisierten Reaktionszentren und der Art des Lösungsmittels zeigte. Somit wurde die Reaktivität von LMoVIN gegenüber Protonen und Elektronen, welches zur Bildung von NH3 führt, unter Verwendung verschiedener Lösungsmittel untersucht und optimiert. Innerhalb des kugelförmigen Polymers verläuft die Protonierung und Reduktion von LMoVIN stufenweise. Aktive Zentren, die sich an der „äußeren Schale“ des Polymers befinden, sind gut zugänglich und reagieren schnell nach H+/e− Zugabe. Aktive Zentren im „Inneren des Polymers“ hingegen sind schlechter zugänglich und zeigen langsame diffusions-kontrollierte Reaktionen, wobei drei H+/e− Schritte gefolgt von einer Ligandenaustausch-Reaktion erforderlich sind, um NH3 freizusetzen: LMoVIN  LMoVNH  LMoIVNH2  LMoIIINH3 und anschließender Ligandenaustausch führt zur Freisetzung von NH3.rnIn einem weiteren Projekt wurde der Bis(ddpd)-Kupfer(II)-Komplex EPR-spektroskopisch in Hinblick auf Jahn−Teller-Verzerrung und -Dynamik untersucht. Dabei wurden die EPR-Spektren bei variabler Temperatur (70−293 K) aufgenommen. Im Festkörperspektrum bei T < 100 K erscheint der Kupfer(II)-Komplex als gestreckter Oktaeder, wohingegen das EPR-Spektrum bei höheren Temperaturen g-Werte aufzeigt, die einer pseudo-gestauchten oktaedrischen Kupfer(II)-Spezies zuzuordnen sind. Diese Tatsache wird einem intramolekularen dynamischen Jahn−Teller Phänomen zugeschrieben, welcher bei 100 K eingefroren wird.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background The reduction in the amount of food available for European avian scavengers as a consequence of restrictive public health policies is a concern for managers and conservationists. Since 2002, the application of several sanitary regulations has limited the availability of feeding resources provided by domestic carcasses, but theoretical studies assessing whether the availability of food resources provided by wild ungulates are enough to cover energetic requirements are lacking. Methodology/Findings We assessed food provided by a wild ungulate population in two areas of NE Spain inhabited by three vulture species and developed a P System computational model to assess the effects of the carrion resources provided on their population dynamics. We compared the real population trend with to a hypothetical scenario in which only food provided by wild ungulates was available. Simulation testing of the model suggests that wild ungulates constitute an important food resource in the Pyrenees and the vulture population inhabiting this area could grow if only the food provided by wild ungulates would be available. On the contrary, in the Pre-Pyrenees there is insufficient food to cover the energy requirements of avian scavenger guilds, declining sharply if biomass from domestic animals would not be available. Conclusions/Significance Our results suggest that public health legislation can modify scavenger population trends if a large number of domestic ungulate carcasses disappear from the mountains. In this case, food provided by wild ungulates could be not enough and supplementary feeding could be necessary if other alternative food resources are not available (i.e. the reintroduction of wild ungulates), preferably in European Mediterranean scenarios sharing similar and socio-economic conditions where there are low densities of wild ungulates. Managers should anticipate the conservation actions required by assessing food availability and the possible scenarios in order to make the most suitable decisions.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We present a high performance-yet low cost-system for multi-view rendering in virtual reality (VR) applications. In contrast to complex CAVE installations, which are typically driven by one render client per view, we arrange eight displays in an octagon around the viewer to provide a full 360° projection, and we drive these eight displays by a single PC equipped with multiple graphics units (GPUs). In this paper we describe the hardware and software setup, as well as the necessary low-level and high-level optimizations to optimally exploit the parallelism of this multi-GPU multi-view VR system.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We report about a lung-on-chip array that mimics the pulmonary parenchymal environment, including the thin, alveolar barrier and the three-dimensional cyclic strain induced by the breathing movements. A micro-diaphragm used to stretch the alveolar barrier is inspired by the in-vivo diaphragm, the main muscle responsible for inspiration. The design of this device aims not only at best reproducing the in-vivo conditions found in the lung parenchyma, but also at making its handling easy and robust. An innovative concept, based on the reversible bonding of the device, is presented that enables to accurately control the concentration of cells cultured on the membrane by easily accessing both sides of the membranes. The functionality of the alveolar barrier could be restored by co-culturing epithelial and endothelial cells that formed tight monolayers on each side of a thin, porous and stretchable membrane. We showed that cyclic stretch significantly affects the permeability properties of epithelial cell layers. Furthermore, we could also demonstrate that the strain influences the metabolic activity and the cytokine secretion of primary human pulmonary alveolar epithelial cells obtained from patients. These results demonstrate the potential of this device and confirm the importance of the mechanical strain induced by the breathing in pulmonary research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A generic bio-inspired adaptive architecture for image compression suitable to be implemented in embedded systems is presented. The architecture allows the system to be tuned during its calibration phase. An evolutionary algorithm is responsible of making the system evolve towards the required performance. A prototype has been implemented in a Xilinx Virtex-5 FPGA featuring an adaptive wavelet transform core directed at improving image compression for specific types of images. An Evolution Strategy has been chosen as the search algorithm and its typical genetic operators adapted to allow for a hardware friendly implementation. HW/SW partitioning issues are also considered after a high level description of the algorithm is profiled which validates the proposed resource allocation in the device fabric. To check the robustness of the system and its adaptation capabilities, different types of images have been selected as validation patterns. A direct application of such a system is its deployment in an unknown environment during design time, letting the calibration phase adjust the system parameters so that it performs efcient image compression. Also, this prototype implementation may serve as an accelerator for the automatic design of evolved transform coefficients which are later on synthesized and implemented in a non-adaptive system in the final implementation device, whether it is a HW or SW based computing device. The architecture has been built in a modular way so that it can be easily extended to adapt other types of image processing cores. Details on this pluggable component point of view are also given in the paper.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work, we propose the Networks of Evolutionary Processors (NEP) [2] as a computational model to solve problems related with biological phenomena. In our first approximation, we simulate biological processes related with cellular signaling and their implications in the metabolism, by using an architecture based on NEP (NEP architecture) and their specializations: Networks of Polarized Evolutionary Processors (NPEP) [1] and NEP Transducers (NEPT) [3]. In particular, we use this architecture to simulate the interplay between cellular processes related with the metabolism as the Krebs cycle and the malate-aspartate shuttle pathway (MAS) both being altered by signaling by calcium.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

n this paper we propose the use of Networks of Bio-inspired Processors (NBP) to model some biological phenomena within a computational framework. In particular, we propose the use of an extension of NBP named Network Evolutionary Processors Transducers to simulate chemical transformations of substances. Within a biological process, chemical transformations of substances are basic operations in the change of the state of the cell. Previously, it has been proved that NBP are computationally complete, that is, they are able to solve NP complete problems in linear time, using massively parallel computations. In addition, we propose a multilayer architecture that will allow us to design models of biological processes related to cellular communication as well as their implications in the metabolic pathways. Subsequently, these models can be applied not only to biological-cellular instances but, possibly, also to configure instances of interactive processes in many other fields like population interactions, ecological trophic networks, in dustrial ecosystems, etc.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a new simulation environment aimed at heterogeneous chained modular robots. This simulator allows testing the feasibility of the design, checking how modules are going to perform in the field and verifying hardware, electronics and communication designs before the prototype is built, saving time and resources. The paper shows how the simulator is built and how it can be set up to adapt to new designs. It also gives some examples of its use showing different heterogeneous modular robots running in different environments.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Data centers are easily found in every sector of the worldwide economy. They are composed of thousands of servers, serving millions of users globally and 24-7. In the last years, e-Science applications such e-Health or Smart Cities have experienced a significant development. The need to deal efficiently with the computational needs of next-generation applications together with the increasing demand for higher resources in traditional applications has facilitated the rapid proliferation and growing of Data Centers. A drawback to this capacity growth has been the rapid increase of the energy consumption of these facilities. In 2010, data center electricity represented 1.3% of all the electricity use in the world. In year 2012 alone, global data center power demand grep 63% to 38GW. A further rise of 17% to 43GW was estimated in 2013. Moreover, Data Centers are responsible for more than 2% of total carbon dioxide emissions.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Los Centros de Datos se encuentran actualmente en cualquier sector de la economía mundial. Están compuestos por miles de servidores, dando servicio a los usuarios de forma global, las 24 horas del día y los 365 días del año. Durante los últimos años, las aplicaciones del ámbito de la e-Ciencia, como la e-Salud o las Ciudades Inteligentes han experimentado un desarrollo muy significativo. La necesidad de manejar de forma eficiente las necesidades de cómputo de aplicaciones de nueva generación, junto con la creciente demanda de recursos en aplicaciones tradicionales, han facilitado el rápido crecimiento y la proliferación de los Centros de Datos. El principal inconveniente de este aumento de capacidad ha sido el rápido y dramático incremento del consumo energético de estas infraestructuras. En 2010, la factura eléctrica de los Centros de Datos representaba el 1.3% del consumo eléctrico mundial. Sólo en el año 2012, el consumo de potencia de los Centros de Datos creció un 63%, alcanzando los 38GW. En 2013 se estimó un crecimiento de otro 17%, hasta llegar a los 43GW. Además, los Centros de Datos son responsables de más del 2% del total de emisiones de dióxido de carbono a la atmósfera. Esta tesis doctoral se enfrenta al problema energético proponiendo técnicas proactivas y reactivas conscientes de la temperatura y de la energía, que contribuyen a tener Centros de Datos más eficientes. Este trabajo desarrolla modelos de energía y utiliza el conocimiento sobre la demanda energética de la carga de trabajo a ejecutar y de los recursos de computación y refrigeración del Centro de Datos para optimizar el consumo. Además, los Centros de Datos son considerados como un elemento crucial dentro del marco de la aplicación ejecutada, optimizando no sólo el consumo del Centro de Datos sino el consumo energético global de la aplicación. Los principales componentes del consumo en los Centros de Datos son la potencia de computación utilizada por los equipos de IT, y la refrigeración necesaria para mantener los servidores dentro de un rango de temperatura de trabajo que asegure su correcto funcionamiento. Debido a la relación cúbica entre la velocidad de los ventiladores y el consumo de los mismos, las soluciones basadas en el sobre-aprovisionamiento de aire frío al servidor generalmente tienen como resultado ineficiencias energéticas. Por otro lado, temperaturas más elevadas en el procesador llevan a un consumo de fugas mayor, debido a la relación exponencial del consumo de fugas con la temperatura. Además, las características de la carga de trabajo y las políticas de asignación de recursos tienen un impacto importante en los balances entre corriente de fugas y consumo de refrigeración. La primera gran contribución de este trabajo es el desarrollo de modelos de potencia y temperatura que permiten describes estos balances entre corriente de fugas y refrigeración; así como la propuesta de estrategias para minimizar el consumo del servidor por medio de la asignación conjunta de refrigeración y carga desde una perspectiva multivariable. Cuando escalamos a nivel del Centro de Datos, observamos un comportamiento similar en términos del balance entre corrientes de fugas y refrigeración. Conforme aumenta la temperatura de la sala, mejora la eficiencia de la refrigeración. Sin embargo, este incremente de la temperatura de sala provoca un aumento en la temperatura de la CPU y, por tanto, también del consumo de fugas. Además, la dinámica de la sala tiene un comportamiento muy desigual, no equilibrado, debido a la asignación de carga y a la heterogeneidad en el equipamiento de IT. La segunda contribución de esta tesis es la propuesta de técnicas de asigación conscientes de la temperatura y heterogeneidad que permiten optimizar conjuntamente la asignación de tareas y refrigeración a los servidores. Estas estrategias necesitan estar respaldadas por modelos flexibles, que puedan trabajar en tiempo real, para describir el sistema desde un nivel de abstracción alto. Dentro del ámbito de las aplicaciones de nueva generación, las decisiones tomadas en el nivel de aplicación pueden tener un impacto dramático en el consumo energético de niveles de abstracción menores, como por ejemplo, en el Centro de Datos. Es importante considerar las relaciones entre todos los agentes computacionales implicados en el problema, de forma que puedan cooperar para conseguir el objetivo común de reducir el coste energético global del sistema. La tercera contribución de esta tesis es el desarrollo de optimizaciones energéticas para la aplicación global por medio de la evaluación de los costes de ejecutar parte del procesado necesario en otros niveles de abstracción, que van desde los nodos hasta el Centro de Datos, por medio de técnicas de balanceo de carga. Como resumen, el trabajo presentado en esta tesis lleva a cabo contribuciones en el modelado y optimización consciente del consumo por fugas y la refrigeración de servidores; el modelado de los Centros de Datos y el desarrollo de políticas de asignación conscientes de la heterogeneidad; y desarrolla mecanismos para la optimización energética de aplicaciones de nueva generación desde varios niveles de abstracción. ABSTRACT Data centers are easily found in every sector of the worldwide economy. They consist of tens of thousands of servers, serving millions of users globally and 24-7. In the last years, e-Science applications such e-Health or Smart Cities have experienced a significant development. The need to deal efficiently with the computational needs of next-generation applications together with the increasing demand for higher resources in traditional applications has facilitated the rapid proliferation and growing of data centers. A drawback to this capacity growth has been the rapid increase of the energy consumption of these facilities. In 2010, data center electricity represented 1.3% of all the electricity use in the world. In year 2012 alone, global data center power demand grew 63% to 38GW. A further rise of 17% to 43GW was estimated in 2013. Moreover, data centers are responsible for more than 2% of total carbon dioxide emissions. This PhD Thesis addresses the energy challenge by proposing proactive and reactive thermal and energy-aware optimization techniques that contribute to place data centers on a more scalable curve. This work develops energy models and uses the knowledge about the energy demand of the workload to be executed and the computational and cooling resources available at data center to optimize energy consumption. Moreover, data centers are considered as a crucial element within their application framework, optimizing not only the energy consumption of the facility, but the global energy consumption of the application. The main contributors to the energy consumption in a data center are the computing power drawn by IT equipment and the cooling power needed to keep the servers within a certain temperature range that ensures safe operation. Because of the cubic relation of fan power with fan speed, solutions based on over-provisioning cold air into the server usually lead to inefficiencies. On the other hand, higher chip temperatures lead to higher leakage power because of the exponential dependence of leakage on temperature. Moreover, workload characteristics as well as allocation policies also have an important impact on the leakage-cooling tradeoffs. The first key contribution of this work is the development of power and temperature models that accurately describe the leakage-cooling tradeoffs at the server level, and the proposal of strategies to minimize server energy via joint cooling and workload management from a multivariate perspective. When scaling to the data center level, a similar behavior in terms of leakage-temperature tradeoffs can be observed. As room temperature raises, the efficiency of data room cooling units improves. However, as we increase room temperature, CPU temperature raises and so does leakage power. Moreover, the thermal dynamics of a data room exhibit unbalanced patterns due to both the workload allocation and the heterogeneity of computing equipment. The second main contribution is the proposal of thermal- and heterogeneity-aware workload management techniques that jointly optimize the allocation of computation and cooling to servers. These strategies need to be backed up by flexible room level models, able to work on runtime, that describe the system from a high level perspective. Within the framework of next-generation applications, decisions taken at this scope can have a dramatical impact on the energy consumption of lower abstraction levels, i.e. the data center facility. It is important to consider the relationships between all the computational agents involved in the problem, so that they can cooperate to achieve the common goal of reducing energy in the overall system. The third main contribution is the energy optimization of the overall application by evaluating the energy costs of performing part of the processing in any of the different abstraction layers, from the node to the data center, via workload management and off-loading techniques. In summary, the work presented in this PhD Thesis, makes contributions on leakage and cooling aware server modeling and optimization, data center thermal modeling and heterogeneityaware data center resource allocation, and develops mechanisms for the energy optimization for next-generation applications from a multi-layer perspective.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Biopolymers do not have competitive prices, which has prevented their industrial exploitation on a global scale so far. In this context, Using nanoclays, improvements in certain biopolymer properties, mainly mechanical and thermal, have been achieved. However, research has been much less focused on changing optical properties through the incorporation of nanoclays. At the same time, current research has focused on obtaining nanopigments, by organic dyes adsoptions into different nanoclays in order to achieve sustainable colouring and high performance materials. By combining advances in these lines of research, biodegradable composites with optimal mechanical and optical properties can be obtained. The aim of this work is to find the optimal formulation of naturally sourced nanopigments, incorporate them into a biological origin epoxy resin, and obtain a significant improvement in their mechanical, and optical properties. We combine three structural modifiers in the nanopigment synthesis: surfactant, silane and mordant salt. The latter was selected in order to replicate the mordant textile dyeing with natural dyes. Using a Taguchi’s desing L8, we look for the effect of the presence of the modifiers, the pH acidification, and the interactions effect between the synthesis factors. Three natural dyes were selected: chlorophyll, beta-carotene, and beetroot extract. Furthermore we use two kinds of laminar nanoclays, differentiated by the ion exchange charge: montmorillonite, and hydrotalcite. Then the thermal, mechanical and colorimetric characterization of the bionanocomposite materials was carried out. The optimal conditions to obtain the best bionanocomposite materials are using acid pH, and modifying the nanoclays with mordant and surfactant.