886 resultados para Systems Simulation


Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the advent of cloud computing model, distributed caches have become the cornerstone for building scalable applications. Popular systems like Facebook [1] or Twitter use Memcached [5], a highly scalable distributed object cache, to speed up applications by avoiding database accesses. Distributed object caches assign objects to cache instances based on a hashing function, and objects are not moved from a cache instance to another unless more instances are added to the cache and objects are redistributed. This may lead to situations where some cache instances are overloaded when some of the objects they store are frequently accessed, while other cache instances are less frequently used. In this paper we propose a multi-resource load balancing algorithm for distributed cache systems. The algorithm aims at balancing both CPU and Memory resources among cache instances by redistributing stored data. Considering the possible conflict of balancing multiple resources at the same time, we give CPU and Memory resources weighted priorities based on the runtime load distributions. A scarcer resource is given a higher weight than a less scarce resource when load balancing. The system imbalance degree is evaluated based on monitoring information, and the utility load of a node, a unit for resource consumption. Besides, since continuous rebalance of the system may affect the QoS of applications utilizing the cache system, our data selection policy ensures that each data migration minimizes the system imbalance degree and hence, the total reconfiguration cost can be minimized. An extensive simulation is conducted to compare our policy with other policies. Our policy shows a significant improvement in time efficiency and decrease in reconfiguration cost.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper a new method for fault isolation in a class of continuous-time stochastic dynamical systems is proposed. The method is framed in the context of model-based analytical redundancy, consisting in the generation of a residual signal by means of a diagnostic observer, for its posterior analysis. Once a fault has been detected, and assuming some basic a priori knowledge about the set of possible failures in the plant, the isolation task is then formulated as a type of on-line statistical classification problem. The proposed isolation scheme employs in parallel different hypotheses tests on a statistic of the residual signal, one test for each possible fault. This isolation method is characterized by deriving for the unidimensional case, a sufficient isolability condition as well as an upperbound of the probability of missed isolation. Simulation examples illustrate the applicability of the proposed scheme.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computational homogenization by means of the finite element analysis of a representative volume element of the microstructure is used to simulate the deformation of nanostructured Ti. The behavior of each grain is taken into account using a single crystal elasto-viscoplastic model which includes the microscopic mechanisms of plastic deformation by slip along basal, prismatic and pyramidal systems. Two different representations of the polycrystal were used. Each grain was modeled with one cubic finite element in the first one while many cubic elements were used to represent each grain in the second one, leading to a model which includes the effect of grain shape and size in a limited number of grains due to the computational cost. Both representations were used to simulate the tensile deformation of nanostructured Ti processed by ECAP-C as well as the drawing process of nanostructured Ti billets. It was found that the first representation based in one finite element per grain led to a stiffer response in tension and was not able to predict the texture evolution during drawing because the strain gradient within each grain could not be captured. On the contrary, the second representation of the polycrystal microstructure with many finite elements per grain was able to predict accurately the deformation of nanostructured Ti.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Purpose – The purpose of this paper is to present a simulation‐based evaluation method for the comparison of different organizational forms and software support levels in the field of supply chain management (SCM). Design/methodology/approach – Apart from widely known logistic performance indicators, the discrete event simulation model considers explicitly coordination cost as stemming from iterative administration procedures. Findings - The method is applied to an exemplary supply chain configuration considering various parameter settings. Curiously, additional coordination cost does not always result in improved logistic performance. Influence factor variations lead to different organizational recommendations. The results confirm the high importance of (up to now) disregarded dimensions when evaluating SCM concepts and IT tools. Research limitations/implications – The model is based on simplified product and network structures. Future research shall include more complex, real world configurations. Practical implications – The developed method is designed for the identification of improvement potential when SCM software is employed. Coordination schemes based only on ERP systems are valid alternatives in industrial practice because significant investment IT can be avoided. Therefore, the evaluation of these coordination procedures, in particular the cost due to iterations, is of high managerial interest and the method provides a comprehensive tool for strategic IT decision making. Originality/value – Reviewed literature is mostly focused on the benefits of SCM software implementations. However, ERP system based supply chain coordination is still widespread industrial practice but associated coordination cost has not been addressed by researchers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Electrical Protection systems and Automatic Voltage Regulators (AVR) are essential components of actual power plants. Its installation and setting is performed during the commissioning, and it needs extensive experience since any failure in this process or in the setting, may entails some risk not only for the generator of the power plant, but also for the reliability of the power grid. In this paper, a real time power plant simulation platform is presented as a tool for improving the training and learning process on electrical protections and automatic voltage regulators. The activities of the commissioning procedure which can be practiced are described, and the applicability of this tool for improving the comprehension of this important part of the power plants is discussed. A commercial AVR and a multifunction protective relay have been tested with satisfactory results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The SESAR (Single European Sky ATM Research) program is an ambitious re-search and development initiative to design the future European air traffic man-agement (ATM) system. The study of the behavior of ATM systems using agent-based modeling and simulation tools can help the development of new methods to improve their performance. This paper presents an overview of existing agent-based approaches in air transportation (paying special attention to the challenges that exist for the design of future ATM systems) and, subsequently, describes a new agent-based approach that we proposed in the CASSIOPEIA project, which was developed according to the goals of the SESAR program. In our approach, we use agent models for different ATM stakeholders, and, in contrast to previous work, our solution models new collaborative decision processes for flow traffic management, it uses an intermediate level of abstraction (useful for simulations at larger scales), and was designed to be a practical tool (open and reusable) for the development of different ATM studies. It was successfully applied in three stud-ies related to the design of future ATM systems in Europe.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Foliage Penetration (FOPEN) radar systems were introduced in 1960, and have been constantly improved by several organizations since that time. The use of Synthetic Aperture Radar (SAR) approaches for this application has important advantages, due to the need for high resolution in two dimensions. The design of this type of systems, however, includes some complications that are not present in standard SAR systems. FOPEN SAR systems need to operate with a low central frequency (VHF or UHF bands) in order to be able to penetrate the foliage. High bandwidth is also required to obtain high resolution. Due to the low central frequency, large integration angles are required during SAR image formation, and therefore the Range Migration Algorithm (RMA) is used. This project thesis identifies the three main complications that arise due to these requirements. First, a high fractional bandwidth makes narrowband propagation models no longer valid. Second, the VHF and UHF bands are used by many communications systems. The transmitted signal spectrum needs to be notched to avoid interfering them. Third, those communications systems cause Radio Frequency Interference (RFI) on the received signal. The thesis carries out a thorough analysis of the three problems, their degrading effects and possible solutions to compensate them. The UWB model is applied to the SAR signal, and the degradation induced by it is derived. The result is tested through simulation of both a single pulse stretch processor and the complete RMA image formation. Both methods show that the degradation is negligible, and therefore the UWB propagation effect does not need compensation. A technique is derived to design a notched transmitted signal. Then, its effect on the SAR image formation is evaluated analytically. It is shown that the stretch processor introduces a processing gain that reduces the degrading effects of the notches. The remaining degrading effect after processing gain is assessed through simulation, and an experimental graph of degradation as a function of percentage of nulled frequencies is obtained. The RFI is characterized and its effect on the SAR processor is derived. Once again, a processing gain is found to be introduced by the receiver. As the RFI power can be much higher than that of the desired signal, an algorithm is proposed to remove the RFI from the received signal before RMA processing. This algorithm is a modification of the Chirp Least Squares Algorithm (CLSA) explained in [4], which adapts it to deramped signals. The algorithm is derived analytically and then its performance is evaluated through simulation, showing that it is effective in removing the RFI and reducing the degradation caused by both RFI and notching. Finally, conclusions are drawn as to the importance of each one of the problems in SAR system design.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Water balance simulation in cropping systems is a very useful tool to study how water can be used efficiently. However this requires that models simulate an accurate water balance. Comparing model results with field observations will provide information on the performance of the models. The objective of this study was to test the performance of DSSAT model in simulating the water balance by comparing the simulations with observed measurements. The soil water balance in DSSAT uses a one dimensional ?tipping bucket? soil water balance approach where available soil water is determined by the drained upper limit (DUL), lower limit (LL) and saturated water content (SAT). A continuous weighing lysimeter was used to get the observed values of drainage and evapotranspiration (ET). An automated agrometeorological weather station close to the lisymeter was also used to record the climatic data. The model simulated accurately the soil water content after the optimization of the soil parameters. However it was found the inability of the model to capture small changes in daily drainage and ET. For that reason simulated cumulative values had larger errors as the time passed by. These results suggested the need to compare outputs of DSSAT and some hydrological model that simulates soil water movement with a more mechanistic approach. The comparison of the two models will allow us to find which mechanism can be modified or incorporated in DSSAT model to improve the simulations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Hybrid Stepper Motors are widely used in open-loop position applications. They are the choice of actuation for the collimators in the Large Hadron Collider, the largest particle accelerator at CERN. In this case the positioning requirements and the highly radioactive operating environment are unique. The latter forces both the use of long cables to connect the motors to the drives which act as transmission lines and also prevents the use of standard position sensors. However, reliable and precise operation of the collimators is critical for the machine, requiring the prevention of step loss in the motors and maintenance to be foreseen in case of mechanical degradation. In order to make the above possible, an approach is proposed for the application of an Extended Kalman Filter to a sensorless stepper motor drive, when the motor is separated from its drive by long cables. When the long cables and high frequency pulse width modulated control voltage signals are used together, the electrical signals difer greatly between the motor and drive-side of the cable. Since in the considered case only drive-side data is available, it is therefore necessary to estimate the motor-side signals. Modelling the entire cable and motor system in an Extended Kalman Filter is too computationally intensive for standard embedded real-time platforms. It is, in consequence, proposed to divide the problem into an Extended Kalman Filter, based only on the motor model, and separated motor-side signal estimators, the combination of which is less demanding computationally. The efectiveness of this approach is shown in simulation. Then its validity is experimentally demonstrated via implementation in a DSP based drive. A testbench to test its performance when driving an axis of a Large Hadron Collider collimator is presented along with the results achieved. It is shown that the proposed method is capable of achieving position and load torque estimates which allow step loss to be detected and mechanical degradation to be evaluated without the need for physical sensors. These estimation algorithms often require a precise model of the motor, but the standard electrical model used for hybrid stepper motors is limited when currents, which are high enough to produce saturation of the magnetic circuit, are present. New model extensions are proposed in order to have a more precise model of the motor independently of the current level, whilst maintaining a low computational cost. It is shown that a significant improvement in the model It is achieved with these extensions, and their computational performance is compared to study the cost of model improvement versus computation cost. The applicability of the proposed model extensions is demonstrated via their use in an Extended Kalman Filter running in real-time for closed-loop current control and mechanical state estimation. An additional problem arises from the use of stepper motors. The mechanics of the collimators can wear due to the abrupt motion and torque profiles that are applied by them when used in the standard way, i.e. stepping in open-loop. Closed-loop position control, more specifically Field Oriented Control, would allow smoother profiles, more respectful to the mechanics, to be applied but requires position feedback. As mentioned already, the use of sensors in radioactive environments is very limited for reliability reasons. Sensorless control is a known option but when the speed is very low or zero, as is the case most of the time for the motors used in the LHC collimator, the loss of observability prevents its use. In order to allow the use of position sensors without reducing the long term reliability of the whole system, the possibility to switch from closed to open loop is proposed and validated, allowing the use of closed-loop control when the position sensors function correctly and open-loop when there is a sensor failure. A different approach to deal with the switched drive working with long cables is also presented. Switched mode stepper motor drives tend to have poor performance or even fail completely when the motor is fed through a long cable due to the high oscillations in the drive-side current. The design of a stepper motor output fillter which solves this problem is thus proposed. A two stage filter, one devoted to dealing with the diferential mode and the other with the common mode, is designed and validated experimentally. With this ?lter the drive performance is greatly improved, achieving a positioning repeatability even better than with the drive working without a long cable, the radiated emissions are reduced and the overvoltages at the motor terminals are eliminated.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The primary hypothesis stated by this paper is that the use of social choice theory in Ambient Intelligence systems can improve significantly users satisfaction when accessing shared resources. A research methodology based on agent based social simulations is employed to support this hypothesis and to evaluate these benefits. The result is a six-fold contribution summarized as follows. Firstly, several considerable differences between this application case and the most prominent social choice application, political elections, have been found and described. Secondly, given these differences, a number of metrics to evaluate different voting systems in this scope have been proposed and formalized. Thirdly, given the presented application and the metrics proposed, the performance of a number of well known electoral systems is compared. Fourthly, as a result of the performance study, a novel voting algorithm capable of obtaining the best balance between the metrics reviewed is introduced. Fifthly, to improve the social welfare in the experiments, the voting methods are combined with cluster analysis techniques. Finally, the article is complemented by a free and open-source tool, VoteSim, which ensures not only the reproducibility of the experimental results presented, but also allows the interested reader to adapt the case study presented to different environments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Los dispositivos móviles modernos disponen cada vez de más funcionalidad debido al rápido avance de las tecnologías de las comunicaciones y computaciones móviles. Sin embargo, la capacidad de la batería no ha experimentado un aumento equivalente. Por ello, la experiencia de usuario en los sistemas móviles modernos se ve muy afectada por la vida de la batería, que es un factor inestable de difícil de control. Para abordar este problema, investigaciones anteriores han propuesto un esquema de gestion del consumo (PM) centrada en la energía y que proporciona una garantía sobre la vida operativa de la batería mediante la gestión de la energía como un recurso de primera clase en el sistema. Como el planificador juega un papel fundamental en la administración del consumo de energía y en la garantía del rendimiento de las aplicaciones, esta tesis explora la optimización de la experiencia de usuario para sistemas móviles con energía limitada desde la perspectiva de un planificador que tiene en cuenta el consumo de energía en un contexto en el que ésta es un recurso de primera clase. En esta tesis se analiza en primer lugar los factores que contribuyen de forma general a la experiencia de usuario en un sistema móvil. Después se determinan los requisitos esenciales que afectan a la experiencia de usuario en la planificación centrada en el consumo de energía, que son el reparto proporcional de la potencia, el cumplimiento de las restricciones temporales, y cuando sea necesario, el compromiso entre la cuota de potencia y las restricciones temporales. Para cumplir con los requisitos, el algoritmo clásico de fair queueing y su modelo de referencia se extienden desde los dominios de las comunicaciones y ancho de banda de CPU hacia el dominio de la energía, y en base a ésto, se propone el algoritmo energy-based fair queueing (EFQ) para proporcionar una planificación basada en la energía. El algoritmo EFQ está diseñado para compartir la potencia consumida entre las tareas mediante su planificación en función de la energía consumida y de la cuota reservada. La cuota de consumo de cada tarea con restricciones temporales está protegida frente a diversos cambios que puedan ocurrir en el sistema. Además, para dar mejor soporte a las tareas en tiempo real y multimedia, se propone un mecanismo para combinar con el algoritmo EFQ para dar preferencia en la planificación durante breves intervalos de tiempo a las tareas más urgentes con restricciones temporales.Las propiedades del algoritmo EFQ se evaluan a través del modelado de alto nivel y la simulación. Los resultados de las simulaciones indican que los requisitos esenciales de la planificación centrada en la energía pueden lograrse. El algoritmo EFQ se implementa más tarde en el kernel de Linux. Para evaluar las propiedades del planificador EFQ basado en Linux, se desarrolló un banco de pruebas experimental basado en una sitema empotrado, un programa de banco de pruebas multihilo, y un conjunto de pruebas de código abierto. A través de experimentos específicamente diseñados, esta tesis verifica primero las propiedades de EFQ en la gestión de la cuota de consumo de potencia y la planificación en tiempo real y, a continuación, explora los beneficios potenciales de emplear la planificación EFQ en la optimización de la experiencia de usuario para sistemas móviles con energía limitada. Los resultados experimentales sobre la gestión de la cuota de energía muestran que EFQ es más eficaz que el planificador de Linux-CFS en la gestión de energía, logrando un reparto proporcional de la energía del sistema independientemente de en qué dispositivo se consume la energía. Los resultados experimentales en la planificación en tiempo real demuestran que EFQ puede lograr de forma eficaz, flexible y robusta el cumplimiento de las restricciones temporales aunque se dé el caso de aumento del el número de tareas o del error en la estimación de energía. Por último, un análisis comparativo de los resultados experimentales sobre la optimización de la experiencia del usuario demuestra que, primero, EFQ es más eficaz y flexible que los algoritmos tradicionales de planificación del procesador, como el que se encuentra por defecto en el planificador de Linux y, segundo, que proporciona la posibilidad de optimizar y preservar la experiencia de usuario para los sistemas móviles con energía limitada. Abstract Modern mobiledevices have been becoming increasingly powerful in functionality and entertainment as the next-generation mobile computing and communication technologies are rapidly advanced. However, the battery capacity has not experienced anequivalent increase. The user experience of modern mobile systems is therefore greatly affected by the battery lifetime,which is an unstable factor that is hard to control. To address this problem, previous works proposed energy-centric power management (PM) schemes to provide strong guarantee on the battery lifetime by globally managing energy as the first-class resource in the system. As the processor scheduler plays a pivotal role in power management and application performance guarantee, this thesis explores the user experience optimization of energy-limited mobile systemsfrom the perspective of energy-centric processor scheduling in an energy-centric context. This thesis first analyzes the general contributing factors of the mobile system user experience.Then itdetermines the essential requirements on the energy-centric processor scheduling for user experience optimization, which are proportional power sharing, time-constraint compliance, and when necessary, a tradeoff between the power share and the time-constraint compliance. To meet the requirements, the classical fair queuing algorithm and its reference model are extended from the network and CPU bandwidth sharing domain to the energy sharing domain, and based on that, the energy-based fair queuing (EFQ) algorithm is proposed for performing energy-centric processor scheduling. The EFQ algorithm is designed to provide proportional power shares to tasks by scheduling the tasks based on their energy consumption and weights. The power share of each time-sensitive task is protected upon the change of the scheduling environment to guarantee a stable performance, and any instantaneous power share that is overly allocated to one time-sensitive task can be fairly re-allocated to the other tasks. In addition, to better support real-time and multimedia scheduling, certain real-time friendly mechanism is combined into the EFQ algorithm to give time-limited scheduling preference to the time-sensitive tasks. Through high-level modelling and simulation, the properties of the EFQ algorithm are evaluated. The simulation results indicate that the essential requirements of energy-centric processor scheduling can be achieved. The EFQ algorithm is later implemented in the Linux kernel. To assess the properties of the Linux-based EFQ scheduler, an experimental test-bench based on an embedded platform, a multithreading test-bench program, and an open-source benchmark suite is developed. Through specifically-designed experiments, this thesis first verifies the properties of EFQ in power share management and real-time scheduling, and then, explores the potential benefits of employing EFQ scheduling in the user experience optimization for energy-limited mobile systems. Experimental results on power share management show that EFQ is more effective than the Linux-CFS scheduler in managing power shares and it can achieve a proportional sharing of the system power regardless of on which device the energy is spent. Experimental results on real-time scheduling demonstrate that EFQ can achieve effective, flexible and robust time-constraint compliance upon the increase of energy estimation error and task number. Finally, a comparative analysis of the experimental results on user experience optimization demonstrates that EFQ is more effective and flexible than traditional processor scheduling algorithms, such as those of the default Linux scheduler, in optimizing and preserving the user experience of energy-limited mobile systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El estudio de los ciclos del combustible nuclear requieren de herramientas computacionales o "códigos" versátiles para dar respuestas al problema multicriterio de evaluar los actuales ciclos o las capacidades de las diferentes estrategias y escenarios con potencial de desarrollo en a nivel nacional, regional o mundial. Por otra parte, la introducción de nuevas tecnologías para reactores y procesos industriales hace que los códigos existentes requieran nuevas capacidades para evaluar la transición del estado actual del ciclo del combustible hacia otros más avanzados y sostenibles. Brevemente, esta tesis se centra en dar respuesta a las principales preguntas, en términos económicos y de recursos, al análisis de escenarios de ciclos de combustible, en particular, para el análisis de los diferentes escenarios del ciclo del combustible de relativa importancia para España y Europa. Para alcanzar este objetivo ha sido necesaria la actualización y el desarrollo de nuevas capacidades del código TR_EVOL (Transition Evolution code). Este trabajo ha sido desarrollado en el Programa de Innovación Nuclear del CIEMAT desde el año 2010. Esta tesis se divide en 6 capítulos. El primer capítulo ofrece una visión general del ciclo de combustible nuclear, sus principales etapas y los diferentes tipos utilizados en la actualidad o en desarrollo para el futuro. Además, se describen las fuentes de material nuclear que podrían ser utilizadas como combustible (uranio y otros). También se puntualizan brevemente una serie de herramientas desarrolladas para el estudio de estos ciclos de combustible nuclear. El capítulo 2 está dirigido a dar una idea básica acerca de los costes involucrados en la generación de electricidad mediante energía nuclear. Aquí se presentan una clasificación de estos costos y sus estimaciones, obtenidas en la bibliografía, y que han sido evaluadas y utilizadas en esta tesis. Se ha incluido también una breve descripción del principal indicador económico utilizado en esta tesis, el “coste nivelado de la electricidad”. El capítulo 3 se centra en la descripción del código de simulación desarrollado para el estudio del ciclo del combustible nuclear, TR_EVOL, que ha sido diseñado para evaluar diferentes opciones de ciclos de combustibles. En particular, pueden ser evaluados las diversos reactores con, posiblemente, diferentes tipos de combustibles y sus instalaciones del ciclo asociadas. El módulo de evaluaciones económica de TR_EVOL ofrece el coste nivelado de la electricidad haciendo uso de las cuatro fuentes principales de información económica y de la salida del balance de masas obtenido de la simulación del ciclo en TR_EVOL. Por otra parte, la estimación de las incertidumbres en los costes también puede ser efectuada por el código. Se ha efectuado un proceso de comprobación cruzada de las funcionalidades del código y se descrine en el Capítulo 4. El proceso se ha aplicado en cuatro etapas de acuerdo con las características más importantes de TR_EVOL, balance de masas, composición isotópica y análisis económico. Así, la primera etapa ha consistido en el balance masas del ciclo de combustible nuclear actual de España. La segunda etapa se ha centrado en la comprobación de la composición isotópica del flujo de masas mediante el la simulación del ciclo del combustible definido en el proyecto CP-ESFR UE. Las dos últimas etapas han tenido como objetivo validar el módulo económico. De este modo, en la tercera etapa han sido evaluados los tres principales costes (financieros, operación y mantenimiento y de combustible) y comparados con los obtenidos por el proyecto ARCAS, omitiendo los costes del fin del ciclo o Back-end, los que han sido evaluado solo en la cuarta etapa, haciendo uso de costes unitarios y parámetros obtenidos a partir de la bibliografía. En el capítulo 5 se analizan dos grupos de opciones del ciclo del combustible nuclear de relevante importancia, en términos económicos y de recursos, para España y Europa. Para el caso español, se han simulado dos grupos de escenarios del ciclo del combustible, incluyendo estrategias de reproceso y extensión de vida de los reactores. Este análisis se ha centrado en explorar las ventajas y desventajas de reprocesado de combustible irradiado en un país con una “relativa” pequeña cantidad de reactores nucleares. Para el grupo de Europa se han tratado cuatro escenarios, incluyendo opciones de transmutación. Los escenarios incluyen los reactores actuales utilizando la tecnología reactor de agua ligera y ciclo abierto, un reemplazo total de los reactores actuales con reactores rápidos que queman combustible U-Pu MOX y dos escenarios del ciclo del combustible con transmutación de actínidos minoritarios en una parte de los reactores rápidos o en sistemas impulsados por aceleradores dedicados a transmutación. Finalmente, el capítulo 6 da las principales conclusiones obtenidas de esta tesis y los trabajos futuros previstos en el campo del análisis de ciclos de combustible nuclear. ABSTRACT The study of the nuclear fuel cycle requires versatile computational tools or “codes” to provide answers to the multicriteria problem of assessing current nuclear fuel cycles or the capabilities of different strategies and scenarios with potential development in a country, region or at world level. Moreover, the introduction of new technologies for reactors and industrial processes makes the existing codes to require new capabilities to assess the transition from current status of the fuel cycle to the more advanced and sustainable ones. Briefly, this thesis is focused in providing answers to the main questions about resources and economics in fuel cycle scenario analyses, in particular for the analysis of different fuel cycle scenarios with relative importance for Spain and Europe. The upgrade and development of new capabilities of the TR_EVOL code (Transition Evolution code) has been necessary to achieve this goal. This work has been developed in the Nuclear Innovation Program at CIEMAT since year 2010. This thesis is divided in 6 chapters. The first one gives an overview of the nuclear fuel cycle, its main stages and types currently used or in development for the future. Besides the sources of nuclear material that could be used as fuel (uranium and others) are also viewed here. A number of tools developed for the study of these nuclear fuel cycles are also briefly described in this chapter. Chapter 2 is aimed to give a basic idea about the cost involved in the electricity generation by means of the nuclear energy. The main classification of these costs and their estimations given by bibliography, which have been evaluated in this thesis, are presented. A brief description of the Levelized Cost of Electricity, the principal economic indicator used in this thesis, has been also included. Chapter 3 is focused on the description of the simulation tool TR_EVOL developed for the study of the nuclear fuel cycle. TR_EVOL has been designed to evaluate different options for the fuel cycle scenario. In particular, diverse nuclear power plants, having possibly different types of fuels and the associated fuel cycle facilities can be assessed. The TR_EVOL module for economic assessments provides the Levelized Cost of Electricity making use of the TR_EVOL mass balance output and four main sources of economic information. Furthermore, uncertainties assessment can be also carried out by the code. A cross checking process of the performance of the code has been accomplished and it is shown in Chapter 4. The process has been applied in four stages according to the most important features of TR_EVOL. Thus, the first stage has involved the mass balance of the current Spanish nuclear fuel cycle. The second stage has been focused in the isotopic composition of the mass flow using the fuel cycle defined in the EU project CP-ESFR. The last two stages have been aimed to validate the economic module. In the third stage, the main three generation costs (financial cost, O&M and fuel cost) have been assessed and compared to those obtained by ARCAS project, omitting the back-end costs. This last cost has been evaluated alone in the fourth stage, making use of some unit cost and parameters obtained from the bibliography. In Chapter 5 two groups of nuclear fuel cycle options with relevant importance for Spain and Europe are analyzed in economic and resources terms. For the Spanish case, two groups of fuel cycle scenarios have been simulated including reprocessing strategies and life extension of the current reactor fleet. This analysis has been focused on exploring the advantages and disadvantages of spent fuel reprocessing in a country with relatively small amount of nuclear power plants. For the European group, four fuel cycle scenarios involving transmutation options have been addressed. Scenarios include the current fleet using Light Water Reactor technology and open fuel cycle, a full replacement of the initial fleet with Fast Reactors burning U-Pu MOX fuel and two fuel cycle scenarios with Minor Actinide transmutation in a fraction of the FR fleet or in dedicated Accelerator Driven Systems. Finally, Chapter 6 gives the main conclusions obtained from this thesis and the future work foreseen in the field of nuclear fuel cycle analysis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a new simulation environment aimed at heterogeneous chained modular robots. This simulator allows testing the feasibility of the design, checking how modules are going to perform in the field and verifying hardware, electronics and communication designs before the prototype is built, saving time and resources. The paper shows how the simulator is built and how it can be set up to adapt to new designs. It also gives some examples of its use showing different heterogeneous modular robots running in different environments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Optical filters are crucial elements in optical communications. The influence of cascaded filters in the optical signal will affect the communications quality seriously. In this paper we will study and simulate the optical signal impairment caused by different kinds of filters which include Butterworth, Bessel, Fiber Bragg Grating (FBG) and Fabry-Perot (FP). Optical signal impairment is analyzed from an Eye Opening Penalty (EOP) and optical spectrum point of view. The simulation results show that when the center frequency of all filters aligns with the laser’s frequency, the Butterworth has the smallest influence to the signal while the F-P has the biggest. With a -1dB EOP, the amount of cascaded Butterworth optical filters with a bandwidth of 50 GHz is 18 in 40 Gbps NRZ-DQPSK systems and 12 in 100 Gbps PMNRZ- DQPSK systems. The value is reduced to 9 and 6 respectively for Febry-Perot optical filters. In the situation of frequency misalignment, the impairment caused by filters is more serious. Our research shows that with a frequency deviation of 5 GHz, only 12 and 9 Butterworth optical filters can be cascaded in 40 Gbps NRZ-DQPSK and 100 Gbps PM-NRZ-DQPSK systems respectively. We also study the signal impairment caused by different orders of the Butterworth filter model. Our study shows that although the higher-order has a smaller clipping effect in the transmission spectrum, it will introduce a more serious phase ripple which seriously affects the signal. Simulation result shows that the 2nd order Butterworth filter has the best performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Optical filters are crucial elements in optical communication networks. Their influence toward the optical signal will affect the communication quality seriously. In this paper we will study and simulate the optical signal impairment and crosstalk penalty caused by different kinds of filters, which include Butterworth, Bessel, Fiber Bragg Grating (FBG) and Fabry-Perot (F-P). Signal impairment from filter concatenation effect and crosstalk penalty from out-band and in-band are analyzed from Q-penalty, eye opening penalty (EOP) and optical spectrum. The simulation results show that signal impairment and crosstalk penalty induced by the Butterworth filter is the minimum among these four types of filters. Signal impairment caused by filter concatenation effect shows that when center frequency of all filters is aligned perfectly with the laser's frequency, 12 50-GHz Butterworth filters can be cascaded, with 1-dB EOP. This value is reduced to 9 when the center frequency is misaligned with 5 GHz. In the 50-GHz channel spacing DWDM networks, total Q-penalty induced by a pair of Butterworth filters based demultiplexer and multiplexer is lower than 0.5 dB when the filter bandwidth is in the range of 42-46 GHz.