997 resultados para reactor safety experiments


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The dissertation titled "Driver Safety in Far-side and Far-oblique Crashes" presents a novel approach to assessing vehicle cockpit safety by integrating Human Factors and Applied Mechanics. The methodology of this approach is aimed at improving safety in compact mobile workspaces such as patrol vehicle cockpits. A statistical analysis performed using Michigan state's traffic crash data to assess various contributing factors that affect the risk of severe driver injuries showed that the risk was greater for unrestrained drivers (OR=3.38, p<0.0001) and for incidents involving front and far-side crashes without seatbelts (OR=8.0 and 23.0 respectively, p<0.005). Statistics also showed that near-side and far-side crashes pose similar threat to driver injury severity. A Human Factor survey was conducted to assess various Human-Machine/Human-Computer Interaction aspects in patrol vehicle cockpits. Results showed that tasks requiring manual operation, especially the usage of laptop, would require more attention and potentially cause more distraction. A vehicle survey conducted to evaluate ergonomics-related issues revealed that some of the equipment was in airbag deployment zones. In addition, experiments were conducted to assess the effects on driver distraction caused by changing the position of in-car accessories. A driving simulator study was conducted to mimic HMI/HCI in a patrol vehicle cockpit (20 subjects, average driving experience = 5.35 years, s.d. = 1.8). It was found that the mounting locations of manual tasks did not result in a significant change in response times. Visual displays resulted in response times less than 1.5sec. It can also be concluded that the manual task was equally distracting regardless of mounting positions (average response time was 15 secs). Average speeds and lane deviations did not show any significant results. Data from 13 full-scale sled tests conducted to simulate far-side impacts at 70 PDOF and 40 PDOF was used to analyze head injuries and HIC/AIS values. It was found that accelerations generated by the vehicle deceleration alone were high enough to cause AIS 3 - AIS 6 injuries. Pretensioners could mitigated injuries only in 40 PDOF (oblique) impacts but are useless in 70 PDOF impacts. Seat belts were ineffective in protecting the driver's head from injuries. Head would come in contact with the laptop during a far-oblique (40 PDOF) crash and far-side door for an angle-type crash (70 PDOF). Finite Element analysis head-laptop impact interaction showed that the contact velocity was the most crucial factor in causing a severe (and potentially fatal) head injury. Results indicate that no equipment may be mounted in driver trajectory envelopes. A very narrow band of space is left in patrol vehicles for installation of manual-task equipment to be both safe and ergonomic. In case of a contact, the material stiffness and damping properties play a very significant role in determining the injury outcome. Future work may be done on improving the interiors' material properties to better absorb and dissipate kinetic energy of the head. The design of seat belts and pretensioners may also be seen as an essential aspect to be further improved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This article addresses the issue of kriging-based optimization of stochastic simulators. Many of these simulators depend on factors that tune the level of precision of the response, the gain in accuracy being at a price of computational time. The contribution of this work is two-fold: first, we propose a quantile-based criterion for the sequential design of experiments, in the fashion of the classical expected improvement criterion, which allows an elegant treatment of heterogeneous response precisions. Second, we present a procedure for the allocation of the computational time given to each measurement, allowing a better distribution of the computational effort and increased efficiency. Finally, the optimization method is applied to an original application in nuclear criticality safety. This article has supplementary material available online. The proposed criterion is available in the R package DiceOptim.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Since the Three Mile Island accident, an important focus of pressurized water reactor (PWR) transient analyses has been a small-break loss-of-coolant accident (SBLOCA). In 2002, the discovery of thinning of the vessel head wall at the Davis Besse nuclear power plant reactor indicated the possibility of an SBLOCA in the upper head of the reactor vessel as a result of circumferential cracking of a control rod drive mechanism penetration nozzle - which has cast even greater importance on the study of SBLOCAs. Several experimental tests have been performed at the Large Scale Test Facility to simulate the behavior of a PWR during an upper-head SBLOCA. The last of these tests, Organisation for Economic Co-operation and Development Nuclear Energy Agency Rig of Safety Assessment (OECD/NEA ROSA) Test 6.1, was performed in 2005. This test was simulated with the TRACE 5.0 code, and good agreement with the experimental results was obtained. Additionally, a broad analysis of an upper-head SBLOCA with high-pressure safety injection failed in a Westinghouse PWR was performed taking into account different accident management actions and conditions in order to check their suitability. This issue has been analyzed also in the framework of the OECD/NEA ROSA project and the Code Applications and Maintenance Program (CAMP). The main conclusion is that the current emergency operating procedures for Westinghouse reactor design are adequate for these kinds of sequences, and they do not need to be modified.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Determining as accurate as possible spent nuclear fuel isotopic content is gaining importance due to its safety and economic implications. Since nowadays higher burn ups are achievable through increasing initial enrichments, more efficient burn up strategies within the reactor cores and the extension of the irradiation periods, establishing and improving computation methodologies is mandatory in order to carry out reliable criticality and isotopic prediction calculations. Several codes (WIMSD5, SERPENT 1.1.7, SCALE 6.0, MONTEBURNS 2.0 and MCNP-ACAB) and methodologies are tested here and compared to consolidated benchmarks (OECD/NEA pin cell moderated with light water) with the purpose of validating them and reviewing the state of the isotopic prediction capabilities. These preliminary comparisons will suggest what can be generally expected of these codes when applied to real problems. In the present paper, SCALE 6.0 and MONTEBURNS 2.0 are used to model the same reported geometries, material compositions and burn up history of the Spanish Van de llós II reactor cycles 7-11 and to reproduce measured isotopies after irradiation and decay times. We analyze comparisons between measurements and each code results for several grades of geometrical modelization detail, using different libraries and cross-section treatment methodologies. The power and flux normalization method implemented in MONTEBURNS 2.0 is discussed and a new normalization strategy is developed to deal with the selected and similar problems, further options are included to reproduce temperature distributions of the materials within the fuel assemblies and it is introduced a new code to automate series of simulations and manage material information between them. In order to have a realistic confidence level in the prediction of spent fuel isotopic content, we have estimated uncertainties using our MCNP-ACAB system. This depletion code, which combines the neutron transport code MCNP and the inventory code ACAB, propagates the uncertainties in the nuclide inventory assessing the potential impact of uncertainties in the basic nuclear data: cross-section, decay data and fission yields

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Isotopic content assessment has a paramount importance for safety and storage reasons. During the latest years, a great variety of codes have been developed to perform transport and decay calculations, but only those that couple both in an iterative manner achieve an accurate prediction of the final isotopic content of irradiated fuels. Needless to say, them all are supposed to pass the test of the comparison of their predictions against the corresponding experimental measures.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Real time Tritium concentrations in air in two chemical forms, HT and HTO, coming from an ITER-like fusion reactor as source were coupled the European Centre Medium Range Weather Forecast (ECMWF) numerical model with the Lagrangian Atmospheric-particle dispersion model FLEXPART. This tool was analyzed in nominal tritium discharge operational reference and selected incidental conditions affecting the Western Mediterranean Basin during 45 days during summer 2010 together with surface “wind observations” or weather data based in real hourly observations of wind direction and velocity providing a real approximation of the tritium behavior after the release to the atmosphere from a fusion reactor. From comparison with NORMTRI - a code using climatologically sequences as input - over the same area, the real time results have demonstrated an apparent overestimation of the corresponding climatologically sequence of Tritium concentrations in air outputs, at several distances from the reactor. For this purpose two development patterns were established. The first one was following a cyclonic circulation over the Mediterranean Sea and the second one was based on the plume delivered over the Interior of the Iberian Peninsula and Continental Europe by another stabilized circulation corresponding to a High Pressure System. One of the important remaining activities defined then, was the qualification tool. In order to validate the model of ECMWF/FLEXPART we have developed of a new complete data base of tritium concentrations for the months from November 2010 to March 2011 and defined a new set of four patterns of HT transport in air, in each case using real boundary conditions: stationary to the North, stationary to the South, fast and very fast displacement. Finally the differences corresponding to those four early patterns (each one in assessments 1 and 2) has been analyzed in terms of the tuning of safety related issues and taking into account the primary phase o- - f tritium modeling, from its discharge to the atmosphere to the deposition on the ground, will affect to the complete tritium environmental pathway altering the chronic dose by absorption, reemission and ingestion both from elemental tritium, HT and from the oxide of tritium, HTO

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work is based on the prototype High Engineering Test Reactor (HTTR) of the Japan Agency of Energy Atomic (JAEA). Its objective is to describe an adequate deterministic model to be used in the assessment of its design safety margins via damage domains. The concept of damage domain is defined and it is shown its relevance in the ongoing effort to apply dynamic risk assessment methods and tools based on the Theory of Stimulated Dynamics (TSD). To illustrate, we present results of an abnormal control rod (CR) withdrawal during subcritical condition and its comparison with results obtained by JAEA. No attempt is made yet to actually assess the detailed scenarios, rather to show how the approach may handle events of its kind

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Proof carrying code (PCC) is a general is originally a roof in ñrst-order logic of certain vermethodology for certifying that the execution of an un- ification onditions and the checking process involves trusted mobile code is safe. The baste idea is that the ensuring that the certifícate is indeed a valid ñrst-order code supplier attaches a certifícate to the mobile code proof. which the consumer checks in order to ensure that the The main practical difñculty of PCC techniques is in code is indeed safe. The potential benefit is that the generating safety certiñeates which at the same time: i) consumer's task is reduced from the level of proving to allow expressing interesting safety properties, ii) can be the level of checking. Recently, the abstract interpre- generated automatically and, iii) are easy and efficient tation techniques developed, in logic programming have to check. In [1], the abstract interpretation techniques been proposed as a basis for PCC. This extended ab- [5] developed in logic programming1 are proposed as stract reports on experiments which illustrate several is- a basis for PCC. They offer a number of advantages sues involved in abstract interpretation-based certifica- for dealing with the aforementioned issues. In particution. First, we describe the implementation of our sys- lar, the xpressiveness of existing abstract domains will tem in the context of CiaoPP: the preprocessor of the be implicitly available in abstract interpretation-based Ciao multi-paradigm programming system. Then, by code certification to deñne a wide range of safety propermeans of some experiments, we show how code certifi- ties. Furthermore, the approach inherits the automation catión is aided in the implementation of the framework. and inference power of the abstract interpretation en- Finally, we discuss the application of our method within gines used in (Constraint) Logic Programming, (C)LP. the área, of pervasive systems

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Proof carrying code is a general methodology for certifying that the execution of an untrusted mobile code is safe, according to a predefined safety policy. The basic idea is that the code supplier attaches a certifícate (or proof) to the mobile code which, then, the consumer checks in order to ensure that the code is indeed safe. The potential benefit is that the consumer's task is reduced from the level of proving to the level of checking, a much simpler task. Recently, the abstract interpretation techniques developed in logic programming have been proposed as a basis for proof carrying code [1]. To this end, the certifícate is generated from an abstract interpretation-based proof of safety. Intuitively, the verification condition is extracted from a set of assertions guaranteeing safety and the answer table generated during the analysis. Given this information, it is relatively simple and fast to verify that the code does meet this proof and so its execution is safe. This extended abstract reports on experiments which illustrate several issues involved in abstract interpretation-based code certification. First, we describe the implementation of our system in the context of CiaoPP: the preprocessor of the Ciao multi-paradigm (constraint) logic programming system. Then, by means of some experiments, we show how code certification is aided in the implementation of the framework. Finally, we discuss the application of our method within the área of pervasive systems which may lack the necessary computing resources to verify safety on their own. We herein illustrate the relevance of the information inferred by existing cost analysis to control resource usage in this context. Moreover, since the (rather complex) analysis phase is replaced by a simpler, efficient checking process at the code consumer side, we believe that our abstract interpretation-based approach to proof-carrying code becomes practically applicable to this kind of systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El accidente de rotura de tubos de un generador de vapor (Steam Generator Tube Rupture, SGTR) en los reactores de agua a presión es uno de los transitorios más exigentes desde el punto de vista de operación. Los transitorios de SGTR son especiales, ya que podría dar lugar a emisiones radiológicas al exterior sin necesidad de daño en el núcleo previo o sin que falle la contención, ya que los SG pueden constituir una vía directa desde el reactor al medio ambiente en este transitorio. En los análisis de seguridad, el SGTR se analiza desde un punto determinista y probabilista, con distintos enfoques con respecto a las acciones del operador y las consecuencias analizadas. Cuando comenzaron los Análisis Deterministas de Seguridad (DSA), la forma de analizar el SGTR fue sin dar crédito a la acción del operador durante los primeros 30 min del transitorio, lo que suponía que el grupo de operación era capaz de detener la fuga por el tubo roto dentro de ese tiempo. Sin embargo, los diferentes casos reales de accidentes de SGTR sucedidos en los EE.UU. y alrededor del mundo demostraron que los operadores pueden emplear más de 30 minutos para detener la fuga en la vida real. Algunas metodologías fueron desarrolladas en los EEUU y en Europa para abordar esa cuestión. En el Análisis Probabilista de Seguridad (PSA), las acciones del operador se tienen en cuenta para diseñar los cabeceros en el árbol de sucesos. Los tiempos disponibles se utilizan para establecer los criterios de éxito para dichos cabeceros. Sin embargo, en una secuencia dinámica como el SGTR, las acciones de un operador son muy dependientes del tiempo disponible por las acciones humanas anteriores. Además, algunas de las secuencias de SGTR puede conducir a la liberación de actividad radiológica al exterior sin daño previo en el núcleo y que no se tienen en cuenta en el APS, ya que desde el punto de vista de la integridad de núcleo son de éxito. Para ello, para analizar todos estos factores, la forma adecuada de analizar este tipo de secuencias pueden ser a través de una metodología que contemple Árboles de Sucesos Dinámicos (Dynamic Event Trees, DET). En esta Tesis Doctoral se compara el impacto en la evolución temporal y la dosis al exterior de la hipótesis más relevantes encontradas en los Análisis Deterministas a nivel mundial. La comparación se realiza con un modelo PWR Westinghouse de tres lazos (CN Almaraz) con el código termohidráulico TRACE, con hipótesis de estimación óptima, pero con hipótesis deterministas como criterio de fallo único o pérdida de energía eléctrica exterior. Las dosis al exterior se calculan con RADTRAD, ya que es uno de los códigos utilizados normalmente para los cálculos de dosis del SGTR. El comportamiento del reactor y las dosis al exterior son muy diversas, según las diferentes hipótesis en cada metodología. Por otra parte, los resultados están bastante lejos de los límites de regulación, pese a los conservadurismos introducidos. En el siguiente paso de la Tesis Doctoral, se ha realizado un análisis de seguridad integrado del SGTR según la metodología ISA, desarrollada por el Consejo de Seguridad Nuclear español (CSN). Para ello, se ha realizado un análisis termo-hidráulico con un modelo de PWR Westinghouse de 3 lazos con el código MAAP. La metodología ISA permite la obtención del árbol de eventos dinámico del SGTR, teniendo en cuenta las incertidumbres en los tiempos de actuación del operador. Las simulaciones se realizaron con SCAIS (sistema de simulación de códigos para la evaluación de la seguridad integrada), que incluye un acoplamiento dinámico con MAAP. Las dosis al exterior se calcularon también con RADTRAD. En los resultados, se han tenido en cuenta, por primera vez en la literatura, las consecuencias de las secuencias en términos no sólo de daños en el núcleo sino de dosis al exterior. Esta tesis doctoral demuestra la necesidad de analizar todas las consecuencias que contribuyen al riesgo en un accidente como el SGTR. Para ello se ha hecho uso de una metodología integrada como ISA-CSN. Con este enfoque, la visión del DSA del SGTR (consecuencias radiológicas) se une con la visión del PSA del SGTR (consecuencias de daño al núcleo) para evaluar el riesgo total del accidente. Abstract Steam Generator Tube Rupture accidents in Pressurized Water Reactors are known to be one of the most demanding transients for the operating crew. SGTR are special transient as they could lead to radiological releases without core damage or containment failure, as they can constitute a direct path to the environment. The SGTR is analyzed from a Deterministic and Probabilistic point of view in the Safety Analysis, although the assumptions of the different approaches regarding the operator actions are quite different. In the beginning of Deterministic Safety Analysis, the way of analyzing the SGTR was not crediting the operator action for the first 30 min of the transient, assuming that the operating crew was able to stop the primary to secondary leakage within that time. However, the different real SGTR accident cases happened in the USA and over the world demonstrated that operators can took more than 30 min to stop the leakage in actual sequences. Some methodologies were raised in the USA and in Europe to cover that issue. In the Probabilistic Safety Analysis, the operator actions are taken into account to set the headers in the event tree. The available times are used to establish the success criteria for the headers. However, in such a dynamic sequence as SGTR, the operator actions are very dependent on the time available left by the other human actions. Moreover, some of the SGTR sequences can lead to offsite doses without previous core damage and they are not taken into account in PSA as from the point of view of core integrity are successful. Therefore, to analyze all this factors, the appropriate way of analyzing that kind of sequences could be through a Dynamic Event Tree methodology. This Thesis compares the impact on transient evolution and the offsite dose of the most relevant hypothesis of the different SGTR analysis included in the Deterministic Safety Analysis. The comparison is done with a PWR Westinghouse three loop model in TRACE code (Almaraz NPP), with best estimate assumptions but including deterministic hypothesis such as single failure criteria or loss of offsite power. The offsite doses are calculated with RADTRAD code, as it is one of the codes normally used for SGTR offsite dose calculations. The behaviour of the reactor and the offsite doses are quite diverse depending on the different assumptions made in each methodology. On the other hand, although the high conservatism, such as the single failure criteria, the results are quite far from the regulatory limits. In the next stage of the Thesis, the Integrated Safety Assessment (ISA) methodology, developed by the Spanish Nuclear Safety Council (CSN), has been applied to a thermohydraulical analysis of a Westinghouse 3-loop PWR plant with the MAAP code. The ISA methodology allows obtaining the SGTR Dynamic Event Tree taking into account the uncertainties on the operator actuation times. Simulations are performed with SCAIS (Simulation Code system for Integrated Safety Assessment), which includes a dynamic coupling with MAAP thermal hydraulic code. The offsite doses are calculated also with RADTRAD. The results shows the consequences of the sequences in terms not only of core damage but of offsite doses. This Thesis shows the need of analyzing all the consequences in an accident such as SGTR. For that, an it has been used an integral methodology like ISA-CSN. With this approach, the DSA vision of the SGTR (radiological consequences) is joined with the PSA vision of the SGTR (core damage consequences) to measure the total risk of the accident.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El futuro de la energía nuclear de fisión dependerá, entre otros factores, de la capacidad que las nuevas tecnologías demuestren para solventar los principales retos a largo plazo que se plantean. Los principales retos se pueden resumir en los siguientes aspectos: la capacidad de proporcionar una solución final, segura y fiable a los residuos radiactivos; así como dar solución a la limitación de recursos naturales necesarios para alimentar los reactores nucleares; y por último, una mejora robusta en la seguridad de las centrales que en definitiva evite cualquier daño potencial tanto en la población como en el medio ambiente como consecuencia de cualquier escenario imaginable o más allá de lo imaginable. Siguiendo estas motivaciones, la Generación IV de reactores nucleares surge con el compromiso de proporcionar electricidad de forma sostenible, segura, económica y evitando la proliferación de material fisible. Entre los sistemas conceptuales que se consideran para la Gen IV, los reactores rápidos destacan por su capacidad potencial de transmutar actínidos a la vez que permiten una utilización óptima de los recursos naturales. Entre los refrigerantes que se plantean, el sodio parece una de las soluciones más prometedoras. Como consecuencia, esta tesis surgió dentro del marco del proyecto europeo CP-ESFR con el principal objetivo de evaluar la física de núcleo y seguridad de los reactores rápidos refrigerados por sodio, al tiempo que se desarrollaron herramientas apropiadas para dichos análisis. Efectivamente, en una primera parte de la tesis, se abarca el estudio de la física del núcleo de un reactor rápido representativo, incluyendo el análisis detallado de la capacidad de transmutar actínidos minoritarios. Como resultado de dichos análisis, se publicó un artículo en la revista Annals of Nuclear Energy [96]. Por otra parte, a través de un análisis de un hipotético escenario nuclear español, se evalúo la disponibilidad de recursos naturales necesarios en el caso particular de España para alimentar una flota específica de reactores rápidos, siguiendo varios escenarios de demanda, y teniendo en cuenta la capacidad de reproducción de plutonio que tienen estos sistemas. Como resultado de este trabajo también surgió una publicación en otra revista científica de prestigio internacional como es Energy Conversion and Management [97]. Con objeto de realizar esos y otros análisis, se desarrollaron diversos modelos del núcleo del ESFR siguiendo varias configuraciones, y para diferentes códigos. Por otro lado, con objeto de poder realizar análisis de seguridad de reactores rápidos, son necesarias herramientas multidimensionales de alta fidelidad específicas para reactores rápidos. Dichas herramientas deben integrar fenómenos relacionados con la neutrónica y con la termo-hidráulica, entre otros, mediante una aproximación multi-física. Siguiendo este objetivo, se evalúo el código de difusión neutrónica ANDES para su aplicación a reactores rápidos. ANDES es un código de resolución nodal que se encuentra implementado dentro del sistema COBAYA3 y está basado en el método ACMFD. Por lo tanto, el método ACMFD fue sometido a una revisión en profundidad para evaluar su aptitud para la aplicación a reactores rápidos. Durante ese proceso, se identificaron determinadas limitaciones que se discutirán a lo largo de este trabajo, junto con los desarrollos que se han elaborado e implementado para la resolución de dichas dificultades. Por otra parte, se desarrolló satisfactoriamente el acomplamiento del código neutrónico ANDES con un código termo-hidráulico de subcanales llamado SUBCHANFLOW, desarrollado recientemente en el KIT. Como conclusión de esta parte, todos los desarrollos implementados son evaluados y verificados. En paralelo con esos desarrollos, se calcularon para el núcleo del ESFR las secciones eficaces en multigrupos homogeneizadas a nivel nodal, así como otros parámetros neutrónicos, mediante los códigos ERANOS, primero, y SERPENT, después. Dichos parámetros se utilizaron más adelante para realizar cálculos estacionarios con ANDES. Además, como consecuencia de la contribución de la UPM al paquete de seguridad del proyecto CP-ESFR, se calcularon mediante el código SERPENT los parámetros de cinética puntual que se necesitan introducir en los típicos códigos termo-hidráulicos de planta, para estudios de seguridad. En concreto, dichos parámetros sirvieron para el análisis del impacto que tienen los actínidos minoritarios en el comportamiento de transitorios. Concluyendo, la tesis presenta una aproximación sistemática y multidisciplinar aplicada al análisis de seguridad y comportamiento neutrónico de los reactores rápidos de sodio de la Gen-IV, usando herramientas de cálculo existentes y recién desarrolladas ad' hoc para tal aplicación. Se ha empleado una cantidad importante de tiempo en identificar limitaciones de los métodos nodales analíticos en su aplicación en multigrupos a reactores rápidos, y se proponen interesantes soluciones para abordarlas. ABSTRACT The future of nuclear reactors will depend, among other aspects, on the capability to solve the long-term challenges linked to this technology. These are the capability to provide a definite, safe and reliable solution to the nuclear wastes; the limitation of natural resources, needed to fuel the reactors; and last but not least, the improved safety, which would avoid any potential damage on the public and or environment as a consequence of any imaginable and beyond imaginable circumstance. Following these motivations, the IV Generation of nuclear reactors arises, with the aim to provide sustainable, safe, economic and proliferationresistant electricity. Among the systems considered for the Gen IV, fast reactors have a representative role thanks to their potential capacity to transmute actinides together with the optimal usage of natural resources, being the sodium fast reactors the most promising concept. As a consequence, this thesis was born in the framework of the CP-ESFR project with the generic aim of evaluating the core physics and safety of sodium fast reactors, as well as the development of the approppriated tools to perform such analyses. Indeed, in a first part of this thesis work, the main core physics of the representative sodium fast reactor are assessed, including a detailed analysis of the capability to transmute minor actinides. A part of the results obtained have been published in Annals of Nuclear Energy [96]. Moreover, by means of the analysis of a hypothetical Spanish nuclear scenario, the availability of natural resources required to deploy an specific fleet of fast reactor is assessed, taking into account the breeding properties of such systems. This work also led to a publication in Energy Conversion and Management [97]. In order to perform those and other analyses, several models of the ESFR core were created for different codes. On the other hand, in order to perform safety studies of sodium fast reactors, high fidelity multidimensional analysis tools for sodium fast reactors are required. Such tools should integrate neutronic and thermal-hydraulic phenomena in a multi-physics approach. Following this motivation, the neutron diffusion code ANDES is assessed for sodium fast reactor applications. ANDES is the nodal solver implemented inside the multigroup pin-by-pin diffusion COBAYA3 code, and is based on the analytical method ACMFD. Thus, the ACMFD was verified for SFR applications and while doing so, some limitations were encountered, which are discussed through this work. In order to solve those, some new developments are proposed and implemented in ANDES. Moreover, the code was satisfactorily coupled with the thermal-hydraulic code SUBCHANFLOW, recently developed at KIT. Finally, the different implementations are verified. In addition to those developments, the node homogenized multigroup cross sections and other neutron parameters were obtained for the ESFR core using ERANOS and SERPENT codes, and employed afterwards by ANDES to perform steady state calculations. Moreover, as a result of the UPM contribution to the safety package of the CP-ESFR project, the point kinetic parameters required by the typical plant thermal-hydraulic codes were computed for the ESFR core using SERPENT, which final aim was the assessment of the impact of minor actinides in transient behaviour. All in all, the thesis provides a systematic and multi-purpose approach applied to the assessment of safety and performance parameters of Generation-IV SFR, using existing and newly developed analytical tools. An important amount of time was employed in identifying the limitations that the analytical nodal diffusion methods present when applied to fast reactors following a multigroup approach, and interesting solutions are proposed in order to overcome them.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The assessment of the uncertainty levels on the design and safety parameters for the innovative European Sodium Fast Reactor (ESFR) is mandatory. Some of these relevant safety quantities are the Doppler and void reactivity coefficients, whose uncertainties are quantified. Besides, the nuclear reaction data where an improvement will certainly benefit the design accuracy are identified. This work has been performed with the SCALE 6.1 codes suite and its multigroups cross sections library based on ENDF/B-VII.0 evaluation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper describes an experimental procedure consisting of impact tests that simulate a collision of a human head with an industrial robot with the aim to validate a safety index named as New Index for Robots (NIR) and its outputs. The experiments in this paper are based on lab tests. It is an attempt to characterize the NIR index underlying the main parameters that are involved in crash interaction and to highlight limitations and weakness of suggested impact tests.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El agotamiento, la ausencia o, simplemente, la incertidumbre sobre la cantidad de las reservas de combustibles fósiles se añaden a la variabilidad de los precios y a la creciente inestabilidad en la cadena de aprovisionamiento para crear fuertes incentivos para el desarrollo de fuentes y vectores energéticos alternativos. El atractivo de hidrógeno como vector energético es muy alto en un contexto que abarca, además, fuertes inquietudes por parte de la población sobre la contaminación y las emisiones de gases de efecto invernadero. Debido a su excelente impacto ambiental, la aceptación pública del nuevo vector energético dependería, a priori, del control de los riesgos asociados su manipulación y almacenamiento. Entre estos, la existencia de un innegable riesgo de explosión aparece como el principal inconveniente de este combustible alternativo. Esta tesis investiga la modelización numérica de explosiones en grandes volúmenes, centrándose en la simulación de la combustión turbulenta en grandes dominios de cálculo en los que la resolución que es alcanzable está fuertemente limitada. En la introducción, se aborda una descripción general de los procesos de explosión. Se concluye que las restricciones en la resolución de los cálculos hacen necesario el modelado de los procesos de turbulencia y de combustión. Posteriormente, se realiza una revisión crítica de las metodologías disponibles tanto para turbulencia como para combustión, que se lleva a cabo señalando las fortalezas, deficiencias e idoneidad de cada una de las metodologías. Como conclusión de esta investigación, se obtiene que la única estrategia viable para el modelado de la combustión, teniendo en cuenta las limitaciones existentes, es la utilización de una expresión que describa la velocidad de combustión turbulenta en función de distintos parámetros. Este tipo de modelos se denominan Modelos de velocidad de llama turbulenta y permiten cerrar una ecuación de balance para la variable de progreso de combustión. Como conclusión también se ha obtenido, que la solución más adecuada para la simulación de la turbulencia es la utilización de diferentes metodologías para la simulación de la turbulencia, LES o RANS, en función de la geometría y de las restricciones en la resolución de cada problema particular. Sobre la base de estos hallazgos, el crea de un modelo de combustión en el marco de los modelos de velocidad de la llama turbulenta. La metodología propuesta es capaz de superar las deficiencias existentes en los modelos disponibles para aquellos problemas en los que se precisa realizar cálculos con una resolución moderada o baja. Particularmente, el modelo utiliza un algoritmo heurístico para impedir el crecimiento del espesor de la llama, una deficiencia que lastraba el célebre modelo de Zimont. Bajo este enfoque, el énfasis del análisis se centra en la determinación de la velocidad de combustión, tanto laminar como turbulenta. La velocidad de combustión laminar se determina a través de una nueva formulación capaz de tener en cuenta la influencia simultánea en la velocidad de combustión laminar de la relación de equivalencia, la temperatura, la presión y la dilución con vapor de agua. La formulación obtenida es válida para un dominio de temperaturas, presiones y dilución con vapor de agua más extenso de cualquiera de las formulaciones previamente disponibles. Por otra parte, el cálculo de la velocidad de combustión turbulenta puede ser abordado mediante el uso de correlaciones que permiten el la determinación de esta magnitud en función de distintos parámetros. Con el objetivo de seleccionar la formulación más adecuada, se ha realizado una comparación entre los resultados obtenidos con diversas expresiones y los resultados obtenidos en los experimentos. Se concluye que la ecuación debida a Schmidt es la más adecuada teniendo en cuenta las condiciones del estudio. A continuación, se analiza la importancia de las inestabilidades de la llama en la propagación de los frentes de combustión. Su relevancia resulta significativa para mezclas pobres en combustible en las que la intensidad de la turbulencia permanece moderada. Estas condiciones son importantes dado que son habituales en los accidentes que ocurren en las centrales nucleares. Por ello, se lleva a cabo la creación de un modelo que permita estimar el efecto de las inestabilidades, y en concreto de la inestabilidad acústica-paramétrica, en la velocidad de propagación de llama. El modelado incluye la derivación matemática de la formulación heurística de Bauwebs et al. para el cálculo de la incremento de la velocidad de combustión debido a las inestabilidades de la llama, así como el análisis de la estabilidad de las llamas con respecto a una perturbación cíclica. Por último, los resultados se combinan para concluir el modelado de la inestabilidad acústica-paramétrica. Tras finalizar esta fase, la investigación se centro en la aplicación del modelo desarrollado en varios problemas de importancia para la seguridad industrial y el posterior análisis de los resultados y la comparación de los mismos con los datos experimentales correspondientes. Concretamente, se abordo la simulación de explosiones en túneles y en contenedores, con y sin gradiente de concentración y ventilación. Como resultados generales, se logra validar el modelo confirmando su idoneidad para estos problemas. Como última tarea, se ha realizado un analisis en profundidad de la catástrofe de Fukushima-Daiichi. El objetivo del análisis es determinar la cantidad de hidrógeno que explotó en el reactor número uno, en contraste con los otros estudios sobre el tema que se han centrado en la determinación de la cantidad de hidrógeno generado durante el accidente. Como resultado de la investigación, se determinó que la cantidad más probable de hidrogeno que fue consumida durante la explosión fue de 130 kg. Es un hecho notable el que la combustión de una relativamente pequeña cantidad de hidrogeno pueda causar un daño tan significativo. Esta es una muestra de la importancia de este tipo de investigaciones. Las ramas de la industria para las que el modelo desarrollado será de interés abarca la totalidad de la futura economía de hidrógeno (pilas de combustible, vehículos, almacenamiento energético, etc) con un impacto especial en los sectores del transporte y la energía nuclear, tanto para las tecnologías de fisión y fusión. ABSTRACT The exhaustion, absolute absence or simply the uncertainty on the amount of the reserves of fossil fuels sources added to the variability of their prices and the increasing instability and difficulties on the supply chain are strong incentives for the development of alternative energy sources and carriers. The attractiveness of hydrogen in a context that additionally comprehends concerns on pollution and emissions is very high. Due to its excellent environmental impact, the public acceptance of the new energetic vector will depend on the risk associated to its handling and storage. Fromthese, the danger of a severe explosion appears as the major drawback of this alternative fuel. This thesis investigates the numerical modeling of large scale explosions, focusing on the simulation of turbulent combustion in large domains where the resolution achievable is forcefully limited. In the introduction, a general description of explosion process is undertaken. It is concluded that the restrictions of resolution makes necessary the modeling of the turbulence and combustion processes. Subsequently, a critical review of the available methodologies for both turbulence and combustion is carried out pointing out their strengths and deficiencies. As a conclusion of this investigation, it appears clear that the only viable methodology for combustion modeling is the utilization of an expression for the turbulent burning velocity to close a balance equation for the combustion progress variable, a model of the Turbulent flame velocity kind. Also, that depending on the particular resolution restriction of each problem and on its geometry the utilization of different simulation methodologies, LES or RANS, is the most adequate solution for modeling the turbulence. Based on these findings, the candidate undertakes the creation of a combustion model in the framework of turbulent flame speed methodology which is able to overcome the deficiencies of the available ones for low resolution problems. Particularly, the model utilizes a heuristic algorithm to maintain the thickness of the flame brush under control, a serious deficiency of the Zimont model. Under the approach utilized by the candidate, the emphasis of the analysis lays on the accurate determination of the burning velocity, both laminar and turbulent. On one side, the laminar burning velocity is determined through a newly developed correlation which is able to describe the simultaneous influence of the equivalence ratio, temperature, steam dilution and pressure on the laminar burning velocity. The formulation obtained is valid for a larger domain of temperature, steam dilution and pressure than any of the previously available formulations. On the other side, a certain number of turbulent burning velocity correlations are available in the literature. For the selection of the most suitable, they have been compared with experiments and ranked, with the outcome that the formulation due to Schmidt was the most adequate for the conditions studied. Subsequently, the role of the flame instabilities on the development of explosions is assessed. Their significance appears to be of importance for lean mixtures in which the turbulence intensity remains moderate. These are important conditions which are typical for accidents on Nuclear Power Plants. Therefore, the creation of a model to account for the instabilities, and concretely, the acoustic parametric instability is undertaken. This encloses the mathematical derivation of the heuristic formulation of Bauwebs et al. for the calculation of the burning velocity enhancement due to flame instabilities as well as the analysis of the stability of flames with respect to a cyclic velocity perturbation. The results are combined to build a model of the acoustic-parametric instability. The following task in this research has been to apply the model developed to several problems significant for the industrial safety and the subsequent analysis of the results and comparison with the corresponding experimental data was performed. As a part of such task simulations of explosions in a tunnel and explosions in large containers, with and without gradient of concentration and venting have been carried out. As a general outcome, the validation of the model is achieved, confirming its suitability for the problems addressed. As a last and final undertaking, a thorough study of the Fukushima-Daiichi catastrophe has been carried out. The analysis performed aims at the determination of the amount of hydrogen participating on the explosion that happened in the reactor one, in contrast with other analysis centered on the amount of hydrogen generated during the accident. As an outcome of the research, it was determined that the most probable amount of hydrogen exploding during the catastrophe was 130 kg. It is remarkable that the combustion of such a small quantity of material can cause tremendous damage. This is an indication of the importance of these types of investigations. The industrial branches that can benefit from the applications of the model developed in this thesis include the whole future hydrogen economy, as well as nuclear safety both in fusion and fission technology.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Esta tesis se centra en desarrollo de tecnologías para la interacción hombre-robot en entornos nucleares de fusión. La problemática principal del sector de fusión nuclear radica en las condiciones ambientales tan extremas que hay en el interior del reactor, y la necesidad de que los equipos cumplan requisitos muy restrictivos para poder aguantar esos niveles de radiación, magnetismo, ultravacío, temperatura... Como no es viable la ejecución de tareas directamente por parte de humanos, habrá que utilizar dispositivos de manipulación remota para llevar a cabo los procesos de operación y mantenimiento. En las instalaciones de ITER es obligatorio tener un entorno controlado de extrema seguridad, que necesita de estándares validados. La definición y uso de protocolos es indispensable para regir su buen funcionamiento. Si nos centramos en la telemanipulación con algo grado de escalado, surge la necesidad de definir protocolos para sistemas abiertos que permitan la interacción entre equipos y dispositivos de diversa índole. En este contexto se plantea la definición del Protocolo de Teleoperación que permita la interconexión entre dispositivos maestros y esclavos de distinta tipología, pudiéndose comunicar bilateralmente entre sí y utilizar distintos algoritmos de control según la tarea a desempeñar. Este protocolo y su interconectividad se han puesto a prueba en la Plataforma Abierta de Teleoperación (P.A.T.) que se ha desarrollado e integrado en la ETSII UPM como una herramienta que permita probar, validar y realizar experimentos de telerrobótica. Actualmente, este Protocolo de Teleoperación se ha propuesto a través de AENOR al grupo ISO de Telerobotics como una solución válida al problema existente y se encuentra bajo revisión. Con el diseño de dicho protocolo se ha conseguido enlazar maestro y esclavo, sin embargo con los niveles de radiación tan altos que hay en ITER la electrónica del controlador no puede entrar dentro del tokamak. Por ello se propone que a través de una mínima electrónica convenientemente protegida se puedan multiplexar las señales de control que van a través del cableado umbilical desde el controlador hasta la base del robot. En este ejercicio teórico se demuestra la utilidad y viabilidad de utilizar este tipo de solución para reducir el volumen y peso del cableado umbilical en cifras aproximadas de un 90%, para ello hay que desarrollar una electrónica específica y con certificación RadHard para soportar los enormes niveles de radiación de ITER. Para este manipulador de tipo genérico y con ayuda de la Plataforma Abierta de Teleoperación, se ha desarrollado un algoritmo que mediante un sensor de fuerza/par y una IMU colocados en la muñeca del robot, y convenientemente protegidos ante la radiación, permiten calcular las fuerzas e inercias que produce la carga, esto es necesario para poder transmitirle al operador unas fuerzas escaladas, y que pueda sentir la carga que manipula, y no otras fuerzas que puedan influir en el esclavo remoto, como ocurre con otras técnicas de estimación de fuerzas. Como el blindaje de los sensores no debe ser grande ni pesado, habrá que destinar este tipo de tecnología a las tareas de mantenimiento de las paradas programadas de ITER, que es cuando los niveles de radiación están en sus valores mínimos. Por otro lado para que el operador sienta lo más fielmente posible la fuerza de carga se ha desarrollado una electrónica que mediante el control en corriente de los motores permita realizar un control en fuerza a partir de la caracterización de los motores del maestro. Además para aumentar la percepción del operador se han realizado unos experimentos que demuestran que al aplicar estímulos multimodales (visuales, auditivos y hápticos) aumenta su inmersión y el rendimiento en la consecución de la tarea puesto que influyen directamente en su capacidad de respuesta. Finalmente, y en referencia a la realimentación visual del operador, en ITER se trabaja con cámaras situadas en localizaciones estratégicas, si bien el humano cuando manipula objetos hace uso de su visión binocular cambiando constantemente el punto de vista adecuándose a las necesidades visuales de cada momento durante el desarrollo de la tarea. Por ello, se ha realizado una reconstrucción tridimensional del espacio de la tarea a partir de una cámara-sensor RGB-D, lo cual nos permite obtener un punto de vista binocular virtual móvil a partir de una cámara situada en un punto fijo que se puede proyectar en un dispositivo de visualización 3D para que el operador pueda variar el punto de vista estereoscópico según sus preferencias. La correcta integración de estas tecnologías para la interacción hombre-robot en la P.A.T. ha permitido validar mediante pruebas y experimentos para verificar su utilidad en la aplicación práctica de la telemanipulación con alto grado de escalado en entornos nucleares de fusión. Abstract This thesis focuses on developing technologies for human-robot interaction in nuclear fusion environments. The main problem of nuclear fusion sector resides in such extreme environmental conditions existing in the hot-cell, leading to very restrictive requirements for equipment in order to deal with these high levels of radiation, magnetism, ultravacuum, temperature... Since it is not feasible to carry out tasks directly by humans, we must use remote handling devices for accomplishing operation and maintenance processes. In ITER facilities it is mandatory to have a controlled environment of extreme safety and security with validated standards. The definition and use of protocols is essential to govern its operation. Focusing on Remote Handling with some degree of escalation, protocols must be defined for open systems to allow interaction among different kind of equipment and several multifunctional devices. In this context, a Teleoperation Protocol definition enables interconnection between master and slave devices from different typologies, being able to communicate bilaterally one each other and using different control algorithms depending on the task to perform. This protocol and its interconnectivity have been tested in the Teleoperation Open Platform (T.O.P.) that has been developed and integrated in the ETSII UPM as a tool to test, validate and conduct experiments in Telerobotics. Currently, this protocol has been proposed for Teleoperation through AENOR to the ISO Telerobotics group as a valid solution to the existing problem, and it is under review. Master and slave connection has been achieved with this protocol design, however with such high radiation levels in ITER, the controller electronics cannot enter inside the tokamak. Therefore it is proposed a multiplexed electronic board, that through suitable and RadHard protection processes, to transmit control signals through an umbilical cable from the controller to the robot base. In this theoretical exercise the utility and feasibility of using this type of solution reduce the volume and weight of the umbilical wiring approximate 90% less, although it is necessary to develop specific electronic hardware and validate in RadHard qualifications in order to handle huge levels of ITER radiation. Using generic manipulators does not allow to implement regular sensors for force feedback in ITER conditions. In this line of research, an algorithm to calculate the forces and inertia produced by the load has been developed using a force/torque sensor and IMU, both conveniently protected against radiation and placed on the robot wrist. Scaled forces should be transmitted to the operator, feeling load forces but not other undesirable forces in slave system as those resulting from other force estimation techniques. Since shielding of the sensors should not be large and heavy, it will be necessary to allocate this type of technology for programmed maintenance periods of ITER, when radiation levels are at their lowest levels. Moreover, the operator perception needs to feel load forces as accurate as possible, so some current control electronics were developed to perform a force control of master joint motors going through a correct motor characterization. In addition to increase the perception of the operator, some experiments were conducted to demonstrate applying multimodal stimuli (visual, auditory and haptic) increases immersion and performance in achieving the task since it is directly correlated with response time. Finally, referring to the visual feedback to the operator in ITER, it is usual to work with 2D cameras in strategic locations, while humans use binocular vision in direct object manipulation, constantly changing the point of view adapting it to the visual needs for performing manipulation during task procedures. In this line a three-dimensional reconstruction of non-structured scenarios has been developed using RGB-D sensor instead of cameras in the remote environment. Thus a mobile virtual binocular point of view could be generated from a camera at a fixed point, projecting stereoscopic images in 3D display device according to operator preferences. The successful integration of these technologies for human-robot interaction in the T.O.P., and validating them through tests and experiments, verify its usefulness in practical application of high scaling remote handling at nuclear fusion environments.