707 resultados para Hydraulics transients


Relevância:

10.00% 10.00%

Publicador:

Resumo:

A prominent control on the flow over subaqueous dunes is the slope of the downstream leeside. While previous work has focused on steep (~30°), asymmetric dunes with permanent flow separation, little is known about dunes with lower lee-slope angles for which flow separation is absent or intermittent. Here, we present a laboratory investigation where we systematically varied the dune lee-slope, holding other geometric parameters and flow hydraulics constant, to explore effects on the turbulent flow field and flow resistance. Three sets of fixed dunes (lee-slopes of 10°, 20° and 30°) were separately installed in a 15 m long and 1 m wide flume and subjected to 0.20 m deep flow. Measurements consisted of high-frequency, vertical profiles collected with a Laser Doppler Velocimeter (LDV). We show that the temporal and spatial occurrence of flow separation decreases with dune lee-slope. Velocity gradients in the dune leeside depict a free shear layer downstream of the 30° dunes and a weaker shear layer closer to the bed for the 20° and 10° dunes. The decrease in velocity gradients leads to lower magnitude of turbulence production for gentle lee-slopes. Aperiodic, strong ejection events dominate the shear layer, but decrease in strength and frequency for low-angle dunes. Flow resistance of dunes decreases with lee-slope; the transition being non-linear. Over the 10°, 20° and 30° dunes, shear stress is 8%, 33% and 90 % greater than a flat bed, respectively. Our results demonstrate that dune lee-slope plays an important, but often ignored role in flow resistance.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Deep ocean circulation has been considered relatively stable during interglacial periods, yet little is known about its behavior on submillennial time scales. Using a subcentennially resolved epibenthic foraminiferal d13C record we show that North Atlantic Deep Water (NADW) influence was strong at the onset of the last interglacial period and then interrupted by several prominent, centennial-scale reductions. These NADW transients occurred during periods of increased ice rafting and southward expansions of polar water influence, suggesting that a buoyancy threshold for convective instability was triggered by freshwater and circum-Arctic cryosphere changes. The deep Atlantic chemical changes were similar in magnitude to those associated with glaciations, implying that the canonical view of a relatively stable interglacial circulation may not hold for conditions warmer/fresher than at present.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Tara Oceans Expedition (2009-2013) sampled the world oceans on board a 36 m long schooner, collecting environmental data and organisms from viruses to planktonic metazoans for later analyses using modern sequencing and state-of-the-art imaging technologies. Tara Oceans Data are particularly suited to study the genetic, morphological and functional diversity of plankton. The present data set provides continuous measurements made with a FRRF instrument, operating in a flow-through mode during the 2009-2012 part of the expedition. It operates by exciting chlorophyll fluorescence using a series of short flashes of controlled energy and time intervals (Kolber et al, 1998). The fluorescence transients produced by this excitation signal were analysed in real-time to provide estimates of abundance of photosynthetic pigments, the photosynthetic yields (Fv/Fm), the functional absorption cross section (a proxy for efficiency of photosynthetic energy acquisition), the kinetics of photosynthetic electron transport between Photosystem II and Photosystem I, and the size of the PQ pool. These parameters were measured at excitation wavelength of 445 nm, 470nm, 505 nm, and 535 nm, allowing to assess the presence and the photosynthetic performance of different phytoplankton taxa based on the spectral composition of their light harvesting pigments. The FRRF-derived photosynthetic characteristics were used to calculate the initial slope, the half saturation, and the maximum level of Photosynthesis vs Irradiance relationship. FRRF data were acquired continuously, at 1-minute time intervals.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

There exists an interest in performing pin-by-pin calculations coupled with thermal hydraulics so as to improve the accuracy of nuclear reactor analysis. In the framework of the EU NURISP project, INRNE and UPM have generated an experimental version of a few group diffusion cross sections library with discontinuity factors intended for VVER analysis at the pin level with the COBAYA3 code. The transport code APOLLO2 was used to perform the branching calculations. As a first proof of principle the library was created for fresh fuel and covers almost the full parameter space of steady state and transient conditions. The main objective is to test the calculation schemes and post-processing procedures, including multi-pin branching calculations. Two library options are being studied: one based on linear table interpolation and another one using a functional fitting of the cross sections. The libraries generated with APOLLO2 have been tested with the pin-by-pin diffusion model in COBAYA3 including discontinuity factors; first comparing 2D results against the APOLLO2 reference solutions and afterwards using the libraries to compute a 3D assembly problem coupled with a simplified thermal-hydraulic model.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The need to refine models for best-estimate calculations, based on good-quality experimental data, has been expressed in many recent meetings in the field of nuclear applications. The modeling needs arising in this respect should not be limited to the currently available macroscopic methods but should be extended to next-generation analysis techniques that focus on more microscopic processes. One of the most valuable databases identified for the thermalhydraulics modeling was developed by the Nuclear Power Engineering Corporation (NUPEC), Japan. From 1987 to 1995, NUPEC performed steady-state and transient critical power and departure from nucleate boiling (DNB) test series based on the equivalent full-size mock-ups. Considering the reliability not only of the measured data, but also other relevant parameters such as the system pressure, inlet sub-cooling and rod surface temperature, these test series supplied the first substantial database for the development of truly mechanistic and consistent models for boiling transition and critical heat flux. Over the last few years the Pennsylvania State University (PSU) under the sponsorship of the U.S. Nuclear Regulatory Commission (NRC) has prepared, organized, conducted and summarized the OECD/NRC Full-size Fine-mesh Bundle Tests (BFBT) Benchmark. The international benchmark activities have been conducted in cooperation with the Nuclear Energy Agency/Organization for Economic Co-operation and Development (NEA/OECD) and Japan Nuclear Energy Safety (JNES) organization, Japan. Consequently, the JNES has made available the Boiling Water Reactor (BWR) NUPEC database for the purposes of the benchmark. Based on the success of the OECD/NRC BFBT benchmark the JNES has decided to release also the data based on the NUPEC Pressurized Water Reactor (PWR) subchannel and bundle tests for another follow-up international benchmark entitled OECD/NRC PWR Subchannel and Bundle Tests (PSBT) benchmark. This paper presents an application of the joint Penn State University/Technical University of Madrid (UPM) version of the well-known subchannel code COBRA-TF, namely CTF, to the critical power and departure from nucleate boiling (DNB) exercises of the OECD/NRC BFBT and PSBT benchmarks

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Rms voltage regulation may be an attractive possibility for controlling power inverters. Combined with a Hall Effect sensor for current control, it keeps its parallel operation capability while increasing its noise immunity, which may lead to a reduction of the Total Harmonic Distortion (THD). Besides, as voltage regulation is designed in DC, a simple PI regulator can provide accurate voltage tracking. Nevertheless, this approach does not lack drawbacks. Its narrow voltage bandwidth makes transients last longer and it increases the voltage THD when feeding non-linear loads, such as rectifying stages. On the other hand, the implementation can fall into offset voltage error. Furthermore, the information of the output voltage phase is hidden for the control as well, making the synchronization of a 3-phase setup not trivial. This paper explains the concept, design and implementation of the whole control scheme, in an on board inverter able to run in parallel and within a 3-phase setup. Special attention is paid to solve the problems foreseen at implementation level: a third analog loop accounts for the offset level is added and a digital algorithm guarantees 3-phase voltage synchronization.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work explores the automatic recognition of physical activity intensity patterns from multi-axial accelerometry and heart rate signals. Data collection was carried out in free-living conditions and in three controlled gymnasium circuits, for a total amount of 179.80 h of data divided into: sedentary situations (65.5%), light-to-moderate activity (17.6%) and vigorous exercise (16.9%). The proposed machine learning algorithms comprise the following steps: time-domain feature definition, standardization and PCA projection, unsupervised clustering (by k-means and GMM) and a HMM to account for long-term temporal trends. Performance was evaluated by 30 runs of a 10-fold cross-validation. Both k-means and GMM-based approaches yielded high overall accuracy (86.97% and 85.03%, respectively) and, given the imbalance of the dataset, meritorious F-measures (up to 77.88%) for non-sedentary cases. Classification errors tended to be concentrated around transients, what constrains their practical impact. Hence, we consider our proposal to be suitable for 24 h-based monitoring of physical activity in ambulatory scenarios and a first step towards intensity-specific energy expenditure estimators

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La acumulación de material sólido en embalses, cauces fluviales y en zonas marítimas hace que la extracción mecánica de estos materiales por medio de succión sea cada vez mas frecuente, por ello resulta importante estudiar el rendimiento de la succión de estos materiales analizando la forma de las boquillas y los parámetros del flujo incluyendo la bomba. Esta tesis estudia, mediante equipos experimentales, la eficacia de distintos dispositivos de extracción de sólidos (utilizando boquillas de diversas formas y bombas de velocidad variable). El dispositivo experimental ha sido desarrollado en el Laboratorio de Hidráulica de la E.T.S.I. de Caminos, C. y P. de la Universidad Politécnica de Madrid. Dicho dispositivo experimental incluye un lecho sumergido de distintos tipos de sedimentos, boquillas de extracción de sólidos y bomba de velocidad variable, así como un elemento de separación del agua y los sólidos extraídos. Los parámetros básicos analizados son el caudal líquido total bombeado, el caudal sólido extraído, diámetro de la tubería de succión, forma y sección de la boquilla extractora, así como los parámetros de velocidad y rendimiento en la bomba de velocidad variable. Los resultados de las medidas obtenidas en el dispositivo experimental han sido estudiados por medio del análisis dimensional y con métodos estadísticos. A partir de este estudio se ha desarrollado una nueva formulación, que relaciona el caudal sólido extraído con los diámetros de tubería y boquilla, caudal líquido bombeado y velocidad de giro de la bomba. Así mismo, desde el punto de vista práctico, se han analizado la influencia de la forma de la boquilla con la capacidad de extracción de sólidos a igualdad del resto de los parámetros, de forma que se puedan recomendar que forma de la boquilla es la más apropiada. The accumulation of solid material in reservoirs, river channels and sea areas causes the mechanical extraction of these materials by suction is becoming much more common, so it is important to study the performance of the suction of these materials analyzing the shape of the nozzles and flow parameters, including the pump. This thesis studies, using experimental equipment, the effectiveness of different solids removal devices (using nozzles of different shapes and variable speed pumps). The experimental device was developed at the Hydraulics Laboratory of the Civil University of the Polytechnic University of Madrid. The device included a submerged bed with different types of sediment solids, different removal nozzles and variable speed pump. It also includes a water separation element and the solids extracted. The key parameters analyzed are the total liquid volume pumped, the solid volume extracted, diameter of the suction pipe, a section of the nozzle and hood, and the parameters of speed and efficiency of the variable speed pump. The basic parameters analyzed are the total liquid volume pumped, the removed solid volume, the diameter of the suction pipe, the shape and cross-section of the nozzle, and the parameters of speed, efficiency and energy consumed by the variable speed pump. The measurements obtained on the experimental device have been studied with dimensional analysis and statistical methods. The outcome of this study is a new formulation, which relates the solid volume extracted with the pipe and nozzle diameters, the pumped liquid flow and the speed of the pump. Also, from a practical point of view, the influence of the shape of the nozzle was compared with the solid extraction capacity, keeping equal the rest of the parameters. So, a recommendation of the best shape of the nozzle can be given.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

El accidente de rotura de tubos de un generador de vapor (Steam Generator Tube Rupture, SGTR) en los reactores de agua a presión es uno de los transitorios más exigentes desde el punto de vista de operación. Los transitorios de SGTR son especiales, ya que podría dar lugar a emisiones radiológicas al exterior sin necesidad de daño en el núcleo previo o sin que falle la contención, ya que los SG pueden constituir una vía directa desde el reactor al medio ambiente en este transitorio. En los análisis de seguridad, el SGTR se analiza desde un punto determinista y probabilista, con distintos enfoques con respecto a las acciones del operador y las consecuencias analizadas. Cuando comenzaron los Análisis Deterministas de Seguridad (DSA), la forma de analizar el SGTR fue sin dar crédito a la acción del operador durante los primeros 30 min del transitorio, lo que suponía que el grupo de operación era capaz de detener la fuga por el tubo roto dentro de ese tiempo. Sin embargo, los diferentes casos reales de accidentes de SGTR sucedidos en los EE.UU. y alrededor del mundo demostraron que los operadores pueden emplear más de 30 minutos para detener la fuga en la vida real. Algunas metodologías fueron desarrolladas en los EEUU y en Europa para abordar esa cuestión. En el Análisis Probabilista de Seguridad (PSA), las acciones del operador se tienen en cuenta para diseñar los cabeceros en el árbol de sucesos. Los tiempos disponibles se utilizan para establecer los criterios de éxito para dichos cabeceros. Sin embargo, en una secuencia dinámica como el SGTR, las acciones de un operador son muy dependientes del tiempo disponible por las acciones humanas anteriores. Además, algunas de las secuencias de SGTR puede conducir a la liberación de actividad radiológica al exterior sin daño previo en el núcleo y que no se tienen en cuenta en el APS, ya que desde el punto de vista de la integridad de núcleo son de éxito. Para ello, para analizar todos estos factores, la forma adecuada de analizar este tipo de secuencias pueden ser a través de una metodología que contemple Árboles de Sucesos Dinámicos (Dynamic Event Trees, DET). En esta Tesis Doctoral se compara el impacto en la evolución temporal y la dosis al exterior de la hipótesis más relevantes encontradas en los Análisis Deterministas a nivel mundial. La comparación se realiza con un modelo PWR Westinghouse de tres lazos (CN Almaraz) con el código termohidráulico TRACE, con hipótesis de estimación óptima, pero con hipótesis deterministas como criterio de fallo único o pérdida de energía eléctrica exterior. Las dosis al exterior se calculan con RADTRAD, ya que es uno de los códigos utilizados normalmente para los cálculos de dosis del SGTR. El comportamiento del reactor y las dosis al exterior son muy diversas, según las diferentes hipótesis en cada metodología. Por otra parte, los resultados están bastante lejos de los límites de regulación, pese a los conservadurismos introducidos. En el siguiente paso de la Tesis Doctoral, se ha realizado un análisis de seguridad integrado del SGTR según la metodología ISA, desarrollada por el Consejo de Seguridad Nuclear español (CSN). Para ello, se ha realizado un análisis termo-hidráulico con un modelo de PWR Westinghouse de 3 lazos con el código MAAP. La metodología ISA permite la obtención del árbol de eventos dinámico del SGTR, teniendo en cuenta las incertidumbres en los tiempos de actuación del operador. Las simulaciones se realizaron con SCAIS (sistema de simulación de códigos para la evaluación de la seguridad integrada), que incluye un acoplamiento dinámico con MAAP. Las dosis al exterior se calcularon también con RADTRAD. En los resultados, se han tenido en cuenta, por primera vez en la literatura, las consecuencias de las secuencias en términos no sólo de daños en el núcleo sino de dosis al exterior. Esta tesis doctoral demuestra la necesidad de analizar todas las consecuencias que contribuyen al riesgo en un accidente como el SGTR. Para ello se ha hecho uso de una metodología integrada como ISA-CSN. Con este enfoque, la visión del DSA del SGTR (consecuencias radiológicas) se une con la visión del PSA del SGTR (consecuencias de daño al núcleo) para evaluar el riesgo total del accidente. Abstract Steam Generator Tube Rupture accidents in Pressurized Water Reactors are known to be one of the most demanding transients for the operating crew. SGTR are special transient as they could lead to radiological releases without core damage or containment failure, as they can constitute a direct path to the environment. The SGTR is analyzed from a Deterministic and Probabilistic point of view in the Safety Analysis, although the assumptions of the different approaches regarding the operator actions are quite different. In the beginning of Deterministic Safety Analysis, the way of analyzing the SGTR was not crediting the operator action for the first 30 min of the transient, assuming that the operating crew was able to stop the primary to secondary leakage within that time. However, the different real SGTR accident cases happened in the USA and over the world demonstrated that operators can took more than 30 min to stop the leakage in actual sequences. Some methodologies were raised in the USA and in Europe to cover that issue. In the Probabilistic Safety Analysis, the operator actions are taken into account to set the headers in the event tree. The available times are used to establish the success criteria for the headers. However, in such a dynamic sequence as SGTR, the operator actions are very dependent on the time available left by the other human actions. Moreover, some of the SGTR sequences can lead to offsite doses without previous core damage and they are not taken into account in PSA as from the point of view of core integrity are successful. Therefore, to analyze all this factors, the appropriate way of analyzing that kind of sequences could be through a Dynamic Event Tree methodology. This Thesis compares the impact on transient evolution and the offsite dose of the most relevant hypothesis of the different SGTR analysis included in the Deterministic Safety Analysis. The comparison is done with a PWR Westinghouse three loop model in TRACE code (Almaraz NPP), with best estimate assumptions but including deterministic hypothesis such as single failure criteria or loss of offsite power. The offsite doses are calculated with RADTRAD code, as it is one of the codes normally used for SGTR offsite dose calculations. The behaviour of the reactor and the offsite doses are quite diverse depending on the different assumptions made in each methodology. On the other hand, although the high conservatism, such as the single failure criteria, the results are quite far from the regulatory limits. In the next stage of the Thesis, the Integrated Safety Assessment (ISA) methodology, developed by the Spanish Nuclear Safety Council (CSN), has been applied to a thermohydraulical analysis of a Westinghouse 3-loop PWR plant with the MAAP code. The ISA methodology allows obtaining the SGTR Dynamic Event Tree taking into account the uncertainties on the operator actuation times. Simulations are performed with SCAIS (Simulation Code system for Integrated Safety Assessment), which includes a dynamic coupling with MAAP thermal hydraulic code. The offsite doses are calculated also with RADTRAD. The results shows the consequences of the sequences in terms not only of core damage but of offsite doses. This Thesis shows the need of analyzing all the consequences in an accident such as SGTR. For that, an it has been used an integral methodology like ISA-CSN. With this approach, the DSA vision of the SGTR (radiological consequences) is joined with the PSA vision of the SGTR (core damage consequences) to measure the total risk of the accident.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Se desarrollan varias técnicas basadas en descomposición ortogonal propia (DOP) local y proyección de tipo Galerkin para acelerar la integración numérica de problemas de evolución, de tipo parabólico, no lineales. Las ideas y métodos que se presentan conllevan un nuevo enfoque para la modelización de tipo DOP, que combina intervalos temporales cortos en que se usa un esquema numérico estándard con otros intervalos temporales en que se utilizan los sistemas de tipo Galerkin que resultan de proyectar las ecuaciones de evolución sobre la variedad lineal generada por los modos DOP, obtenidos a partir de instantáneas calculadas en los intervalos donde actúa el código numérico. La variedad DOP se construye completamente en el primer intervalo, pero solamente se actualiza en los demás intervalos según las dinámicas de la solución, aumentando de este modo la eficiencia del modelo de orden reducido resultante. Además, se aprovechan algunas propiedades asociadas a la dependencia débil de los modos DOP tanto en la variable temporal como en los posibles parámetros de que pueda depender el problema. De esta forma, se aumentan la flexibilidad y la eficiencia computacional del proceso. La aplicación de los métodos resultantes es muy prometedora, tanto en la simulación de transitorios en flujos laminares como en la construcción de diagramas de bifurcación en sistemas dependientes de parámetros. Las ideas y los algoritmos desarrollados en la tesis se ilustran en dos problemas test, la ecuación unidimensional compleja de Ginzburg-Landau y el problema bidimensional no estacionario de la cavidad. Abstract Various ideas and methods involving local proper orthogonal decomposition (POD) and Galerkin projection are presented aiming at accelerating the numerical integration of nonlinear time dependent parabolic problems. The proposed methods come from a new approach to the POD-based model reduction procedures, which combines short runs with a given numerical solver and a reduced order model constructed by expanding the solution of the problem into appropriate POD modes, which span a POD manifold, and Galerkin projecting some evolution equations onto that linear manifold. The POD manifold is completely constructed from the outset, but only updated as time proceeds according to the dynamics, which yields an adaptive and flexible procedure. In addition, some properties concerning the weak dependence of the POD modes on time and possible parameters in the problem are exploited in order to increase the flexibility and efficiency of the low dimensional model computation. Application of the developed techniques to the approximation of transients in laminar fluid flows and the simulation of attractors in bifurcation problems shows very promising results. The test problems considered to illustrate the various ideas and check the performance of the algorithms are the onedimensional complex Ginzburg-Landau equation and the two-dimensional unsteady liddriven cavity problem.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Integrated Safety Assessment (ISA) methodology, developed by the Spanish Nuclear Safety Council (CSN), has been applied to a thermo-hydraulical analysis of a Westinghouse 3-loop PWR plant by means of the dynamic event trees (DET) for Steam Generator Tube Rupture (SGTR) sequences. The ISA methodology allows obtaining the SGTR Dynamic Event Tree taking into account the operator actuation times. Simulations are performed with SCAIS (Simulation Code system for Integrated Safety Assessment), which includes a dynamic coupling with MAAP thermal hydraulic code. The results show the capability of the ISA methodology and SCAIS platform to obtain the DET of complex sequences.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work is related to the improvement of the output impedance of the Buck converter by means of introducing an additional power path that virtually increases the output capacitance during transients. It is well known that in VRM applications, with wide load steps, voltage overshoots and undershoots may lead to undesired performance of the load. To solve this problem, high-bandwidth high-switching frequency power converters can be applied to reduce the transient time or a big output capacitor can be applied to reduce the output impedance. The first solution can degrade the efficiency by increasing switching losses of the MOSFETS, and the second solution is penalizing the cost and size of the output filter. The Output Impedance Correction Circuit (OICC), as presented here, is used to inject or extract a current n-1 times larger than the output capacitor current, thus virtually increasing n times the value of the output capacitance during the transients. This feature allows the usage of a low frequency Buck converter with smaller capacitor but satisfying the dynamic requirements.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work is related to the improvement of the dynamic performance of the Buck converter by means of introducing an additional power path that virtually increase s the output capacitance during transients, thus improving the output impedance of the converter. It is well known that in VRM applications, with wide load steps, voltage overshoots and undershoots ma y lead to undesired performance of the load. To solve this problem, high-bandwidth high-switching frequency power converter s can be applied to reduce the transient time or a big output capacitor can be applied to reduce the output impedance. The first solution can degrade the efficiency by increasing switching losses of the MOSFETS, and the second solution is penalizing the cost and size of the output filter. The additional energy path, as presented here, is introduced with the Output Impedance Correction Circuit (OICC) based on the Controlled Current Source (CCS). The OICC is using CCS to inject or extract a current n - 1 times larger than the output capacitor current, thus virtually increasing n times the value of the output capacitance during the transients. This feature allows the usage of a low frequency Buck converter with smaller capacitor but satisfying the dynamic requirements.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The design of nuclear power plant has to follow a number of regulations aimed at limiting the risks inherent in this type of installation. The goal is to prevent and to limit the consequences of any possible incident that might threaten the public or the environment. To verify that the safety requirements are met a safety assessment process is followed. Safety analysis is as key component of a safety assessment, which incorporates both probabilistic and deterministic approaches. The deterministic approach attempts to ensure that the various situations, and in particular accidents, that are considered to be plausible, have been taken into account, and that the monitoring systems and engineered safety and safeguard systems will be capable of ensuring the safety goals. On the other hand, probabilistic safety analysis tries to demonstrate that the safety requirements are met for potential accidents both within and beyond the design basis, thus identifying vulnerabilities not necessarily accessible through deterministic safety analysis alone. Probabilistic safety assessment (PSA) methodology is widely used in the nuclear industry and is especially effective in comprehensive assessment of the measures needed to prevent accidents with small probability but severe consequences. Still, the trend towards a risk informed regulation (RIR) demanded a more extended use of risk assessment techniques with a significant need to further extend PSA’s scope and quality. Here is where the theory of stimulated dynamics (TSD) intervenes, as it is the mathematical foundation of the integrated safety assessment (ISA) methodology developed by the CSN(Consejo de Seguridad Nuclear) branch of Modelling and Simulation (MOSI). Such methodology attempts to extend classical PSA including accident dynamic analysis, an assessment of the damage associated to the transients and a computation of the damage frequency. The application of this ISA methodology requires a computational framework called SCAIS (Simulation Code System for Integrated Safety Assessment). SCAIS provides accident dynamic analysis support through simulation of nuclear accident sequences and operating procedures. Furthermore, it includes probabilistic quantification of fault trees and sequences; and integration and statistic treatment of risk metrics. SCAIS comprehensively implies an intensive use of code coupling techniques to join typical thermal hydraulic analysis, severe accident and probability calculation codes. The integration of accident simulation in the risk assessment process and thus requiring the use of complex nuclear plant models is what makes it so powerful, yet at the cost of an enormous increase in complexity. As the complexity of the process is primarily focused on such accident simulation codes, the question of whether it is possible to reduce the number of required simulation arises, which will be the focus of the present work. This document presents the work done on the investigation of more efficient techniques applied to the process of risk assessment inside the mentioned ISA methodology. Therefore such techniques will have the primary goal of decreasing the number of simulation needed for an adequate estimation of the damage probability. As the methodology and tools are relatively recent, there is not much work done inside this line of investigation, making it a quite difficult but necessary task, and because of time limitations the scope of the work had to be reduced. Therefore, some assumptions were made to work in simplified scenarios best suited for an initial approximation to the problem. The following section tries to explain in detail the process followed to design and test the developed techniques. Then, the next section introduces the general concepts and formulae of the TSD theory which are at the core of the risk assessment process. Afterwards a description of the simulation framework requirements and design is given. Followed by an introduction to the developed techniques, giving full detail of its mathematical background and its procedures. Later, the test case used is described and result from the application of the techniques is shown. Finally the conclusions are presented and future lines of work are exposed.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Realistic operation of helicopter flight simulators in complex topographies (such as urban environments) requires appropriate prediction of the incoming wind, and this prediction should be made in real time. Unfortunately, the wind topology around complex topographies shows time-dependent, fully nonlinear, turbulent patterns (i.e., wakes) whose simulation cannot be made using computationally inexpensive tools based on corrected potential approximations. Instead, the full Navier-Stokes plus some kind of turbulent modeling is necessary, which is quite computationally expensive. The complete unsteady flow depends on two parameters, namely the velocity and orientation of the free stream flow. The aim of this MSc thesis is to develop a methodology for the real time simulation of these complex flows. For simplicity, the flow around a single building (20 mx20 m cross section and 100 m height) is considered, with free stream velocity in the range 5-25 m/s. Because of the square cross section, the problem shows two reflection symmetries, which allows for restricting the orientations to the range 0° < a. < 45°. The methodology includes an offline preprocess and the online operation. The preprocess consists in three steps: An appropriate, unstructured mesh is selected in which the flow is sim¬ulated using OpenFOAM, and this is done for 33 combinations of 3 free stream intensities and 11 orientations. For each of these, the simulation proceeds for a sufficiently large time as to eliminate transients. This step is quite computationally expensive. Each flow field is post-processed using a combination of proper orthogonal decomposition, fast Fourier transform, and a convenient optimization tool, which identifies the relevant frequencies (namely, both the basic frequencies and their harmonics) and modes in the computational mesh. This combination includes several new ingredients to filter errors out and identify the relevant spatio-temporal patterns. Note that, in principle, the basic frequencies depend on both the intensity and the orientation of the free stream flow. The outcome of this step is a set of modes (vectors containing the three velocity components at all mesh points) for the various Fourier components, intensities, and orientations, which can be organized as a third order tensor. This step is fairly computationally inexpensive. The above mentioned tensor is treated using a combination of truncated high order singular value, decomposition and appropriate one-dimensional interpolation (as in Lorente, Velazquez, Vega, J. Aircraft, 45 (2008) 1779-1788). The outcome is a tensor representation of both the relevant fre¬quencies and the associated Fourier modes for a given pair of values of the free stream flow intensity and orientation. This step is fairly compu¬tationally inexpensive. The online, operation requires just reconstructing the time-dependent flow field from its Fourier representation, which is extremely computationally inex¬pensive. The whole method is quite robust.