954 resultados para Test Case Generator


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Dynamic thermal management techniques require a collection of on-chip thermal sensors that imply a significant area and power overhead. Finding the optimum number of temperature monitors and their location on the chip surface to optimize accuracy is an NP-hard problem. In this work we improve the modeling of the problem by including area, power and networking constraints along with the consideration of three inaccuracy terms: spatial errors, sampling rate errors and monitor-inherent errors. The problem is solved by the simulated annealing algorithm. We apply the algorithm to a test case employing three different types of monitors to highlight the importance of the different metrics. Finally we present a case study of the Alpha 21364 processor under two different constraint scenarios.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The design of nuclear power plant has to follow a number of regulations aimed at limiting the risks inherent in this type of installation. The goal is to prevent and to limit the consequences of any possible incident that might threaten the public or the environment. To verify that the safety requirements are met a safety assessment process is followed. Safety analysis is as key component of a safety assessment, which incorporates both probabilistic and deterministic approaches. The deterministic approach attempts to ensure that the various situations, and in particular accidents, that are considered to be plausible, have been taken into account, and that the monitoring systems and engineered safety and safeguard systems will be capable of ensuring the safety goals. On the other hand, probabilistic safety analysis tries to demonstrate that the safety requirements are met for potential accidents both within and beyond the design basis, thus identifying vulnerabilities not necessarily accessible through deterministic safety analysis alone. Probabilistic safety assessment (PSA) methodology is widely used in the nuclear industry and is especially effective in comprehensive assessment of the measures needed to prevent accidents with small probability but severe consequences. Still, the trend towards a risk informed regulation (RIR) demanded a more extended use of risk assessment techniques with a significant need to further extend PSA’s scope and quality. Here is where the theory of stimulated dynamics (TSD) intervenes, as it is the mathematical foundation of the integrated safety assessment (ISA) methodology developed by the CSN(Consejo de Seguridad Nuclear) branch of Modelling and Simulation (MOSI). Such methodology attempts to extend classical PSA including accident dynamic analysis, an assessment of the damage associated to the transients and a computation of the damage frequency. The application of this ISA methodology requires a computational framework called SCAIS (Simulation Code System for Integrated Safety Assessment). SCAIS provides accident dynamic analysis support through simulation of nuclear accident sequences and operating procedures. Furthermore, it includes probabilistic quantification of fault trees and sequences; and integration and statistic treatment of risk metrics. SCAIS comprehensively implies an intensive use of code coupling techniques to join typical thermal hydraulic analysis, severe accident and probability calculation codes. The integration of accident simulation in the risk assessment process and thus requiring the use of complex nuclear plant models is what makes it so powerful, yet at the cost of an enormous increase in complexity. As the complexity of the process is primarily focused on such accident simulation codes, the question of whether it is possible to reduce the number of required simulation arises, which will be the focus of the present work. This document presents the work done on the investigation of more efficient techniques applied to the process of risk assessment inside the mentioned ISA methodology. Therefore such techniques will have the primary goal of decreasing the number of simulation needed for an adequate estimation of the damage probability. As the methodology and tools are relatively recent, there is not much work done inside this line of investigation, making it a quite difficult but necessary task, and because of time limitations the scope of the work had to be reduced. Therefore, some assumptions were made to work in simplified scenarios best suited for an initial approximation to the problem. The following section tries to explain in detail the process followed to design and test the developed techniques. Then, the next section introduces the general concepts and formulae of the TSD theory which are at the core of the risk assessment process. Afterwards a description of the simulation framework requirements and design is given. Followed by an introduction to the developed techniques, giving full detail of its mathematical background and its procedures. Later, the test case used is described and result from the application of the techniques is shown. Finally the conclusions are presented and future lines of work are exposed.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Antecedentes Europa vive una situación insostenible. Desde el 2008 se han reducido los recursos de los gobiernos a raíz de la crisis económica. El continente Europeo envejece con ritmo constante al punto que se prevé que en 2050 habrá sólo dos trabajadores por jubilado [54]. A esta situación se le añade el aumento de la incidencia de las enfermedades crónicas, relacionadas con el envejecimiento, cuyo coste puede alcanzar el 7% del PIB de un país [51]. Es necesario un cambio de paradigma. Una nueva manera de cuidar de la salud de las personas: sustentable, eficaz y preventiva más que curativa. Algunos estudios abogan por el cuidado personalizado de la salud (pHealth). En este modelo las prácticas médicas son adaptadas e individualizadas al paciente, desde la detección de los factores de riesgo hasta la personalización de los tratamientos basada en la respuesta del individuo [81]. El cuidado personalizado de la salud está asociado a menudo al uso de las tecnologías de la información y comunicación (TICs) que, con su desarrollo exponencial, ofrecen oportunidades interesantes para la mejora de la salud. El cambio de paradigma hacia el pHealth está lentamente ocurriendo, tanto en el ámbito de la investigación como en la industria, pero todavía no de manera significativa. Existen todavía muchas barreras relacionadas a la economía, a la política y la cultura. También existen barreras puramente tecnológicas, como la falta de sistemas de información interoperables [199]. A pesar de que los aspectos de interoperabilidad están evolucionando, todavía hace falta un diseño de referencia especialmente direccionado a la implementación y el despliegue en gran escala de sistemas basados en pHealth. La presente Tesis representa un intento de organizar la disciplina de la aplicación de las TICs al cuidado personalizado de la salud en un modelo de referencia, que permita la creación de plataformas de desarrollo de software para simplificar tareas comunes de desarrollo en este dominio. Preguntas de investigación RQ1 >Es posible definir un modelo, basado en técnicas de ingeniería del software, que represente el dominio del cuidado personalizado de la salud de una forma abstracta y representativa? RQ2 >Es posible construir una plataforma de desarrollo basada en este modelo? RQ3 >Esta plataforma ayuda a los desarrolladores a crear sistemas pHealth complejos e integrados? Métodos Para la descripción del modelo se adoptó el estándar ISO/IEC/IEEE 42010por ser lo suficientemente general y abstracto para el amplio enfoque de esta tesis [25]. El modelo está definido en varias partes: un modelo conceptual, expresado a través de mapas conceptuales que representan las partes interesadas (stakeholders), los artefactos y la información compartida; y escenarios y casos de uso para la descripción de sus funcionalidades. El modelo fue desarrollado de acuerdo a la información obtenida del análisis de la literatura, incluyendo 7 informes industriales y científicos, 9 estándares, 10 artículos en conferencias, 37 artículos en revistas, 25 páginas web y 5 libros. Basándose en el modelo se definieron los requisitos para la creación de la plataforma de desarrollo, enriquecidos por otros requisitos recolectados a través de una encuesta realizada a 11 ingenieros con experiencia en la rama. Para el desarrollo de la plataforma, se adoptó la metodología de integración continua [74] que permitió ejecutar tests automáticos en un servidor y también desplegar aplicaciones en una página web. En cuanto a la metodología utilizada para la validación se adoptó un marco para la formulación de teorías en la ingeniería del software [181]. Esto requiere el desarrollo de modelos y proposiciones que han de ser validados dentro de un ámbito de investigación definido, y que sirvan para guiar al investigador en la búsqueda de la evidencia necesaria para justificarla. La validación del modelo fue desarrollada mediante una encuesta online en tres rondas con un número creciente de invitados. El cuestionario fue enviado a 134 contactos y distribuido en algunos canales públicos como listas de correo y redes sociales. El objetivo era evaluar la legibilidad del modelo, su nivel de cobertura del dominio y su potencial utilidad en el diseño de sistemas derivados. El cuestionario incluía preguntas cuantitativas de tipo Likert y campos para recolección de comentarios. La plataforma de desarrollo fue validada en dos etapas. En la primera etapa se utilizó la plataforma en un experimento a pequeña escala, que consistió en una sesión de entrenamiento de 12 horas en la que 4 desarrolladores tuvieron que desarrollar algunos casos de uso y reunirse en un grupo focal para discutir su uso. La segunda etapa se realizó durante los tests de un proyecto en gran escala llamado HeartCycle [160]. En este proyecto un equipo de diseñadores y programadores desarrollaron tres aplicaciones en el campo de las enfermedades cardio-vasculares. Una de estas aplicaciones fue testeada en un ensayo clínico con pacientes reales. Al analizar el proyecto, el equipo de desarrollo se reunió en un grupo focal para identificar las ventajas y desventajas de la plataforma y su utilidad. Resultados Por lo que concierne el modelo que describe el dominio del pHealth, la parte conceptual incluye una descripción de los roles principales y las preocupaciones de los participantes, un modelo de los artefactos TIC que se usan comúnmente y un modelo para representar los datos típicos que son necesarios formalizar e intercambiar entre sistemas basados en pHealth. El modelo funcional incluye un conjunto de 18 escenarios, repartidos en: punto de vista de la persona asistida, punto de vista del cuidador, punto de vista del desarrollador, punto de vista de los proveedores de tecnologías y punto de vista de las autoridades; y un conjunto de 52 casos de uso repartidos en 6 categorías: actividades de la persona asistida, reacciones del sistema, actividades del cuidador, \engagement" del usuario, actividades del desarrollador y actividades de despliegue. Como resultado del cuestionario de validación del modelo, un total de 65 personas revisó el modelo proporcionando su nivel de acuerdo con las dimensiones evaluadas y un total de 248 comentarios sobre cómo mejorar el modelo. Los conocimientos de los participantes variaban desde la ingeniería del software (70%) hasta las especialidades médicas (15%), con declarado interés en eHealth (24%), mHealth (16%), Ambient Assisted Living (21%), medicina personalizada (5%), sistemas basados en pHealth (15%), informática médica (10%) e ingeniería biomédica (8%) con una media de 7.25_4.99 años de experiencia en estas áreas. Los resultados de la encuesta muestran que los expertos contactados consideran el modelo fácil de leer (media de 1.89_0.79 siendo 1 el valor más favorable y 5 el peor), suficientemente abstracto (1.99_0.88) y formal (2.13_0.77), con una cobertura suficiente del dominio (2.26_0.95), útil para describir el dominio (2.02_0.7) y para generar sistemas más específicos (2_0.75). Los expertos también reportan un interés parcial en utilizar el modelo en su trabajo (2.48_0.91). Gracias a sus comentarios, el modelo fue mejorado y enriquecido con conceptos que faltaban, aunque no se pudo demonstrar su mejora en las dimensiones evaluadas, dada la composición diferente de personas en las tres rondas de evaluación. Desde el modelo, se generó una plataforma de desarrollo llamada \pHealth Patient Platform (pHPP)". La plataforma desarrollada incluye librerías, herramientas de programación y desarrollo, un tutorial y una aplicación de ejemplo. Se definieron cuatro módulos principales de la arquitectura: el Data Collection Engine, que permite abstraer las fuentes de datos como sensores o servicios externos, mapeando los datos a bases de datos u ontologías, y permitiendo interacción basada en eventos; el GUI Engine, que abstrae la interfaz de usuario en un modelo de interacción basado en mensajes; y el Rule Engine, que proporciona a los desarrolladores un medio simple para programar la lógica de la aplicación en forma de reglas \if-then". Después de que la plataforma pHPP fue utilizada durante 5 años en el proyecto HeartCycle, 5 desarrolladores fueron reunidos en un grupo de discusión para analizar y evaluar la plataforma. De estas evaluaciones se concluye que la plataforma fue diseñada para encajar las necesidades de los ingenieros que trabajan en la rama, permitiendo la separación de problemas entre las distintas especialidades, y simplificando algunas tareas de desarrollo como el manejo de datos y la interacción asíncrona. A pesar de ello, se encontraron algunos defectos a causa de la inmadurez de algunas tecnologías empleadas, y la ausencia de algunas herramientas específicas para el dominio como el procesado de datos o algunos protocolos de comunicación relacionados con la salud. Dentro del proyecto HeartCycle la plataforma fue utilizada para el desarrollo de la aplicación \Guided Exercise", un sistema TIC para la rehabilitación de pacientes que han sufrido un infarto del miocardio. El sistema fue testeado en un ensayo clínico randomizado en el cual a 55 pacientes se les dio el sistema para su uso por 21 semanas. De los resultados técnicos del ensayo se puede concluir que, a pesar de algunos errores menores prontamente corregidos durante el estudio, la plataforma es estable y fiable. Conclusiones La investigación llevada a cabo en esta Tesis y los resultados obtenidos proporcionan las respuestas a las tres preguntas de investigación que motivaron este trabajo: RQ1 Se ha desarrollado un modelo para representar el dominio de los sistemas personalizados de salud. La evaluación hecha por los expertos de la rama concluye que el modelo representa el dominio con precisión y con un balance apropiado entre abstracción y detalle. RQ2 Se ha desarrollado, con éxito, una plataforma de desarrollo basada en el modelo. RQ3 Se ha demostrado que la plataforma es capaz de ayudar a los desarrolladores en la creación de software pHealth complejos. Las ventajas de la plataforma han sido demostradas en el ámbito de un proyecto de gran escala, aunque el enfoque genérico adoptado indica que la plataforma podría ofrecer beneficios también en otros contextos. Los resultados de estas evaluaciones ofrecen indicios de que, ambos, el modelo y la plataforma serán buenos candidatos para poderse convertir en una referencia para futuros desarrollos de sistemas pHealth. ABSTRACT Background Europe is living in an unsustainable situation. The economic crisis has been reducing governments' economic resources since 2008 and threatening social and health systems, while the proportion of older people in the European population continues to increase so that it is foreseen that in 2050 there will be only two workers per retiree [54]. To this situation it should be added the rise, strongly related to age, of chronic diseases the burden of which has been estimated to be up to the 7% of a country's gross domestic product [51]. There is a need for a paradigm shift, the need for a new way of caring for people's health, shifting the focus from curing conditions that have arisen to a sustainable and effective approach with the emphasis on prevention. Some advocate the adoption of personalised health care (pHealth), a model where medical practices are tailored to the patient's unique life, from the detection of risk factors to the customization of treatments based on each individual's response [81]. Personalised health is often associated to the use of Information and Communications Technology (ICT), that, with its exponential development, offers interesting opportunities for improving healthcare. The shift towards pHealth is slowly taking place, both in research and in industry, but the change is not significant yet. Many barriers still exist related to economy, politics and culture, while others are purely technological, like the lack of interoperable information systems [199]. Though interoperability aspects are evolving, there is still the need of a reference design, especially tackling implementation and large scale deployment of pHealth systems. This thesis contributes to organizing the subject of ICT systems for personalised health into a reference model that allows for the creation of software development platforms to ease common development issues in the domain. Research questions RQ1 Is it possible to define a model, based on software engineering techniques, for representing the personalised health domain in an abstract and representative way? RQ2 Is it possible to build a development platform based on this model? RQ3 Does the development platform help developers create complex integrated pHealth systems? Methods As method for describing the model, the ISO/IEC/IEEE 42010 framework [25] is adopted for its generality and high level of abstraction. The model is specified in different parts: a conceptual model, which makes use of concept maps, for representing stakeholders, artefacts and shared information, and in scenarios and use cases for the representation of the functionalities of pHealth systems. The model was derived from literature analysis, including 7 industrial and scientific reports, 9 electronic standards, 10 conference proceedings papers, 37 journal papers, 25 websites and 5 books. Based on the reference model, requirements were drawn for building the development platform enriched with a set of requirements gathered in a survey run among 11 experienced engineers. For developing the platform, the continuous integration methodology [74] was adopted which allowed to perform automatic tests on a server and also to deploy packaged releases on a web site. As a validation methodology, a theory building framework for SW engineering was adopted from [181]. The framework, chosen as a guide to find evidence for justifying the research questions, imposed the creation of theories based on models and propositions to be validated within a scope. The validation of the model was conducted as an on-line survey in three validation rounds, encompassing a growing number of participants. The survey was submitted to 134 experts of the field and on some public channels like relevant mailing lists and social networks. Its objective was to assess the model's readability, its level of coverage of the domain and its potential usefulness in the design of actual, derived systems. The questionnaires included quantitative Likert scale questions and free text inputs for comments. The development platform was validated in two scopes. As a small-scale experiment, the platform was used in a 12 hours training session where 4 developers had to perform an exercise consisting in developing a set of typical pHealth use cases At the end of the session, a focus group was held to identify benefits and drawbacks of the platform. The second validation was held as a test-case study in a large scale research project called HeartCycle the aim of which was to develop a closed-loop disease management system for heart failure and coronary heart disease patients [160]. During this project three applications were developed by a team of programmers and designers. One of these applications was tested in a clinical trial with actual patients. At the end of the project, the team was interviewed in a focus group to assess the role the platform had within the project. Results For what regards the model that describes the pHealth domain, its conceptual part includes a description of the main roles and concerns of pHealth stakeholders, a model of the ICT artefacts that are commonly adopted and a model representing the typical data that need to be formalized among pHealth systems. The functional model includes a set of 18 scenarios, divided into assisted person's view, caregiver's view, developer's view, technology and services providers' view and authority's view, and a set of 52 Use Cases grouped in 6 categories: assisted person's activities, system reactions, caregiver's activities, user engagement, developer's activities and deployer's activities. For what concerns the validation of the model, a total of 65 people participated in the online survey providing their level of agreement in all the assessed dimensions and a total of 248 comments on how to improve and complete the model. Participants' background spanned from engineering and software development (70%) to medical specialities (15%), with declared interest in the fields of eHealth (24%), mHealth (16%), Ambient Assisted Living (21%), Personalized Medicine (5%), Personal Health Systems (15%), Medical Informatics (10%) and Biomedical Engineering (8%) with an average of 7.25_4.99 years of experience in these fields. From the analysis of the answers it is possible to observe that the contacted experts considered the model easily readable (average of 1.89_0.79 being 1 the most favourable scoring and 5 the worst), sufficiently abstract (1.99_0.88) and formal (2.13_0.77) for its purpose, with a sufficient coverage of the domain (2.26_0.95), useful for describing the domain (2.02_0.7) and for generating more specific systems (2_0.75) and they reported a partial interest in using the model in their job (2.48_0.91). Thanks to their comments, the model was improved and enriched with concepts that were missing at the beginning, nonetheless it was not possible to prove an improvement among the iterations, due to the diversity of the participants in the three rounds. From the model, a development platform for the pHealth domain was generated called pHealth Patient Platform (pHPP). The platform includes a set of libraries, programming and deployment tools, a tutorial and a sample application. The main four modules of the architecture are: the Data Collection Engine, which allows abstracting sources of information like sensors or external services, mapping data to databases and ontologies, and allowing event-based interaction and filtering, the GUI Engine, which abstracts the user interface in a message-like interaction model, the Workow Engine, which allows programming the application's user interaction ows with graphical workows, and the Rule Engine, which gives developers a simple means for programming the application's logic in the form of \if-then" rules. After the 5 years experience of HeartCycle, partially programmed with pHPP, 5 developers were joined in a focus group to discuss the advantages and drawbacks of the platform. The view that emerged from the training course and the focus group was that the platform is well-suited to the needs of the engineers working in the field, it allowed the separation of concerns among the different specialities and it simplified some common development tasks like data management and asynchronous interaction. Nevertheless, some deficiencies were pointed out in terms of a lack of maturity of some technological choices, and for the absence of some domain-specific tools, e.g. for data processing or for health-related communication protocols. Within HeartCycle, the platform was used to develop part of the Guided Exercise system, a composition of ICT tools for the physical rehabilitation of patients who suffered from myocardial infarction. The system developed using the platform was tested in a randomized controlled clinical trial, in which 55 patients used the system for 21 weeks. The technical results of this trial showed that the system was stable and reliable. Some minor bugs were detected, but these were promptly corrected using the platform. This shows that the platform, as well as facilitating the development task, can be successfully used to produce reliable software. Conclusions The research work carried out in developing this thesis provides responses to the three three research questions that were the motivation for the work. RQ1 A model was developed representing the domain of personalised health systems, and the assessment of experts in the field was that it represents the domain accurately, with an appropriate balance between abstraction and detail. RQ2 A development platform based on the model was successfully developed. RQ3 The platform has been shown to assist developers create complex pHealth software. This was demonstrated within the scope of one large-scale project, but the generic approach adopted provides indications that it would offer benefits more widely. The results of these evaluations provide indications that both the model and the platform are good candidates for being a reference for future pHealth developments.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Pressure measurements on the surface of a 1:230 scale model of Bolund Island are presented. The model is smooth and no boundary layer generation has been considered since the experiment is designed as the simplest possible reference case. Measurement have been taken for a range of Reynolds numbers based on the average undisturbed wind speed U? and the maximum height of the island, h [1.7×104 , 8.5×104], and for a range of wind directions. Four minutes time series of pressure in more than 400 points have been acquired and analysed to obtain the spatial distribution of both the time average and the variance of the pressure signal. The horizontal extension of the detachment bubble for the different Reynolds numbers and wind directions is identified by isobars and curves of constant value of pressure variance. The applicability of this technique for evaluating the horizontal topology of high turbulence regions associated to detachment bubbles after escarpments in potential wind farm sites is analysed. The results obtained shows that the behaviour of the mean pressure coefficient, Cp, the std. pressure coefficient, C?p, and the skewness of the pressure, Sp can be used to study the bubble over the island to a certain extent. This experiment is part of the set of different analysis on the Bolund test case that is being undertaken within WAUDIT project by the different scientific groups.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper assesses the main challenges associated with the propagation and channel modeling of broadband radio systems in a complex environment of high speed and metropolitan railways. These challenges comprise practical simulation, modeling interferences, radio planning, test trials and performance evaluation in different railway scenarios using Long Term Evolution (LTE) as test case. This approach requires several steps; the first is the use of a radio propagation simulator based on ray-tracing techniques to accurately predict propagation. Besides the radio propagation simulator, a complete test bed has been constructed to assess LTE performance, channel propagation conditions and interference with other systems in real-world environments by means of standard-compliant LTE transmissions. Such measurement results allowed us to evaluate the propagation and performance of broadband signals and to test the suitability of LTE radio technology for complex railway scenarios.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

La realidad aumentada educativa es una tecnología que actualmente está mejorando la calidad de enseñanza, la utilización de dispositivos móviles permite que el estudiante sea protagonista de su aprendizaje sin estar confinado a un espacio o tiempo específico para aprender. Aplicaciones colaborativas con realidad aumentada están siendo empleadas cada vez más en la educación, de tal forma que fomentan el trabajo en grupo donde los estudiantes comparten conocimiento, dudas, opiniones logrando un mejor nivel cognitivo que trabajando individualmente. En este trabajo se presenta el estado de la cuestión de Aplicaciones Educativas con Realidad Aumentada en dispositivos móviles, y Aplicaciones Educativas colaborativas con Realidad Aumentada, desarrolladas desde el 2002 e implementadas en instituciones educativas. Así mismo se realiza un estudio sobre la Realidad Aumentada, Realidad Aumentada móvil y Aprendizaje Móvil. Además, a partir de las características del estudio de las aplicaciones con Realidad Aumenta, se realiza un análisis y diseño de una Aplicación Móvil para el proyecto de inicio de los alumnos de nuevo ingreso de la UPM. Así como también una herramienta de autoría para las gestiones de las actividades propuestas por los docentes de la UPM. Finalmente se presenta un caso de prueba en el que se implementa parte de la propuesta de este trabajo, logrando construir un parte funcional para el proyecto inicial denominado PIANI – UPM. ---ABSTRACT---Educational Augmented reality is a technology that is improving the quality of teaching, use of mobile devices enables the student to be protagonists of their learning without being confined to a specific space or time to learn. Collaborative augmented reality applications applied in education are being used gradually encourage group work where students share knowledge, doubts, opinions so they achieve better cognitive level than working individually. In this paper the description of educational applications is presented with Augmented Reality using mobile devices, and collaborative educational Augmented Reality applications, developed since 2002 and implemented in educational institutions. Also a study on Augmented Reality, Mobile Augmented Reality and Mobile Learning is performed. Furthermore, from the study of the characteristics of Reality applications increases, an analysis and design of a mobile application for the proposed start of new students of UPM is performed. As well as an authoring tool for the efforts of the activities proposed by the teachers of the UPM. Finally a test case is presented in which part of the proposal of this work is implemented, obtaining building an initial prototype called PIANI - UPM.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The paper presents the application of a new risk-informed methodology for the identification of the Emergency Management Requirements (EMR) to a Generation II, Large size Reactor and a Generation III+ Small Modular Reactor. The results obtained in this test case demonstrate that the actual EMR is conservative, as expected, for the GenII reactor, while the new methodology could be applied for the definition of EMRs for the new generation Nuclear Power Plants, with a possible reduction of the emergency area without loss of safety level. By adopting both probabilistic and deterministic approaches, the study addresses possible accidents and corresponding release scenarios for the two types of reactor, calculates the areas where the accidents have an impact on the population and defines the new EMR considering the health effects on the population.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The purpose of this work is to analyze a complex high lift configuration for which significant regions of separated flow are present. Current state of the art methods have some diffculty to predict the origin and the progression of this separated flow when increasing the angle of attack. The mechanisms responsible for the maximum lift limit on multi-element wing con?gurations are not clear; this stability analysis could help to understand the physics behind the phenomenon and to find a relation between the flow separation and the instability onset. The methodology presented herein consists in the computation of a steady base flow solution based on a finite volume discretization and a proposal of the solution for a generalized eigenvalue problem corresponding to the perturbed and linearized problem. The eigenvalue problem has been solved with the Arnoldi iterative method, one of the Krylov subspace projection methods. The described methodology was applied to the NACA0012 test case in subsonic and in transonic conditions and, finally, for the first time to the authors knowledge, on an industrial multi-component geometry, such as the A310 airfoil, in order to identify low frequency instabilities related to the separation. One important conclusion is that for all the analyzed geometries, one unstable mode related to flow separation appears for an angle of attack greater than the one correspondent to the maximum lift coe?cient condition. Finally, an adjoint study was carried out in order to evaluate the receptivity and the structural sensitivity of the geometries, giving an indication of the domain region that could be modified resulting in the biggest change of the flowfield.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Los ataques a redes de información son cada vez más sofisticados y exigen una constante evolución y mejora de las técnicas de detección. Para ello, en este proyecto se ha diseñado e implementado una plataforma cooperativa para la detección de intrusiones basada en red. En primer lugar, se ha realizado un estudio teórico previo del marco tecnológico relacionado con este ámbito, en el que se describe y caracteriza el software que se utiliza para realizar ataques a sistemas (malware) así como los métodos que se utilizan para llegar a transmitir ese software (vectores de ataque). En el documento también se describen los llamados APT, que son ataques dirigidos con una gran inversión económica y temporal. Estos pueden englobar todos los malware y vectores de ataque existentes. Para poder evitar estos ataques, se estudiarán los sistemas de detección y prevención de intrusiones, describiendo brevemente los algoritmos que se tienden a utilizar en la actualidad. En segundo lugar, se ha planteado y desarrollado una plataforma en red dedicada al análisis de paquetes y conexiones para detectar posibles intrusiones. Este sistema está orientado a sistemas SCADA (Supervisory Control And Data Adquisition) aunque funciona sobre cualquier red IPv4/IPv6, para ello se definirá previamente lo que es un sistema SCADA, así como sus partes principales. Para implementar el sistema se han utilizado dispositivos de bajo consumo llamados Raspberry PI, estos se ubican entre la red y el equipo final que se quiera analizar. En ellos se ejecutan 2 aplicaciones desarrolladas de tipo cliente-servidor (la Raspberry central ejecutará la aplicación servidora y las esclavas la aplicación cliente) que funcionan de forma cooperativa utilizando la tecnología distribuida de Hadoop, la cual se explica previamente. Mediante esta tecnología se consigue desarrollar un sistema completamente escalable. La aplicación servidora muestra una interfaz gráfica que permite administrar la plataforma de análisis de forma centralizada, pudiendo ver así las alarmas de cada dispositivo y calificando cada paquete según su peligrosidad. El algoritmo desarrollado en la aplicación calcula el ratio de paquetes/tiempo que entran/salen del equipo final, procesando los paquetes y analizándolos teniendo en cuenta la información de señalización, creando diferentes bases de datos que irán mejorando la robustez del sistema, reduciendo así la posibilidad de ataques externos. Para concluir, el proyecto inicial incluía el procesamiento en la nube de la aplicación principal, pudiendo administrar así varias infraestructuras concurrentemente, aunque debido al trabajo extra necesario se ha dejado preparado el sistema para poder implementar esta funcionalidad. En el caso experimental actual el procesamiento de la aplicación servidora se realiza en la Raspberry principal, creando un sistema escalable, rápido y tolerante a fallos. ABSTRACT. The attacks to networks of information are increasingly sophisticated and demand a constant evolution and improvement of the technologies of detection. For this project it is developed and implemented a cooperative platform for detect intrusions based on networking. First, there has been a previous theoretical study of technological framework related to this area, which describes the software used for attacks on systems (malware) as well as the methods used in order to transmit this software (attack vectors). In this document it is described the APT, which are attacks directed with a big economic and time inversion. These can contain all existing malware and attack vectors. To prevent these attacks, intrusion detection systems and prevention intrusion systems will be discussed, describing previously the algorithms tend to use today. Secondly, a platform for analyzing network packets has been proposed and developed to detect possible intrusions in SCADA (Supervisory Control And Data Adquisition) systems. This platform is designed for SCADA systems (Supervisory Control And Data Acquisition) but works on any IPv4 / IPv6 network. Previously, it is defined what a SCADA system is and the main parts of it. To implement it, we used low-power devices called Raspberry PI, these are located between the network and the final device to analyze it. In these Raspberry run two applications client-server developed (the central Raspberry runs the server application and the slaves the client application) that work cooperatively using Hadoop distributed technology, which is previously explained. Using this technology is achieved develop a fully scalable system. The server application displays a graphical interface to manage analytics platform centrally, thereby we can see each device alarms and qualifying each packet by dangerousness. The algorithm developed in the application calculates the ratio of packets/time entering/leaving the terminal device, processing the packets and analyzing the signaling information of each packet, reating different databases that will improve the system, thereby reducing the possibility of external attacks. In conclusion, the initial project included cloud computing of the main application, being able to manage multiple concurrent infrastructure, but due to the extra work required has been made ready the system to implement this funcionality. In the current test case the server application processing is made on the main Raspberry, creating a scalable, fast and fault-tolerant system.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this work a p-adaptation (modification of the polynomial order) strategy based on the minimization of the truncation error is developed for high order discontinuous Galerkin methods. The truncation error is approximated by means of a truncation error estimation procedure and enables the identification of mesh regions that require adaptation. Three truncation error estimation approaches are developed and termed a posteriori, quasi-a priori and quasi-a priori corrected. Fine solutions, which are obtained by enriching the polynomial order, are required to solve the numerical problem with adequate accuracy. For the three truncation error estimation methods the former needs time converged solutions, while the last two rely on non-converged solutions, which lead to faster computations. Based on these truncation error estimation methods, algorithms for mesh adaptation were designed and tested. Firstly, an isotropic adaptation approach is presented, which leads to equally distributed polynomial orders in different coordinate directions. This first implementation is improved by incorporating a method to extrapolate the truncation error. This results in a significant reduction of computational cost. Secondly, the employed high order method permits the spatial decoupling of the estimated errors and enables anisotropic p-adaptation. The incorporation of anisotropic features leads to meshes with different polynomial orders in the different coordinate directions such that flow-features related to the geometry are resolved in a better manner. These adaptations result in a significant reduction of degrees of freedom and computational cost, while the amount of improvement depends on the test-case. Finally, this anisotropic approach is extended by using error extrapolation which leads to an even higher reduction in computational cost. These strategies are verified and compared in terms of accuracy and computational cost for the Euler and the compressible Navier-Stokes equations. The main result is that the two quasi-a priori methods achieve a significant reduction in computational cost when compared to a uniform polynomial enrichment. Namely, for a viscous boundary layer flow, we obtain a speedup of a factor of 6.6 and 7.6 for the quasi-a priori and quasi-a priori corrected approaches, respectively. RESUMEN En este trabajo se ha desarrollado una estrategia de adaptación-p (modificación del orden polinómico) para métodos Galerkin discontinuo de alto orden basada en la minimización del error de truncación. El error de truncación se estima utilizando el método tau-estimation. El estimador permite la identificación de zonas de la malla que requieren adaptación. Se distinguen tres técnicas de estimación: a posteriori, quasi a priori y quasi a priori con correción. Todas las estrategias requieren una solución obtenida en una malla fina, la cual es obtenida aumentando de manera uniforme el orden polinómico. Sin embargo, mientras que el primero requiere que esta solución esté convergida temporalmente, el resto utiliza soluciones no convergidas, lo que se traduce en un menor coste computacional. En este trabajo se han diseñado y probado algoritmos de adaptación de malla basados en métodos tau-estimation. En primer lugar, se presenta un algoritmo de adaptacin isótropo, que conduce a discretizaciones con el mismo orden polinómico en todas las direcciones espaciales. Esta primera implementación se mejora incluyendo un método para extrapolar el error de truncación. Esto resulta en una reducción significativa del coste computacional. En segundo lugar, el método de alto orden permite el desacoplamiento espacial de los errores estimados, permitiendo la adaptación anisotropica. Las mallas obtenidas mediante esta técnica tienen distintos órdenes polinómicos en cada una de las direcciones espaciales. La malla final tiene una distribución óptima de órdenes polinómicos, los cuales guardan relación con las características del flujo que, a su vez, depenen de la geometría. Estas técnicas de adaptación reducen de manera significativa los grados de libertad y el coste computacional. Por último, esta aproximación anisotropica se extiende usando extrapolación del error de truncación, lo que conlleva un coste computational aún menor. Las estrategias se verifican y se comparan en téminors de precisión y coste computacional utilizando las ecuaciones de Euler y Navier Stokes. Los dos métodos quasi a priori consiguen una reducción significativa del coste computacional en comparación con aumento uniforme del orden polinómico. En concreto, para una capa límite viscosa, obtenemos una mejora en tiempo de computación de 6.6 y 7.6 respectivamente, para las aproximaciones quasi-a priori y quasi-a priori con corrección.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Este trabajo presenta un método discreto para el cálculo de estabilidad hidrodinámica y análisis de sensibilidad a perturbaciones externas para ecuaciones diferenciales y en particular para las ecuaciones de Navier-Stokes compressible. Se utiliza una aproximación con variable compleja para obtener una precisión analítica en la evaluación de la matriz Jacobiana. Además, mapas de sensibilidad para la sensibilidad a las modificaciones del flujo de base y a una fuerza constante permiten identificar las regiones del campo fluido donde una modificacin (ej. fuerza puntual) tiene un efecto estabilizador del flujo. Se presentan cuatro casos de prueba: (1) un caso analítico para comprobar la derivación discreta, (2) una cavidad cerrada a bajo Reynolds para mostrar la mayor precisión en el cálculo de los valores propios con la aproximación de paso complejo, (3) flujo 2D en un cilindro circular para validar la metodología, y (4) flujo en un cavidad abierta, presentado para validar el método en casos de inestabilidades convectivamente inestables. Los tres últimos casos mencionados (2-4) se resolvieron con las ecuaciones de Navier-Stokes compresibles, utilizando un método Discontinuous Galerkin Spectral Element Method. Se obtuvo una buena concordancia para el caso de validación (3), cuando se comparó el nuevo método con resultados de la literatura. Además, este trabajo muestra que para el cálculo de los modos propios directos y adjuntos, así como para los mapas de sensibilidad, el uso de variables complejas es de suprema importancia para obtener una predicción precisa. El método descrito es aplicado al análisis para la estabilización de la estela generada por un disco actuador, que representa un modelo sencillo para hélices, rotores de helicópteros o turbinas eólicas. Se explora la primera bifurcación del flujo para un disco actuador, y se sugiere que está asociada a una inestabilidad de tipo Kelvin-Helmholtz, cuya estabilidad se controla con en el número de Reynolds y en la resistencia del disco actuador (o fuerza resistente). En primer lugar, se verifica que la disminución de la resistencia del disco tiene un efecto estabilizador parecido a una disminución del Reynolds. En segundo lugar, el análisis hidrodinmico discreto identifica dos regiones para la colocación de una fuerza puntual que controle las inestabilidades, una cerca del disco y otra en una zona aguas abajo. En tercer lugar, se muestra que la inclusión de un forzamiento localizado cerca del actuador produce una estabilización más eficiente que al forzar aguas abajo. El análisis de los campos de flujo controlados confirma que modificando el gradiente de velocidad cerca del actuador es más eficiente para estabilizar la estela. Estos resultados podrían proporcionar nuevas directrices para la estabilización de la estela de turbinas de viento o de marea cuando estén instaladas en un parque eólico y minimizar las interacciones no estacionarias entre turbinas. ABSTRACT A discrete framework for computing the global stability and sensitivity analysis to external perturbations for any set of partial differential equations is presented. In particular, a complex-step approximation is used to achieve near analytical accuracy for the evaluation of the Jacobian matrix. Sensitivity maps for the sensitivity to base flow modifications and to a steady force are computed to identify regions of the flow field where an input could have a stabilising effect. Four test cases are presented: (1) an analytical test case to prove the theory of the discrete framework, (2) a lid-driven cavity at low Reynolds case to show the improved accuracy in the calculation of the eigenvalues when using the complex-step approximation, (3) the 2D flow past a circular cylinder at just below the critical Reynolds number is used to validate the methodology, and finally, (4) the flow past an open cavity is presented to give an example of the discrete method applied to a convectively unstable case. The latter three (2–4) of the aforementioned cases were solved with the 2D compressible Navier–Stokes equations using a Discontinuous Galerkin Spectral Element Method. Good agreement was obtained for the validation test case, (3), with appropriate results in the literature. Furthermore, it is shown that for the calculation of the direct and adjoint eigenmodes and their sensitivity maps to external perturbations, the use of complex variables is paramount for obtaining an accurate prediction. An analysis for stabilising the wake past an actuator disc, which represents a simple model for propellers, helicopter rotors or wind turbines is also presented. We explore the first flow bifurcation for an actuator disc and it suggests that it is associated to a Kelvin- Helmholtz type instability whose stability relies on the Reynolds number and the flow resistance applied through the disc (or actuator forcing). First, we report that decreasing the disc resistance has a similar stabilising effect to an decrease in the Reynolds number. Second, a discrete sensitivity analysis identifies two regions for suitable placement of flow control forcing, one close to the disc and one far downstream where the instability originates. Third, we show that adding a localised forcing close to the actuator provides more stabilisation that forcing far downstream. The analysis of the controlled flow fields, confirms that modifying the velocity gradient close to the actuator is more efficient to stabilise the wake than controlling the sheared flow far downstream. An interesting application of these results is to provide guidelines for stabilising the wake of wind or tidal turbines when placed in an energy farm to minimise unsteady interactions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Molecular differentiation between races or closely related species is often incongruent with the reproductive divergence of the taxa of interest. Shared ancient polymorphism and/or introgression during secondary contact may be responsible for the incongruence. At loci contributing to speciation, these two complications should be minimized (1, 2); hence, their variation may more faithfully reflect the history of the species' reproductive differentiation. In this study, we analyzed DNA polymorphism at the Odysseus (OdsH) locus of hybrid sterility between Drosophila mauritiana and Drosophila simulans and were able to verify such a prediction. Interestingly, DNA variation only a short distance away (1.8 kb) appears not to be influenced by the forces that shape the recent evolution of the OdsH coding region. This locus thus may represent a test case of inferring phylogeny of very closely related species.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Plasma levels of corticosterone are often used as a measure of “stress” in wild animal populations. However, we lack conclusive evidence that different stress levels reflect different survival probabilities between populations. Galápagos marine iguanas offer an ideal test case because island populations are affected differently by recurring El Niño famine events, and population-level survival can be quantified by counting iguanas locally. We surveyed corticosterone levels in six populations during the 1998 El Niño famine and the 1999 La Niña feast period. Iguanas had higher baseline and handling stress-induced corticosterone concentrations during famine than feast conditions. Corticosterone levels differed between islands and predicted survival through an El Niño period. However, among individuals, baseline corticosterone was only elevated when body condition dropped below a critical threshold. Thus, the population-level corticosterone response was variable but nevertheless predicted overall population health. Our results lend support to the use of corticosterone as a rapid quantitative predictor of survival in wild animal populations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We have demonstrated the assembly of two-dimensional patterns of functional antibodies on a surface. In particular, we have selectively adsorbed micrometer-scale regions of biotinylated immunoglobulin that exhibit specific antigen binding after adsorption. The advantage of this technique is its potential adaptability to adsorbing arbitrary proteins in tightly packed monolayers while retaining functionality. The procedure begins with the formation of a self-assembled monolayer of n-octadecyltrimethoxysilane (OTMS) on a silicon dioxide surface. This monolayer can then be selectively removed by UV photolithography. Under appropriate solution conditions, the OTMS regions will adsorb a monolayer of bovine serum albumin (BSA), while the silicon dioxide regions where the OTMS has been removed by UV light will adsorb less than 2% of a monolayer, thus creating high contrast patterned adsorption of BSA. The attachment of the molecule biotin to the BSA allows the pattern to be replicated in a layer of streptavidin, which bonds to the biotinylated BSA and in turn will bond an additional layer of an arbitrary biotinylated protein. In our test case, functionality of the biotinylated goat antibodies raised against mouse immunoglobulin was demonstrated by the specific binding of fluorescently labeled mouse IgG.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A capillary electrophoresis method has been developed to study DNA-protein complexes by mobility-shift assay. This method is at least 100 times more sensitive than conventional gel mobility-shift procedures. Key features of the technique include the use of a neutral coated capillary, a small amount of linear polymer in the separation medium, and use of covalently dye-labeled DNA probes that can be detected with a commercially available laser-induced fluorescence monitor. The capillary method provides quantitative data in runs requiring < 20 min, from which dissociation constants are readily determined. As a test case we studied interactions of a developmentally important sea urchin embryo transcription factor, SpP3A2. As little as 2-10 x 10(6) molecules of specific SpP3A2-oligonucleotide complex were reproducibly detected, using recombinant SpP3A2, crude nuclear extract, egg lysates, and even a single sea urchin egg lysed within the capillary column.