46 resultados para sustainable assessment tools


Relevância:

30.00% 30.00%

Publicador:

Resumo:

An analysis and comparison of daily and yearly solar irradiation from the satellite CM SAF database and a set of 301 stations from the Spanish SIAR network is performed using data of 2010 and 2011. This analysis is completed with the comparison of the estimations of effective irradiation incident on three different tilted planes (fixed, two axis tracking, north-south hori- zontal axis) using irradiation from these two data sources. Finally, a new map of yearly values of irradiation both on the horizontal plane and on inclined planes is produced mixing both sources with geostatistical techniques (kriging with external drift, KED) The Mean Absolute Difference (MAD) between CM SAF and SIAR is approximately 4% for the irradiation on the horizontal plane and is comprised between 5% and 6% for the irradiation incident on the inclined planes. The MAD between KED and SIAR, and KED and CM SAF is approximately 3% for the irradiation on the horizontal plane and is comprised between 3% and 4% for the irradiation incident on the inclined planes. The methods have been implemented using free software, available as supplementary ma- terial, and the data sources are freely available without restrictions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The design of nuclear power plant has to follow a number of regulations aimed at limiting the risks inherent in this type of installation. The goal is to prevent and to limit the consequences of any possible incident that might threaten the public or the environment. To verify that the safety requirements are met a safety assessment process is followed. Safety analysis is as key component of a safety assessment, which incorporates both probabilistic and deterministic approaches. The deterministic approach attempts to ensure that the various situations, and in particular accidents, that are considered to be plausible, have been taken into account, and that the monitoring systems and engineered safety and safeguard systems will be capable of ensuring the safety goals. On the other hand, probabilistic safety analysis tries to demonstrate that the safety requirements are met for potential accidents both within and beyond the design basis, thus identifying vulnerabilities not necessarily accessible through deterministic safety analysis alone. Probabilistic safety assessment (PSA) methodology is widely used in the nuclear industry and is especially effective in comprehensive assessment of the measures needed to prevent accidents with small probability but severe consequences. Still, the trend towards a risk informed regulation (RIR) demanded a more extended use of risk assessment techniques with a significant need to further extend PSA’s scope and quality. Here is where the theory of stimulated dynamics (TSD) intervenes, as it is the mathematical foundation of the integrated safety assessment (ISA) methodology developed by the CSN(Consejo de Seguridad Nuclear) branch of Modelling and Simulation (MOSI). Such methodology attempts to extend classical PSA including accident dynamic analysis, an assessment of the damage associated to the transients and a computation of the damage frequency. The application of this ISA methodology requires a computational framework called SCAIS (Simulation Code System for Integrated Safety Assessment). SCAIS provides accident dynamic analysis support through simulation of nuclear accident sequences and operating procedures. Furthermore, it includes probabilistic quantification of fault trees and sequences; and integration and statistic treatment of risk metrics. SCAIS comprehensively implies an intensive use of code coupling techniques to join typical thermal hydraulic analysis, severe accident and probability calculation codes. The integration of accident simulation in the risk assessment process and thus requiring the use of complex nuclear plant models is what makes it so powerful, yet at the cost of an enormous increase in complexity. As the complexity of the process is primarily focused on such accident simulation codes, the question of whether it is possible to reduce the number of required simulation arises, which will be the focus of the present work. This document presents the work done on the investigation of more efficient techniques applied to the process of risk assessment inside the mentioned ISA methodology. Therefore such techniques will have the primary goal of decreasing the number of simulation needed for an adequate estimation of the damage probability. As the methodology and tools are relatively recent, there is not much work done inside this line of investigation, making it a quite difficult but necessary task, and because of time limitations the scope of the work had to be reduced. Therefore, some assumptions were made to work in simplified scenarios best suited for an initial approximation to the problem. The following section tries to explain in detail the process followed to design and test the developed techniques. Then, the next section introduces the general concepts and formulae of the TSD theory which are at the core of the risk assessment process. Afterwards a description of the simulation framework requirements and design is given. Followed by an introduction to the developed techniques, giving full detail of its mathematical background and its procedures. Later, the test case used is described and result from the application of the techniques is shown. Finally the conclusions are presented and future lines of work are exposed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Antecedentes Europa vive una situación insostenible. Desde el 2008 se han reducido los recursos de los gobiernos a raíz de la crisis económica. El continente Europeo envejece con ritmo constante al punto que se prevé que en 2050 habrá sólo dos trabajadores por jubilado [54]. A esta situación se le añade el aumento de la incidencia de las enfermedades crónicas, relacionadas con el envejecimiento, cuyo coste puede alcanzar el 7% del PIB de un país [51]. Es necesario un cambio de paradigma. Una nueva manera de cuidar de la salud de las personas: sustentable, eficaz y preventiva más que curativa. Algunos estudios abogan por el cuidado personalizado de la salud (pHealth). En este modelo las prácticas médicas son adaptadas e individualizadas al paciente, desde la detección de los factores de riesgo hasta la personalización de los tratamientos basada en la respuesta del individuo [81]. El cuidado personalizado de la salud está asociado a menudo al uso de las tecnologías de la información y comunicación (TICs) que, con su desarrollo exponencial, ofrecen oportunidades interesantes para la mejora de la salud. El cambio de paradigma hacia el pHealth está lentamente ocurriendo, tanto en el ámbito de la investigación como en la industria, pero todavía no de manera significativa. Existen todavía muchas barreras relacionadas a la economía, a la política y la cultura. También existen barreras puramente tecnológicas, como la falta de sistemas de información interoperables [199]. A pesar de que los aspectos de interoperabilidad están evolucionando, todavía hace falta un diseño de referencia especialmente direccionado a la implementación y el despliegue en gran escala de sistemas basados en pHealth. La presente Tesis representa un intento de organizar la disciplina de la aplicación de las TICs al cuidado personalizado de la salud en un modelo de referencia, que permita la creación de plataformas de desarrollo de software para simplificar tareas comunes de desarrollo en este dominio. Preguntas de investigación RQ1 >Es posible definir un modelo, basado en técnicas de ingeniería del software, que represente el dominio del cuidado personalizado de la salud de una forma abstracta y representativa? RQ2 >Es posible construir una plataforma de desarrollo basada en este modelo? RQ3 >Esta plataforma ayuda a los desarrolladores a crear sistemas pHealth complejos e integrados? Métodos Para la descripción del modelo se adoptó el estándar ISO/IEC/IEEE 42010por ser lo suficientemente general y abstracto para el amplio enfoque de esta tesis [25]. El modelo está definido en varias partes: un modelo conceptual, expresado a través de mapas conceptuales que representan las partes interesadas (stakeholders), los artefactos y la información compartida; y escenarios y casos de uso para la descripción de sus funcionalidades. El modelo fue desarrollado de acuerdo a la información obtenida del análisis de la literatura, incluyendo 7 informes industriales y científicos, 9 estándares, 10 artículos en conferencias, 37 artículos en revistas, 25 páginas web y 5 libros. Basándose en el modelo se definieron los requisitos para la creación de la plataforma de desarrollo, enriquecidos por otros requisitos recolectados a través de una encuesta realizada a 11 ingenieros con experiencia en la rama. Para el desarrollo de la plataforma, se adoptó la metodología de integración continua [74] que permitió ejecutar tests automáticos en un servidor y también desplegar aplicaciones en una página web. En cuanto a la metodología utilizada para la validación se adoptó un marco para la formulación de teorías en la ingeniería del software [181]. Esto requiere el desarrollo de modelos y proposiciones que han de ser validados dentro de un ámbito de investigación definido, y que sirvan para guiar al investigador en la búsqueda de la evidencia necesaria para justificarla. La validación del modelo fue desarrollada mediante una encuesta online en tres rondas con un número creciente de invitados. El cuestionario fue enviado a 134 contactos y distribuido en algunos canales públicos como listas de correo y redes sociales. El objetivo era evaluar la legibilidad del modelo, su nivel de cobertura del dominio y su potencial utilidad en el diseño de sistemas derivados. El cuestionario incluía preguntas cuantitativas de tipo Likert y campos para recolección de comentarios. La plataforma de desarrollo fue validada en dos etapas. En la primera etapa se utilizó la plataforma en un experimento a pequeña escala, que consistió en una sesión de entrenamiento de 12 horas en la que 4 desarrolladores tuvieron que desarrollar algunos casos de uso y reunirse en un grupo focal para discutir su uso. La segunda etapa se realizó durante los tests de un proyecto en gran escala llamado HeartCycle [160]. En este proyecto un equipo de diseñadores y programadores desarrollaron tres aplicaciones en el campo de las enfermedades cardio-vasculares. Una de estas aplicaciones fue testeada en un ensayo clínico con pacientes reales. Al analizar el proyecto, el equipo de desarrollo se reunió en un grupo focal para identificar las ventajas y desventajas de la plataforma y su utilidad. Resultados Por lo que concierne el modelo que describe el dominio del pHealth, la parte conceptual incluye una descripción de los roles principales y las preocupaciones de los participantes, un modelo de los artefactos TIC que se usan comúnmente y un modelo para representar los datos típicos que son necesarios formalizar e intercambiar entre sistemas basados en pHealth. El modelo funcional incluye un conjunto de 18 escenarios, repartidos en: punto de vista de la persona asistida, punto de vista del cuidador, punto de vista del desarrollador, punto de vista de los proveedores de tecnologías y punto de vista de las autoridades; y un conjunto de 52 casos de uso repartidos en 6 categorías: actividades de la persona asistida, reacciones del sistema, actividades del cuidador, \engagement" del usuario, actividades del desarrollador y actividades de despliegue. Como resultado del cuestionario de validación del modelo, un total de 65 personas revisó el modelo proporcionando su nivel de acuerdo con las dimensiones evaluadas y un total de 248 comentarios sobre cómo mejorar el modelo. Los conocimientos de los participantes variaban desde la ingeniería del software (70%) hasta las especialidades médicas (15%), con declarado interés en eHealth (24%), mHealth (16%), Ambient Assisted Living (21%), medicina personalizada (5%), sistemas basados en pHealth (15%), informática médica (10%) e ingeniería biomédica (8%) con una media de 7.25_4.99 años de experiencia en estas áreas. Los resultados de la encuesta muestran que los expertos contactados consideran el modelo fácil de leer (media de 1.89_0.79 siendo 1 el valor más favorable y 5 el peor), suficientemente abstracto (1.99_0.88) y formal (2.13_0.77), con una cobertura suficiente del dominio (2.26_0.95), útil para describir el dominio (2.02_0.7) y para generar sistemas más específicos (2_0.75). Los expertos también reportan un interés parcial en utilizar el modelo en su trabajo (2.48_0.91). Gracias a sus comentarios, el modelo fue mejorado y enriquecido con conceptos que faltaban, aunque no se pudo demonstrar su mejora en las dimensiones evaluadas, dada la composición diferente de personas en las tres rondas de evaluación. Desde el modelo, se generó una plataforma de desarrollo llamada \pHealth Patient Platform (pHPP)". La plataforma desarrollada incluye librerías, herramientas de programación y desarrollo, un tutorial y una aplicación de ejemplo. Se definieron cuatro módulos principales de la arquitectura: el Data Collection Engine, que permite abstraer las fuentes de datos como sensores o servicios externos, mapeando los datos a bases de datos u ontologías, y permitiendo interacción basada en eventos; el GUI Engine, que abstrae la interfaz de usuario en un modelo de interacción basado en mensajes; y el Rule Engine, que proporciona a los desarrolladores un medio simple para programar la lógica de la aplicación en forma de reglas \if-then". Después de que la plataforma pHPP fue utilizada durante 5 años en el proyecto HeartCycle, 5 desarrolladores fueron reunidos en un grupo de discusión para analizar y evaluar la plataforma. De estas evaluaciones se concluye que la plataforma fue diseñada para encajar las necesidades de los ingenieros que trabajan en la rama, permitiendo la separación de problemas entre las distintas especialidades, y simplificando algunas tareas de desarrollo como el manejo de datos y la interacción asíncrona. A pesar de ello, se encontraron algunos defectos a causa de la inmadurez de algunas tecnologías empleadas, y la ausencia de algunas herramientas específicas para el dominio como el procesado de datos o algunos protocolos de comunicación relacionados con la salud. Dentro del proyecto HeartCycle la plataforma fue utilizada para el desarrollo de la aplicación \Guided Exercise", un sistema TIC para la rehabilitación de pacientes que han sufrido un infarto del miocardio. El sistema fue testeado en un ensayo clínico randomizado en el cual a 55 pacientes se les dio el sistema para su uso por 21 semanas. De los resultados técnicos del ensayo se puede concluir que, a pesar de algunos errores menores prontamente corregidos durante el estudio, la plataforma es estable y fiable. Conclusiones La investigación llevada a cabo en esta Tesis y los resultados obtenidos proporcionan las respuestas a las tres preguntas de investigación que motivaron este trabajo: RQ1 Se ha desarrollado un modelo para representar el dominio de los sistemas personalizados de salud. La evaluación hecha por los expertos de la rama concluye que el modelo representa el dominio con precisión y con un balance apropiado entre abstracción y detalle. RQ2 Se ha desarrollado, con éxito, una plataforma de desarrollo basada en el modelo. RQ3 Se ha demostrado que la plataforma es capaz de ayudar a los desarrolladores en la creación de software pHealth complejos. Las ventajas de la plataforma han sido demostradas en el ámbito de un proyecto de gran escala, aunque el enfoque genérico adoptado indica que la plataforma podría ofrecer beneficios también en otros contextos. Los resultados de estas evaluaciones ofrecen indicios de que, ambos, el modelo y la plataforma serán buenos candidatos para poderse convertir en una referencia para futuros desarrollos de sistemas pHealth. ABSTRACT Background Europe is living in an unsustainable situation. The economic crisis has been reducing governments' economic resources since 2008 and threatening social and health systems, while the proportion of older people in the European population continues to increase so that it is foreseen that in 2050 there will be only two workers per retiree [54]. To this situation it should be added the rise, strongly related to age, of chronic diseases the burden of which has been estimated to be up to the 7% of a country's gross domestic product [51]. There is a need for a paradigm shift, the need for a new way of caring for people's health, shifting the focus from curing conditions that have arisen to a sustainable and effective approach with the emphasis on prevention. Some advocate the adoption of personalised health care (pHealth), a model where medical practices are tailored to the patient's unique life, from the detection of risk factors to the customization of treatments based on each individual's response [81]. Personalised health is often associated to the use of Information and Communications Technology (ICT), that, with its exponential development, offers interesting opportunities for improving healthcare. The shift towards pHealth is slowly taking place, both in research and in industry, but the change is not significant yet. Many barriers still exist related to economy, politics and culture, while others are purely technological, like the lack of interoperable information systems [199]. Though interoperability aspects are evolving, there is still the need of a reference design, especially tackling implementation and large scale deployment of pHealth systems. This thesis contributes to organizing the subject of ICT systems for personalised health into a reference model that allows for the creation of software development platforms to ease common development issues in the domain. Research questions RQ1 Is it possible to define a model, based on software engineering techniques, for representing the personalised health domain in an abstract and representative way? RQ2 Is it possible to build a development platform based on this model? RQ3 Does the development platform help developers create complex integrated pHealth systems? Methods As method for describing the model, the ISO/IEC/IEEE 42010 framework [25] is adopted for its generality and high level of abstraction. The model is specified in different parts: a conceptual model, which makes use of concept maps, for representing stakeholders, artefacts and shared information, and in scenarios and use cases for the representation of the functionalities of pHealth systems. The model was derived from literature analysis, including 7 industrial and scientific reports, 9 electronic standards, 10 conference proceedings papers, 37 journal papers, 25 websites and 5 books. Based on the reference model, requirements were drawn for building the development platform enriched with a set of requirements gathered in a survey run among 11 experienced engineers. For developing the platform, the continuous integration methodology [74] was adopted which allowed to perform automatic tests on a server and also to deploy packaged releases on a web site. As a validation methodology, a theory building framework for SW engineering was adopted from [181]. The framework, chosen as a guide to find evidence for justifying the research questions, imposed the creation of theories based on models and propositions to be validated within a scope. The validation of the model was conducted as an on-line survey in three validation rounds, encompassing a growing number of participants. The survey was submitted to 134 experts of the field and on some public channels like relevant mailing lists and social networks. Its objective was to assess the model's readability, its level of coverage of the domain and its potential usefulness in the design of actual, derived systems. The questionnaires included quantitative Likert scale questions and free text inputs for comments. The development platform was validated in two scopes. As a small-scale experiment, the platform was used in a 12 hours training session where 4 developers had to perform an exercise consisting in developing a set of typical pHealth use cases At the end of the session, a focus group was held to identify benefits and drawbacks of the platform. The second validation was held as a test-case study in a large scale research project called HeartCycle the aim of which was to develop a closed-loop disease management system for heart failure and coronary heart disease patients [160]. During this project three applications were developed by a team of programmers and designers. One of these applications was tested in a clinical trial with actual patients. At the end of the project, the team was interviewed in a focus group to assess the role the platform had within the project. Results For what regards the model that describes the pHealth domain, its conceptual part includes a description of the main roles and concerns of pHealth stakeholders, a model of the ICT artefacts that are commonly adopted and a model representing the typical data that need to be formalized among pHealth systems. The functional model includes a set of 18 scenarios, divided into assisted person's view, caregiver's view, developer's view, technology and services providers' view and authority's view, and a set of 52 Use Cases grouped in 6 categories: assisted person's activities, system reactions, caregiver's activities, user engagement, developer's activities and deployer's activities. For what concerns the validation of the model, a total of 65 people participated in the online survey providing their level of agreement in all the assessed dimensions and a total of 248 comments on how to improve and complete the model. Participants' background spanned from engineering and software development (70%) to medical specialities (15%), with declared interest in the fields of eHealth (24%), mHealth (16%), Ambient Assisted Living (21%), Personalized Medicine (5%), Personal Health Systems (15%), Medical Informatics (10%) and Biomedical Engineering (8%) with an average of 7.25_4.99 years of experience in these fields. From the analysis of the answers it is possible to observe that the contacted experts considered the model easily readable (average of 1.89_0.79 being 1 the most favourable scoring and 5 the worst), sufficiently abstract (1.99_0.88) and formal (2.13_0.77) for its purpose, with a sufficient coverage of the domain (2.26_0.95), useful for describing the domain (2.02_0.7) and for generating more specific systems (2_0.75) and they reported a partial interest in using the model in their job (2.48_0.91). Thanks to their comments, the model was improved and enriched with concepts that were missing at the beginning, nonetheless it was not possible to prove an improvement among the iterations, due to the diversity of the participants in the three rounds. From the model, a development platform for the pHealth domain was generated called pHealth Patient Platform (pHPP). The platform includes a set of libraries, programming and deployment tools, a tutorial and a sample application. The main four modules of the architecture are: the Data Collection Engine, which allows abstracting sources of information like sensors or external services, mapping data to databases and ontologies, and allowing event-based interaction and filtering, the GUI Engine, which abstracts the user interface in a message-like interaction model, the Workow Engine, which allows programming the application's user interaction ows with graphical workows, and the Rule Engine, which gives developers a simple means for programming the application's logic in the form of \if-then" rules. After the 5 years experience of HeartCycle, partially programmed with pHPP, 5 developers were joined in a focus group to discuss the advantages and drawbacks of the platform. The view that emerged from the training course and the focus group was that the platform is well-suited to the needs of the engineers working in the field, it allowed the separation of concerns among the different specialities and it simplified some common development tasks like data management and asynchronous interaction. Nevertheless, some deficiencies were pointed out in terms of a lack of maturity of some technological choices, and for the absence of some domain-specific tools, e.g. for data processing or for health-related communication protocols. Within HeartCycle, the platform was used to develop part of the Guided Exercise system, a composition of ICT tools for the physical rehabilitation of patients who suffered from myocardial infarction. The system developed using the platform was tested in a randomized controlled clinical trial, in which 55 patients used the system for 21 weeks. The technical results of this trial showed that the system was stable and reliable. Some minor bugs were detected, but these were promptly corrected using the platform. This shows that the platform, as well as facilitating the development task, can be successfully used to produce reliable software. Conclusions The research work carried out in developing this thesis provides responses to the three three research questions that were the motivation for the work. RQ1 A model was developed representing the domain of personalised health systems, and the assessment of experts in the field was that it represents the domain accurately, with an appropriate balance between abstraction and detail. RQ2 A development platform based on the model was successfully developed. RQ3 The platform has been shown to assist developers create complex pHealth software. This was demonstrated within the scope of one large-scale project, but the generic approach adopted provides indications that it would offer benefits more widely. The results of these evaluations provide indications that both the model and the platform are good candidates for being a reference for future pHealth developments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Automated and semi-automated accessibility evaluation tools are key to streamline the process of accessibility assessment, and ultimately ensure that software products, contents, and services meet accessibility requirements. Different evaluation tools may better fit different needs and concerns, accounting for a variety of corporate and external policies, content types, invocation methods, deployment contexts, exploitation models, intended audiences and goals; and the specific overall process where they are introduced. This has led to the proliferation of many evaluation tools tailored to specific contexts. However, tool creators, who may be not familiar with the realm of accessibility and may be part of a larger project, lack any systematic guidance when facing the implementation of accessibility evaluation functionalities. Herein we present a systematic approach to the development of accessibility evaluation tools, leveraging the different artifacts and activities of a standardized development process model (the Unified Software Development Process), and providing templates of these artifacts tailored to accessibility evaluation tools. The work presented specially considers the work in progress in this area by the W3C/WAI Evaluation and Report Working Group (ERT WG)

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Collaborative e-learning is increasingly appealing as a pedagogical approach that can positively affect student learning. We propose a didactical model that integrates multimedia with collaborative tools and peer assessment to foster collaborative e-learning. In this paper, we explain it and present the results of its application to the “International Seminars on Materials Science” online course. The proposed didactical model consists of five educational activities. In the first three, students review the multimedia resources proposed by the teacher in collaboration with their classmates. Then, in the last two activities, they create their own multimedia resources and assess those created by their classmates. These activities foster communication and collaboration among students and their ability to use and create multimedia resources. Our purpose is to encourage the creativity, motivation, and dynamism of the learning process for both teachers and students.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La planificación de la movilidad sostenible urbana es una tarea compleja que implica un alto grado de incertidumbre debido al horizonte de planificación a largo plazo, la amplia gama de paquetes de políticas posibles, la necesidad de una aplicación efectiva y eficiente, la gran escala geográfica, la necesidad de considerar objetivos económicos, sociales y ambientales, y la respuesta del viajero a los diferentes cursos de acción y su aceptabilidad política (Shiftan et al., 2003). Además, con las tendencias inevitables en motorización y urbanización, la demanda de terrenos y recursos de movilidad en las ciudades está aumentando dramáticamente. Como consecuencia de ello, los problemas de congestión de tráfico, deterioro ambiental, contaminación del aire, consumo de energía, desigualdades en la comunidad, etc. se hacen más y más críticos para la sociedad. Esta situación no es estable a largo plazo. Para enfrentarse a estos desafíos y conseguir un desarrollo sostenible, es necesario considerar una estrategia de planificación urbana a largo plazo, que aborde las necesarias implicaciones potencialmente importantes. Esta tesis contribuye a las herramientas de evaluación a largo plazo de la movilidad urbana estableciendo una metodología innovadora para el análisis y optimización de dos tipos de medidas de gestión de la demanda del transporte (TDM). La metodología nueva realizado se basa en la flexibilización de la toma de decisiones basadas en utilidad, integrando diversos mecanismos de decisión contrariedad‐anticipada y combinados utilidad‐contrariedad en un marco integral de planificación del transporte. La metodología propuesta incluye dos aspectos principales: 1) La construcción de escenarios con una o varias medidas TDM usando el método de encuesta que incorpora la teoría “regret”. La construcción de escenarios para este trabajo se hace para considerar específicamente la implementación de cada medida TDM en el marco temporal y marco espacial. Al final, se construyen 13 escenarios TDM en términos del más deseable, el más posible y el de menor grado de “regret” como resultado de una encuesta en dos rondas a expertos en el tema. 2) A continuación se procede al desarrollo de un marco de evaluación estratégica, basado en un Análisis Multicriterio de Toma de Decisiones (Multicriteria Decision Analysis, MCDA) y en un modelo “regret”. Este marco de evaluación se utiliza para comparar la contribución de los distintos escenarios TDM a la movilidad sostenible y para determinar el mejor escenario utilizando no sólo el valor objetivo de utilidad objetivo obtenido en el análisis orientado a utilidad MCDA, sino también el valor de “regret” que se calcula por medio del modelo “regret” MCDA. La función objetivo del MCDA se integra en un modelo de interacción de uso del suelo y transporte que se usa para optimizar y evaluar los impactos a largo plazo de los escenarios TDM previamente construidos. Un modelo de “regret”, llamado “referencedependent regret model (RDRM)” (modelo de contrariedad dependiente de referencias), se ha adaptado para analizar la contribución de cada escenario TDM desde un punto de vista subjetivo. La validación de la metodología se realiza mediante su aplicación a un caso de estudio en la provincia de Madrid. La metodología propuesta define pues un procedimiento técnico detallado para la evaluación de los impactos estratégicos de la aplicación de medidas de gestión de la demanda en el transporte, que se considera que constituye una herramienta de planificación útil, transparente y flexible, tanto para los planificadores como para los responsables de la gestión del transporte. Planning sustainable urban mobility is a complex task involving a high degree of uncertainty due to the long‐term planning horizon, the wide spectrum of potential policy packages, the need for effective and efficient implementation, the large geographical scale, the necessity to consider economic, social, and environmental goals, and the traveller’s response to the various action courses and their political acceptability (Shiftan et al., 2003). Moreover, with the inevitable trends on motorisation and urbanisation, the demand for land and mobility in cities is growing dramatically. Consequently, the problems of traffic congestion, environmental deterioration, air pollution, energy consumption, and community inequity etc., are becoming more and more critical for the society (EU, 2011). Certainly, this course is not sustainable in the long term. To address this challenge and achieve sustainable development, a long‐term perspective strategic urban plan, with its potentially important implications, should be established. This thesis contributes on assessing long‐term urban mobility by establishing an innovative methodology for optimizing and evaluating two types of transport demand management measures (TDM). The new methodology aims at relaxing the utility‐based decision‐making assumption by embedding anticipated‐regret and combined utilityregret decision mechanisms in an integrated transport planning framework. The proposed methodology includes two major aspects: 1) Construction of policy scenarios within a single measure or combined TDM policy‐packages using the survey method incorporating the regret theory. The purpose of building the TDM scenarios in this work is to address the specific implementation in terms of time frame and geographic scale for each TDM measure. Finally, 13 TDM scenarios are built in terms of the most desirable, the most expected and the least regret choice by means of the two‐round Delphi based survey. 2) Development of the combined utility‐regret analysis framework based on multicriteria decision analysis (MCDA). This assessment framework is used to compare the contribution of the TDM scenario towards sustainable mobility and to determine the best scenario considering not only the objective utility value obtained from the utilitybased MCDA, but also a regret value that is calculated via a regret‐based MCDA. The objective function of the utility‐based MCDA is integrated in a land use and transport interaction model and is used for optimizing and assessing the long term impacts of the constructed TDM scenarios. A regret based model, called referente dependent regret model (RDRM) is adapted to analyse the contribution of each TDM scenario in terms of a subjective point of view. The suggested methodology is implemented and validated in the case of Madrid. It defines a comprehensive technical procedure for assessing strategic effects of transport demand management measures, which can be useful, transparent and flexible planning tool both for planners and decision‐makers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nowadays, it is urgent to renovate a great number of residential buildings. The necessity of improving energy efficiency must also be considered as an opportunity to improve indoor comfort. To achieve this goal, it is essential to develop tools to be used in the decision-making process, aiming to refurbish buildings in an integrated, efficient and sustainable way. The integrated system developed is based on a set of indicators. Sustainability indicators are useful to synthesize and organize complex information. They can provide data to evaluate a process in different stages: evaluation, diagnosis, comparison and tracing. The set of proposed indicators aims to accomplish the holistic approach pursued by sustainable development. So, these indicators are divided into three groups: environmental, social and economic. However, the main innovation of the system of indicators is the social ones. The sustainable refurbishment system aims to be a user-focused one. Therefore, the starting point is the needs of the user and social indicators are developed around this. The system tackles the sustainable refurbishment of buildings beyond energy problems. It proposes incorporating users in the decision-making process involving them in the refurbishment and so, contributing to the success of the renovation. In order to achieve this target, three social indicators are used, divided into 10 sub-indicators, and a ?Questionnaire about Sustainable Refurbishment? is drawn up. This research has been carried out in the framework of ?Sustainable Refurbishment? Research and Development Project, an integrated project under the supervision of the Centro para el Desarrollo Tecnológico e Industrial (CDTI) from the Spanish Government, in which University and the Construction Industry collaborate. This research project aims to develop an integrated system for the retrofitting of existing buildings to improve their energy efficiency. Accordingly, an additional objective of the project is to improve quality of life of residents.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La presente tesis se enmarca dentro de los trabajos realizados en el Proyecto CENIT OASIS (Operación de Autopistas Seguras Inteligentes y Sostenibles) sobre el impacto y la integración paisajística de las autopistas, y en los trabajos realizados por el grupo de trabajo GT 13 (paisaje) dentro del comité técnico nacional CTN 157 (proyectos) para normalización del Paisaje en España, del que la doctoranda es secretaria técnica. El objetivo principal de esta tesis es desarrollar una Metodología que permita la normalización del paisaje en España. Por ello, establece las bases para el desarrollo científico y profesional en el ámbito del paisaje, a través de la caracterización de la actividad científica y de la actividad normalizadora internacional. Para después elaborar una propuesta de documentos normativos para su regulación en España. Por último, se pone en práctica la única de las normas aplicables a un caso real, concretamente en la AP-7 a su paso por la provincia de Gerona. La caracterización de la actividad científica en el ámbito del paisaje proporciona una visión global que sirve de referencia a las futuras investigaciones en la materia, no existente hasta la fecha. Entre los múltiples resultados, se identifican las áreas de conocimiento y disciplinas afines desde las que se aborda el paisaje, se analiza la evolución de las temáticas y líneas de investigación en el campo, se determina la distribución e impacto de la producción científica, destacando los países y centros de investigación punteros y sus colaboraciones, y se determinan las publicaciones más destacadas en la materia. La caracterización de la actividad normativa internacional hasta la fecha supone un referente en este campo, habiendo traducido, analizado y clasificado decenas de documentos sobre temas como la terminología, la profesión de paisajista, las reglas generales para las intervenciones en el paisaje, las normas para la protección del paisaje y normas para la evaluación del impacto paisajístico. La tesis desarrolla tres documentos normativos, que se espera sean el germen de los futuros documentos legales para normalización del Paisaje en España. El principal objetivo de la normalización es dotar a los profesionales de las herramientas necesarias para desarrollar sus intervenciones en el paisaje. Para ello, se ha elaborado un documento normativo sobre terminología del concepto clave y los términos asociados en castellano, que sirva de referencia para un futuro documento normativo; un documento normativo que regule los estudios de integración e impacto paisajístico en España, definiendo una serie de pautas que ayuden a los profesionales a desarrollar los proyectos de intervención en el paisaje; un documento que regule y defina la profesión de arquitecto paisajista, identificando sus capacidades, formaciones y competencias. Por último, el documento de impacto e integración paisajística se aplica a un caso concreto de infraestructuras del transporte, dentro del proyecto OASIS, sirviendo como ejemplo a los profesionales de la materia para desarrollar futuras intervenciones. El enfoque de este documento coincide con el de paisaje ecológico, el análisis del paisaje se aborda desde lo visible (fenosistema) y desde los procesos que lo conforman (criptosistema). Y las medidas de integración pretenden conseguir que la infraestructura forme parte del paisaje y de los procesos que ocurren en él, lo que en la tesis se define como Infraestructuras Verdes. ABSTRACT The thesis is within the framework of the CENIT OASIS Project (Operation of Safe, Intelligent and Sustainable Highways) about the landscape impact and integration of highways, and the work done by the working group GT 13 (landscape) in the national technical committee CTN 157 (projects) for landscape standardization in Spain, of which the PhD is technical secretary. The main objective of this thesis is to develop a Methodology that allows the landscape standardization in Spain. Therefore, it establishes the basis for the scientific and professional development in the landscape field, through the characterization of scientific and international normalizing activity. It concludes with the proposal of regulatory documents for its use in Spain. Finally, it implements the only of the rules applicable to a real case, specifically in the AP- 7 passing through the Gerona province. The characterization of scientific activity in the landscape field provides an overview that is a reference in the researches in this field, non-existent to date. Among the many results, the areas of knowledge and related disciplines, from which the landscape is addressed, are identified; the evolution of topics and lines of research in the field are analyzed; the distribution and impact of scientific production is determined, highlighting the countries and leading research centers and collaborations; and the leading publications in the field are determined. The characterization of the international regulatory activity to date is a model in this field, having translated, analyzed and classified dozens of papers about terminology, the landscapist profession, general rules for intervention in the landscape, standards for the landscape protection and rules for the assessment of landscape impact. The thesis develops three normative documents, which are expected to be the germ of future legal documents for standardization landscape in Spain. The main objective of standardization is to provide the necessary tools for professionals who work developing interventions in the landscape. To do this, it has been developed a normative document about terminology on the key concept and the associated terms in Castilian, as a reference for a future normative document; a normative document that regulates studies of landscape integration and impact in Spain, defining a set of guidelines to help professionals to develop intervention projects in the landscape; a document to regulate and define the activities developed by the professionals, defining the profession of landscape architect, their capabilities and competencies. Finally, the document of landscape impact and integration is applied to a particular case of transport infrastructures within the OASIS project, serving as an example to professionals in the field to develop future interventions. The focus of this document coincides with the ecological landscape; the landscape analysis is approached from the visible (fenosystem) and from the processes that shape it (cryptosystem); and integration measures aim to achieve that the infrastructure take part of the landscape and its existing processes, which in this thesis is defined as Green Infrastructures.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El estudio de los ciclos del combustible nuclear requieren de herramientas computacionales o "códigos" versátiles para dar respuestas al problema multicriterio de evaluar los actuales ciclos o las capacidades de las diferentes estrategias y escenarios con potencial de desarrollo en a nivel nacional, regional o mundial. Por otra parte, la introducción de nuevas tecnologías para reactores y procesos industriales hace que los códigos existentes requieran nuevas capacidades para evaluar la transición del estado actual del ciclo del combustible hacia otros más avanzados y sostenibles. Brevemente, esta tesis se centra en dar respuesta a las principales preguntas, en términos económicos y de recursos, al análisis de escenarios de ciclos de combustible, en particular, para el análisis de los diferentes escenarios del ciclo del combustible de relativa importancia para España y Europa. Para alcanzar este objetivo ha sido necesaria la actualización y el desarrollo de nuevas capacidades del código TR_EVOL (Transition Evolution code). Este trabajo ha sido desarrollado en el Programa de Innovación Nuclear del CIEMAT desde el año 2010. Esta tesis se divide en 6 capítulos. El primer capítulo ofrece una visión general del ciclo de combustible nuclear, sus principales etapas y los diferentes tipos utilizados en la actualidad o en desarrollo para el futuro. Además, se describen las fuentes de material nuclear que podrían ser utilizadas como combustible (uranio y otros). También se puntualizan brevemente una serie de herramientas desarrolladas para el estudio de estos ciclos de combustible nuclear. El capítulo 2 está dirigido a dar una idea básica acerca de los costes involucrados en la generación de electricidad mediante energía nuclear. Aquí se presentan una clasificación de estos costos y sus estimaciones, obtenidas en la bibliografía, y que han sido evaluadas y utilizadas en esta tesis. Se ha incluido también una breve descripción del principal indicador económico utilizado en esta tesis, el “coste nivelado de la electricidad”. El capítulo 3 se centra en la descripción del código de simulación desarrollado para el estudio del ciclo del combustible nuclear, TR_EVOL, que ha sido diseñado para evaluar diferentes opciones de ciclos de combustibles. En particular, pueden ser evaluados las diversos reactores con, posiblemente, diferentes tipos de combustibles y sus instalaciones del ciclo asociadas. El módulo de evaluaciones económica de TR_EVOL ofrece el coste nivelado de la electricidad haciendo uso de las cuatro fuentes principales de información económica y de la salida del balance de masas obtenido de la simulación del ciclo en TR_EVOL. Por otra parte, la estimación de las incertidumbres en los costes también puede ser efectuada por el código. Se ha efectuado un proceso de comprobación cruzada de las funcionalidades del código y se descrine en el Capítulo 4. El proceso se ha aplicado en cuatro etapas de acuerdo con las características más importantes de TR_EVOL, balance de masas, composición isotópica y análisis económico. Así, la primera etapa ha consistido en el balance masas del ciclo de combustible nuclear actual de España. La segunda etapa se ha centrado en la comprobación de la composición isotópica del flujo de masas mediante el la simulación del ciclo del combustible definido en el proyecto CP-ESFR UE. Las dos últimas etapas han tenido como objetivo validar el módulo económico. De este modo, en la tercera etapa han sido evaluados los tres principales costes (financieros, operación y mantenimiento y de combustible) y comparados con los obtenidos por el proyecto ARCAS, omitiendo los costes del fin del ciclo o Back-end, los que han sido evaluado solo en la cuarta etapa, haciendo uso de costes unitarios y parámetros obtenidos a partir de la bibliografía. En el capítulo 5 se analizan dos grupos de opciones del ciclo del combustible nuclear de relevante importancia, en términos económicos y de recursos, para España y Europa. Para el caso español, se han simulado dos grupos de escenarios del ciclo del combustible, incluyendo estrategias de reproceso y extensión de vida de los reactores. Este análisis se ha centrado en explorar las ventajas y desventajas de reprocesado de combustible irradiado en un país con una “relativa” pequeña cantidad de reactores nucleares. Para el grupo de Europa se han tratado cuatro escenarios, incluyendo opciones de transmutación. Los escenarios incluyen los reactores actuales utilizando la tecnología reactor de agua ligera y ciclo abierto, un reemplazo total de los reactores actuales con reactores rápidos que queman combustible U-Pu MOX y dos escenarios del ciclo del combustible con transmutación de actínidos minoritarios en una parte de los reactores rápidos o en sistemas impulsados por aceleradores dedicados a transmutación. Finalmente, el capítulo 6 da las principales conclusiones obtenidas de esta tesis y los trabajos futuros previstos en el campo del análisis de ciclos de combustible nuclear. ABSTRACT The study of the nuclear fuel cycle requires versatile computational tools or “codes” to provide answers to the multicriteria problem of assessing current nuclear fuel cycles or the capabilities of different strategies and scenarios with potential development in a country, region or at world level. Moreover, the introduction of new technologies for reactors and industrial processes makes the existing codes to require new capabilities to assess the transition from current status of the fuel cycle to the more advanced and sustainable ones. Briefly, this thesis is focused in providing answers to the main questions about resources and economics in fuel cycle scenario analyses, in particular for the analysis of different fuel cycle scenarios with relative importance for Spain and Europe. The upgrade and development of new capabilities of the TR_EVOL code (Transition Evolution code) has been necessary to achieve this goal. This work has been developed in the Nuclear Innovation Program at CIEMAT since year 2010. This thesis is divided in 6 chapters. The first one gives an overview of the nuclear fuel cycle, its main stages and types currently used or in development for the future. Besides the sources of nuclear material that could be used as fuel (uranium and others) are also viewed here. A number of tools developed for the study of these nuclear fuel cycles are also briefly described in this chapter. Chapter 2 is aimed to give a basic idea about the cost involved in the electricity generation by means of the nuclear energy. The main classification of these costs and their estimations given by bibliography, which have been evaluated in this thesis, are presented. A brief description of the Levelized Cost of Electricity, the principal economic indicator used in this thesis, has been also included. Chapter 3 is focused on the description of the simulation tool TR_EVOL developed for the study of the nuclear fuel cycle. TR_EVOL has been designed to evaluate different options for the fuel cycle scenario. In particular, diverse nuclear power plants, having possibly different types of fuels and the associated fuel cycle facilities can be assessed. The TR_EVOL module for economic assessments provides the Levelized Cost of Electricity making use of the TR_EVOL mass balance output and four main sources of economic information. Furthermore, uncertainties assessment can be also carried out by the code. A cross checking process of the performance of the code has been accomplished and it is shown in Chapter 4. The process has been applied in four stages according to the most important features of TR_EVOL. Thus, the first stage has involved the mass balance of the current Spanish nuclear fuel cycle. The second stage has been focused in the isotopic composition of the mass flow using the fuel cycle defined in the EU project CP-ESFR. The last two stages have been aimed to validate the economic module. In the third stage, the main three generation costs (financial cost, O&M and fuel cost) have been assessed and compared to those obtained by ARCAS project, omitting the back-end costs. This last cost has been evaluated alone in the fourth stage, making use of some unit cost and parameters obtained from the bibliography. In Chapter 5 two groups of nuclear fuel cycle options with relevant importance for Spain and Europe are analyzed in economic and resources terms. For the Spanish case, two groups of fuel cycle scenarios have been simulated including reprocessing strategies and life extension of the current reactor fleet. This analysis has been focused on exploring the advantages and disadvantages of spent fuel reprocessing in a country with relatively small amount of nuclear power plants. For the European group, four fuel cycle scenarios involving transmutation options have been addressed. Scenarios include the current fleet using Light Water Reactor technology and open fuel cycle, a full replacement of the initial fleet with Fast Reactors burning U-Pu MOX fuel and two fuel cycle scenarios with Minor Actinide transmutation in a fraction of the FR fleet or in dedicated Accelerator Driven Systems. Finally, Chapter 6 gives the main conclusions obtained from this thesis and the future work foreseen in the field of nuclear fuel cycle analysis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The new reactor concepts proposed in the Generation IV International Forum (GIF) are conceived to improve the use of natural resources, reduce the amount of high-level radioactive waste and excel in their reliability and safe operation. Among these novel designs sodium fast reactors (SFRs) stand out due to their technological feasibility as demonstrated in several countries during the last decades. As part of the contribution of EURATOM to GIF the CP-ESFR is a collaborative project with the objective, among others, to perform extensive analysis on safety issues involving renewed SFR demonstrator designs. The verification of computational tools able to simulate the plant behaviour under postulated accidental conditions by code-to-code comparison was identified as a key point to ensure reactor safety. In this line, several organizations employed coupled neutronic and thermal-hydraulic system codes able to simulate complex and specific phenomena involving multi-physics studies adapted to this particular fast reactor technology. In the “Introduction” of this paper the framework of this study is discussed, the second section describes the envisaged plant design and the commonly agreed upon modelling guidelines. The third section presents a comparative analysis of the calculations performed by each organisation applying their models and codes to a common agreed transient with the objective to harmonize the models as well as validating the implementation of all relevant physical phenomena in the different system codes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The new reactor concepts proposed in the Generation IV International Forum require the development and validation of computational tools able to assess their safety performance. In the first part of this paper the models of the ESFR design developed by several organisations in the framework of the CP-ESFR project were presented and their reliability validated via a benchmarking exercise. This second part of the paper includes the application of those tools for the analysis of design basis accident (DBC) scenarios of the reference design. Further, this paper also introduces the main features of the core optimisation process carried out within the project with the objective to enhance the core safety performance through the reduction of the positive coolant density reactivity effect. The influence of this optimised core design on the reactor safety performance during the previously analysed transients is also discussed. The conclusion provides an overview of the work performed by the partners involved in the project towards the development and enhancement of computational tools specifically tailored to the evaluation of the safety performance of the Generation IV innovative nuclear reactor designs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work analysed the feasibility of using a fast, customized Monte Carlo (MC) method to perform accurate computation of dose distributions during pre- and intraplanning of intraoperative electron radiation therapy (IOERT) procedures. The MC method that was implemented, which has been integrated into a specific innovative simulation and planning tool, is able to simulate the fate of thousands of particles per second, and it was the aim of this work to determine the level of interactivity that could be achieved. The planning workflow enabled calibration of the imaging and treatment equipment, as well as manipulation of the surgical frame and insertion of the protection shields around the organs at risk and other beam modifiers. In this way, the multidisciplinary team involved in IOERT has all the tools necessary to perform complex MC dosage simulations adapted to their equipment in an efficient and transparent way. To assess the accuracy and reliability of this MC technique, dose distributions for a monoenergetic source were compared with those obtained using a general-purpose software package used widely in medical physics applications. Once accuracy of the underlying simulator was confirmed, a clinical accelerator was modelled and experimental measurements in water were conducted. A comparison was made with the output from the simulator to identify the conditions under which accurate dose estimations could be obtained in less than 3 min, which is the threshold imposed to allow for interactive use of the tool in treatment planning. Finally, a clinically relevant scenario, namely early-stage breast cancer treatment, was simulated with pre- and intraoperative volumes to verify that it was feasible to use the MC tool intraoperatively and to adjust dose delivery based on the simulation output, without compromising accuracy. The workflow provided a satisfactory model of the treatment head and the imaging system, enabling proper configuration of the treatment planning system and providing good accuracy in the dosage simulation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Engineering aims to work with what knowledge is available to achieve society's goals (Coyle, Murphy, and Grimson 2007). The current environmental challenges and the characteristics of the labour market mean that the effectiveness of Engineering activities in Geosciences must be increased through the development of technical knowledge and the inclusion of suitable training aimed at solving real cases (European Commission 2010). Human capital – understood as the talents, skills and capabilities of higher education graduates – is perceived as an essential element for sustainable economic growth and development in the globalised economy (Sianesi and Van Reenan 2003). We need, therefore, to rethink our approaches to curriculum, instruction and assessment in science education, particularly because of the rapid growth of the scientific knowledge, tools/technologies and theories that have originated over the last 50 years (Duschl and Grandy 2013).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Automating the assessment of programming assignments brings benefits for both students and teachers, since it helps the formers to gain a timely feedback and releases the latter from tedious tasks. The related literature in the domain has usually focused on the assessment process and the tools required for it, proposing libraries and systems that teachers can use in this process. However, few of them have work rowards reducing the effort and time teacher require to properly set up new assessente processes. This paper describes our experience with the analysis and design of a new tool to support teachers in visually developing automatic grades of programming assignments, introducing the underlying concepts and technologies and presenting the system architecture.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the ever growing trend of smart phones and tablets, Android is becoming more and more popular everyday. With more than one billion active users i to date, Android is the leading technology in smart phone arena. In addition to that, Android also runs on Android TV, Android smart watches and cars. Therefore, in recent years, Android applications have become one of the major development sectors in software industry. As of mid 2013, the number of published applications on Google Play had exceeded one million and the cumulative number of downloads was more than 50 billionii. A 2013 survey also revealed that 71% of the mobile application developers work on developing Android applicationsiii. Considering this size of Android applications, it is quite evident that people rely on these applications on a daily basis for the completion of simple tasks like keeping track of weather to rather complex tasks like managing one’s bank accounts. Hence, like every other kind of code, Android code also needs to be verified in order to work properly and achieve a certain confidence level. Because of the gigantic size of the number of applications, it becomes really hard to manually test Android applications specially when it has to be verified for various versions of the OS and also, various device configurations such as different screen sizes and different hardware availability. Hence, recently there has been a lot of work on developing different testing methods for Android applications in Computer Science fraternity. The model of Android attracts researchers because of its open source nature. It makes the whole research model more streamlined when the code for both, application and the platform are readily available to analyze. And hence, there has been a great deal of research in testing and static analysis of Android applications. A great deal of this research has been focused on the input test generation for Android applications. Hence, there are a several testing tools available now, which focus on automatic generation of test cases for Android applications. These tools differ with one another on the basis of their strategies and heuristics used for this generation of test cases. But there is still very little work done on the comparison of these testing tools and the strategies they use. Recently, some research work has been carried outiv in this regard that compared the performance of various available tools with respect to their respective code coverage, fault detection, ability to work on multiple platforms and their ease of use. It was done, by running these tools on a total of 60 real world Android applications. The results of this research showed that although effective, these strategies being used by the tools, also face limitations and hence, have room for improvement. The purpose of this thesis is to extend this research into a more specific and attribute-­‐ oriented way. Attributes refer to the tasks that can be completed using the Android platform. It can be anything ranging from a basic system call for receiving an SMS to more complex tasks like sending the user to another application from the current one. The idea is to develop a benchmark for Android testing tools, which is based on the performance related to these attributes. This will allow the comparison of these tools with respect to these attributes. For example, if there is an application that plays some audio file, will the testing tool be able to generate a test input that will warrant the execution of this audio file? Using multiple applications using different attributes, it can be visualized that which testing tool is more useful for which kinds of attributes. In this thesis, it was decided that 9 attributes covering the basic nature of tasks, will be targeted for the assessment of three testing tools. Later this can be done for much more attributes to compare even more testing tools. The aim of this work is to show that this approach is effective and can be used on a much larger scale. One of the flagship features of this work, which also differentiates it with the previous work, is that the applications used, are all specially made for this research. The reason for doing that is to analyze just that specific attribute in isolation, which the application is focused on, and not allow the tool to get bottlenecked by something trivial, which is not the main attribute under testing. This means 9 applications, each focused on one specific attribute. The main contributions of this thesis are: A summary of the three existing testing tools and their respective techniques for automatic test input generation of Android Applications. • A detailed study of the usage of these testing tools using the 9 applications specially designed and developed for this study. • The analysis of the obtained results of the study carried out. And a comparison of the performance of the selected tools.