991 resultados para Flood Proof Design


Relevância:

30.00% 30.00%

Publicador:

Resumo:

As a consequence of flood impacts, communities inhabiting mountain areas are increasingly affected by considerable damage to infrastructure and property. The design of effective flood risk mitigation strategies and their subsequent implementation is crucial for a sustainable development in mountain areas. The assessment of the dynamic evolution of flood risk is the pillar of any subsequent planning process that is targeted at a reduction of the expected adverse consequences of the hazard impact. Given these premises, firstly, a comprehensive method to derive flood hazard process scenarios for well-defined areas at risk is presented. Secondly, conceptualisations of a static and dynamic flood risk assessment are provided. These are based on formal schemes to compute the risk mitigation performance of devised mitigation strategies within the framework of economic cost-benefit analysis. In this context, techniques suitable to quantify the expected losses induced by the identified flood impacts are provided.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Standard procedures for forecasting flood risk (Bulletin 17B) assume annual maximum flood (AMF) series are stationary, meaning the distribution of flood flows is not significantly affected by climatic trends/cycles, or anthropogenic activities within the watershed. Historical flood events are therefore considered representative of future flood occurrences, and the risk associated with a given flood magnitude is modeled as constant over time. However, in light of increasing evidence to the contrary, this assumption should be reconsidered, especially as the existence of nonstationarity in AMF series can have significant impacts on planning and management of water resources and relevant infrastructure. Research presented in this thesis quantifies the degree of nonstationarity evident in AMF series for unimpaired watersheds throughout the contiguous U.S., identifies meteorological, climatic, and anthropogenic causes of this nonstationarity, and proposes an extension of the Bulletin 17B methodology which yields forecasts of flood risk that reflect climatic influences on flood magnitude. To appropriately forecast flood risk, it is necessary to consider the driving causes of nonstationarity in AMF series. Herein, large-scale climate patterns—including El Niño-Southern Oscillation (ENSO), Pacific Decadal Oscillation (PDO), North Atlantic Oscillation (NAO), and Atlantic Multidecadal Oscillation (AMO)—are identified as influencing factors on flood magnitude at numerous stations across the U.S. Strong relationships between flood magnitude and associated precipitation series were also observed for the majority of sites analyzed in the Upper Midwest and Northeastern regions of the U.S. Although relationships between flood magnitude and associated temperature series are not apparent, results do indicate that temperature is highly correlated with the timing of flood peaks. Despite consideration of watersheds classified as unimpaired, analyses also suggest that identified change-points in AMF series are due to dam construction, and other types of regulation and diversion. Although not explored herein, trends in AMF series are also likely to be partially explained by changes in land use and land cover over time. Results obtained herein suggest that improved forecasts of flood risk may be obtained using a simple modification of the Bulletin 17B framework, wherein the mean and standard deviation of the log-transformed flows are modeled as functions of climate indices associated with oceanic-atmospheric patterns (e.g. AMO, ENSO, NAO, and PDO) with lead times between 3 and 9 months. Herein, one-year ahead forecasts of the mean and standard deviation, and subsequently flood risk, are obtained by applying site specific multivariate regression models, which reflect the phase and intensity of a given climate pattern, as well as possible impacts of coupling of the climate cycles. These forecasts of flood risk are compared with forecasts derived using the existing Bulletin 17B model; large differences in the one-year ahead forecasts are observed in some locations. The increased knowledge of the inherent structure of AMF series and an improved understanding of physical and/or climatic causes of nonstationarity gained from this research should serve as insight for the formulation of a physical-casual based statistical model, incorporating both climatic variations and human impacts, for flood risk over longer planning horizons (e.g., 10-, 50, 100-years) necessary for water resources design, planning, and management.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Self-stabilization is a property of a distributed system such that, regardless of the legitimacy of its current state, the system behavior shall eventually reach a legitimate state and shall remain legitimate thereafter. The elegance of self-stabilization stems from the fact that it distinguishes distributed systems by a strong fault tolerance property against arbitrary state perturbations. The difficulty of designing and reasoning about self-stabilization has been witnessed by many researchers; most of the existing techniques for the verification and design of self-stabilization are either brute-force, or adopt manual approaches non-amenable to automation. In this dissertation, we first investigate the possibility of automatically designing self-stabilization through global state space exploration. In particular, we develop a set of heuristics for automating the addition of recovery actions to distributed protocols on various network topologies. Our heuristics equally exploit the computational power of a single workstation and the available parallelism on computer clusters. We obtain existing and new stabilizing solutions for classical protocols like maximal matching, ring coloring, mutual exclusion, leader election and agreement. Second, we consider a foundation for local reasoning about self-stabilization; i.e., study the global behavior of the distributed system by exploring the state space of just one of its components. It turns out that local reasoning about deadlocks and livelocks is possible for an interesting class of protocols whose proof of stabilization is otherwise complex. In particular, we provide necessary and sufficient conditions – verifiable in the local state space of every process – for global deadlock- and livelock-freedom of protocols on ring topologies. Local reasoning potentially circumvents two fundamental problems that complicate the automated design and verification of distributed protocols: (1) state explosion and (2) partial state information. Moreover, local proofs of convergence are independent of the number of processes in the network, thereby enabling our assertions about deadlocks and livelocks to apply on rings of arbitrary sizes without worrying about state explosion.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

ABSTRACT ONTOLOGIES AND METHODS FOR INTEROPERABILITY OF ENGINEERING ANALYSIS MODELS (EAMS) IN AN E-DESIGN ENVIRONMENT SEPTEMBER 2007 NEELIMA KANURI, B.S., BIRLA INSTITUTE OF TECHNOLOGY AND SCIENCES PILANI INDIA M.S., UNIVERSITY OF MASSACHUSETTS AMHERST Directed by: Professor Ian Grosse Interoperability is the ability of two or more systems to exchange and reuse information efficiently. This thesis presents new techniques for interoperating engineering tools using ontologies as the basis for representing, visualizing, reasoning about, and securely exchanging abstract engineering knowledge between software systems. The specific engineering domain that is the primary focus of this report is the modeling knowledge associated with the development of engineering analysis models (EAMs). This abstract modeling knowledge has been used to support integration of analysis and optimization tools in iSIGHT FD , a commercial engineering environment. ANSYS , a commercial FEA tool, has been wrapped as an analysis service available inside of iSIGHT-FD. Engineering analysis modeling (EAM) ontology has been developed and instantiated to form a knowledge base for representing analysis modeling knowledge. The instances of the knowledge base are the analysis models of real world applications. To illustrate how abstract modeling knowledge can be exploited for useful purposes, a cantilever I-Beam design optimization problem has been used as a test bed proof-of-concept application. Two distinct finite element models of the I-beam are available to analyze a given beam design- a beam-element finite element model with potentially lower accuracy but significantly reduced computational costs and a high fidelity, high cost, shell-element finite element model. The goal is to obtain an optimized I-beam design at minimum computational expense. An intelligent KB tool was developed and implemented in FiPER . This tool reasons about the modeling knowledge to intelligently shift between the beam and the shell element models during an optimization process to select the best analysis model for a given optimization design state. In addition to improved interoperability and design optimization, methods are developed and presented that demonstrate the ability to operate on ontological knowledge bases to perform important engineering tasks. One such method is the automatic technical report generation method which converts the modeling knowledge associated with an analysis model to a flat technical report. The second method is a secure knowledge sharing method which allocates permissions to portions of knowledge to control knowledge access and sharing. Both the methods acting together enable recipient specific fine grain controlled knowledge viewing and sharing in an engineering workflow integration environment, such as iSIGHT-FD. These methods together play a very efficient role in reducing the large scale inefficiencies existing in current product design and development cycles due to poor knowledge sharing and reuse between people and software engineering tools. This work is a significant advance in both understanding and application of integration of knowledge in a distributed engineering design framework.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE Successful repair of defects in the avascular zone of meniscus remains a challenge in orthopedics. This proof of concept study aimed to investigate a guided tissue regeneration approach for treatment of tears in meniscus avascular zone in a goat model. DESIGN Full-depth longitudinal tear was created in the avascular zone of the meniscus and sutured. In the two treatment groups, porcine collagen membrane was wrapped around the tear without (CM) or with injection of expanded autologous chondrocytes (CM+cells), whereas in the control group the tear remained only sutured. Gait recovery was evaluated during the entire follow-up period. On explantation at 3 and 6 months, macroscopic gross inspection assessed healing of tears, degradation of collagen membrane, potential signs of inflammation, and osteoarthritic changes. Microscopic histology scoring criteria were developed to evaluate healing of tears, the cellular response, and the inflammatory response. RESULTS Gait recovery suggested protective effect of collagen membrane and was supported by macroscopical evaluation where improved tear healing was noted in both treated groups. Histology scoring in CM compared to suture group revealed an increase in tear margins contact, newly formed connective tissue between margins, and cell formations surrounded with new matrix after 3 months yet not maintained after 6 months. In contrast, in the CM+cells group these features were observed after 3 and 6 months. CONCLUSIONS A transient, short-term guided tissue regeneration of avascular meniscal tears occurred upon application of collagen membrane, whereas addition of expanded autologous chondrocytes supported more sustainable longer term tear healing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The development of a novel optical design for the high concentration photovoltaics (HPCV) nonimaging concentrator (>500x) that utilizes a built-in spectrum splitting concept is presented. The primary optical element (POE) is a flat Fresnel lens and the secondary optical element (SOE) is a free-form RXI-type concentrator with a band-pass filter embedded in it. The POE and SOE perform Köhler integration to produce light homogenization on the receiver. The system uses a combination of a commercial concentration GaInP/GaInAs/Ge 3J cell and a concentration Back-PointContact (BPC) silicon cell for efficient spectral utilization, and an external confinement technique for recovering the 3J cell’s reflection. A design target of an “equivalent” cell efficiency ~46% is predicted using commercial 39% 3J and 26% Si cells. A projected CPV module efficiency of greater than 38% is achievable at a concentration level greater than 500X with a wide acceptance angle of ±1º. A first proof-of concept receiver prototype has been manufactured using a simpler optical architecture (with a lower concentration, ~100x and lower simulated added efficiency), and experimental measurements have shown up to 39.8% 4J receiver efficiency using a 3J cell with a peak efficiency of 36.9%

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Development of a novel HCPV nonimaging concentrator with high concentration (>500x) and built-in spectrum splitting concept is presented. It uses the combination of a commercial concentration GaInP/GaInAs/Ge 3J cell and a concentration Back-Point-Contact (BPC) silicon cell for efficient spectral utilization, and external confinement techniques for recovering the 3J cell's reflection. The primary optical element (POE) is a flat Fresnel lens and the secondary optical element (SOE) is a free-form RXI-type concentrator with a band-pass filter embedded in it - Both the POE and SOE performing Köhler integration to produce light homogenization on the receiver. The band-pass filter transmits the IR photons in the 900-1200 nm band to the silicon cell. A design target of an "equivalent" cell efficiency ~46% is predicted using commercial 39% 3J and 26% Si cells. A projected CPV module efficiency of greater than 38% is achievable at a concentration level larger than 500X with a wide acceptance angle of ±1°. A first proof-of concept receiver prototype has been manufactured using a simpler optical architecture (with a lower concentration, ~100x and lower simulated added efficiency), and experimental measurements have shown up to 39.8% 4J receiver efficiency using a 3J cell with a peak efficiency of 36.9%.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Las técnicas de cirugía de mínima invasión (CMI) se están consolidando hoy en día como alternativa a la cirugía tradicional, debido a sus numerosos beneficios para los pacientes. Este cambio de paradigma implica que los cirujanos deben aprender una serie de habilidades distintas de aquellas requeridas en cirugía abierta. El entrenamiento y evaluación de estas habilidades se ha convertido en una de las mayores preocupaciones en los programas de formación de cirujanos, debido en gran parte a la presión de una sociedad que exige cirujanos bien preparados y una reducción en el número de errores médicos. Por tanto, se está prestando especial atención a la definición de nuevos programas que permitan el entrenamiento y la evaluación de las habilidades psicomotoras en entornos seguros antes de que los nuevos cirujanos puedan operar sobre pacientes reales. Para tal fin, hospitales y centros de formación están gradualmente incorporando instalaciones de entrenamiento donde los residentes puedan practicar y aprender sin riesgos. Es cada vez más común que estos laboratorios dispongan de simuladores virtuales o simuladores físicos capaces de registrar los movimientos del instrumental de cada residente. Estos simuladores ofrecen una gran variedad de tareas de entrenamiento y evaluación, así como la posibilidad de obtener información objetiva de los ejercicios. Los diferentes estudios de validación llevados a cabo dan muestra de su utilidad; pese a todo, los niveles de evidencia presentados son en muchas ocasiones insuficientes. Lo que es más importante, no existe un consenso claro a la hora de definir qué métricas son más útiles para caracterizar la pericia quirúrgica. El objetivo de esta tesis doctoral es diseñar y validar un marco de trabajo conceptual para la definición y validación de entornos para la evaluación de habilidades en CMI, en base a un modelo en tres fases: pedagógica (tareas y métricas a emplear), tecnológica (tecnologías de adquisición de métricas) y analítica (interpretación de la competencia en base a las métricas). Para tal fin, se describe la implementación práctica de un entorno basado en (1) un sistema de seguimiento de instrumental fundamentado en el análisis del vídeo laparoscópico; y (2) la determinación de la pericia en base a métricas de movimiento del instrumental. Para la fase pedagógica se diseñó e implementó un conjunto de tareas para la evaluación de habilidades psicomotoras básicas, así como una serie de métricas de movimiento. La validación de construcción llevada a cabo sobre ellas mostró buenos resultados para tiempo, camino recorrido, profundidad, velocidad media, aceleración media, economía de área y economía de volumen. Adicionalmente, los resultados obtenidos en la validación de apariencia fueron en general positivos en todos los grupos considerados (noveles, residentes, expertos). Para la fase tecnológica, se introdujo el EVA Tracking System, una solución para el seguimiento del instrumental quirúrgico basado en el análisis del vídeo endoscópico. La precisión del sistema se evaluó a 16,33ppRMS para el seguimiento 2D de la herramienta en la imagen; y a 13mmRMS para el seguimiento espacial de la misma. La validación de construcción con una de las tareas de evaluación mostró buenos resultados para tiempo, camino recorrido, profundidad, velocidad media, aceleración media, economía de área y economía de volumen. La validación concurrente con el TrEndo® Tracking System por su parte presentó valores altos de correlación para 8 de las 9 métricas analizadas. Finalmente, para la fase analítica se comparó el comportamiento de tres clasificadores supervisados a la hora de determinar automáticamente la pericia quirúrgica en base a la información de movimiento del instrumental, basados en aproximaciones lineales (análisis lineal discriminante, LDA), no lineales (máquinas de soporte vectorial, SVM) y difusas (sistemas adaptativos de inferencia neurodifusa, ANFIS). Los resultados muestran que en media SVM presenta un comportamiento ligeramente superior: 78,2% frente a los 71% y 71,7% obtenidos por ANFIS y LDA respectivamente. Sin embargo las diferencias estadísticas medidas entre los tres no fueron demostradas significativas. En general, esta tesis doctoral corrobora las hipótesis de investigación postuladas relativas a la definición de sistemas de evaluación de habilidades para cirugía de mínima invasión, a la utilidad del análisis de vídeo como fuente de información y a la importancia de la información de movimiento de instrumental a la hora de caracterizar la pericia quirúrgica. Basándose en estos cimientos, se han de abrir nuevos campos de investigación que contribuyan a la definición de programas de formación estructurados y objetivos, que puedan garantizar la acreditación de cirujanos sobradamente preparados y promocionen la seguridad del paciente en el quirófano. Abstract Minimally invasive surgery (MIS) techniques have become a standard in many surgical sub-specialties, due to their many benefits for patients. However, this shift in paradigm implies that surgeons must acquire a complete different set of skills than those normally attributed to open surgery. Training and assessment of these skills has become a major concern in surgical learning programmes, especially considering the social demand for better-prepared professionals and for the decrease of medical errors. Therefore, much effort is being put in the definition of structured MIS learning programmes, where practice with real patients in the operating room (OR) can be delayed until the resident can attest for a minimum level of psychomotor competence. To this end, skills’ laboratory settings are being introduced in hospitals and training centres where residents may practice and be assessed on their psychomotor skills. Technological advances in the field of tracking technologies and virtual reality (VR) have enabled the creation of new learning systems such as VR simulators or enhanced box trainers. These systems offer a wide range of tasks, as well as the capability of registering objective data on the trainees’ performance. Validation studies give proof of their usefulness; however, levels of evidence reported are in many cases low. More importantly, there is still no clear consensus on topics such as the optimal metrics that must be used to assess competence, the validity of VR simulation, the portability of tracking technologies into real surgeries (for advanced assessment) or the degree to which the skills measured and obtained in laboratory environments transfer to the OR. The purpose of this PhD is to design and validate a conceptual framework for the definition and validation of MIS assessment environments based on a three-pillared model defining three main stages: pedagogical (tasks and metrics to employ), technological (metric acquisition technologies) and analytical (interpretation of competence based on metrics). To this end, a practical implementation of the framework is presented, focused on (1) a video-based tracking system and (2) the determination of surgical competence based on the laparoscopic instruments’ motionrelated data. The pedagogical stage’s results led to the design and implementation of a set of basic tasks for MIS psychomotor skills’ assessment, as well as the definition of motion analysis parameters (MAPs) to measure performance on said tasks. Validation yielded good construct results for parameters such as time, path length, depth, average speed, average acceleration, economy of area and economy of volume. Additionally, face validation results showed positive acceptance on behalf of the experts, residents and novices. For the technological stage the EVA Tracking System is introduced. EVA provides a solution for tracking laparoscopic instruments from the analysis of the monoscopic video image. Accuracy tests for the system are presented, which yielded an average RMSE of 16.33pp for 2D tracking of the instrument on the image and of 13mm for 3D spatial tracking. A validation experiment was conducted using one of the tasks and the most relevant MAPs. Construct validation showed significant differences for time, path length, depth, average speed, average acceleration, economy of area and economy of volume; especially between novices and residents/experts. More importantly, concurrent validation with the TrEndo® Tracking System presented high correlation values (>0.7) for 8 of the 9 MAPs proposed. Finally, the analytical stage allowed comparing the performance of three different supervised classification strategies in the determination of surgical competence based on motion-related information. The three classifiers were based on linear (linear discriminant analysis, LDA), non-linear (support vector machines, SVM) and fuzzy (adaptive neuro fuzzy inference systems, ANFIS) approaches. Results for SVM show slightly better performance than the other two classifiers: on average, accuracy for LDA, SVM and ANFIS was of 71.7%, 78.2% and 71% respectively. However, when confronted, no statistical significance was found between any of the three. Overall, this PhD corroborates the investigated research hypotheses regarding the definition of MIS assessment systems, the use of endoscopic video analysis as the main source of information and the relevance of motion analysis in the determination of surgical competence. New research fields in the training and assessment of MIS surgeons can be proposed based on these foundations, in order to contribute to the definition of structured and objective learning programmes that guarantee the accreditation of well-prepared professionals and the promotion of patient safety in the OR.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study characterises the abatement effect of large dams with fixed-crest spillways under extreme design flood conditions. In contrast to previous studies using specific hydrographs for flow into the reservoir and simplifications to obtain analytical solutions, an automated tool was designed for calculations based on a Monte Carlo simulation environment, which integrates models that represent the different physical processes in watersheds with areas of 150?2000 km2. The tool was applied to 21 sites that were uniformly distributed throughout continental Spain, with 105 fixed-crest dam configurations. This tool allowed a set of hydrographs to be obtained as an approximation for the hydrological forcing of a dam and the characterisation of the response of the dam to this forcing. For all cases studied, we obtained a strong linear correlation between the peak flow entering the reservoir and the peak flow discharged by the dam, and a simple general procedure was proposed to characterise the peak-flow attenuation behaviour of the reservoir. Additionally, two dimensionless coefficients were defined to relate the variables governing both the generation of the flood and its abatement in the reservoir. Using these coefficients, a model was defined to allow for the estimation of the flood abatement effect of a reservoir based on the available information. This model should be useful in the hydrological design of spillways and the evaluation of the hydrological safety of dams. Finally, the proposed procedure and model were evaluated and representative applications were presented

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A multivariate analysis on flood variables is needed to design some hydraulic structures like dams, as the complexity of the routing process in a reservoir requires a representation of the full hydrograph. In this work, a bivariate copula model was used to obtain the bivariate joint distribution of flood peak and volume, in order to know the probability of occurrence of a given inflow hydrograph. However, the risk of dam overtopping is given by the maximum water elevation reached during the routing process, which depends on the hydrograph variables, the reservoir volume and the spillway crest length. Consequently, an additional bivariate return period, the so-called routed return period, was defined in terms of risk of dam overtopping based on this maximum water elevation obtained after routing the inflow hydrographs. The theoretical return periods, which give the probability of occurrence of a hydrograph prior to accounting for the reservoir routing, were compared with the routed return period, as in both cases hydrographs with the same probability will draw a curve in the peak-volume space. The procedure was applied to the case study of the Santillana reservoir in Spain. Different reservoir volumes and spillway lengths were considered to investigate the influence of the dam and reservoir characteristics on the results. The methodology improves the estimation of the Design Flood Hydrograph and can be applied to assess the risk of dam overtopping

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La adecuada estimación de avenidas de diseño asociadas a altos periodos de retorno es necesaria para el diseño y gestión de estructuras hidráulicas como presas. En la práctica, la estimación de estos cuantiles se realiza normalmente a través de análisis de frecuencia univariados, basados en su mayoría en el estudio de caudales punta. Sin embargo, la naturaleza de las avenidas es multivariada, siendo esencial tener en cuenta características representativas de las avenidas, tales como caudal punta, volumen y duración del hidrograma, con el fin de llevar a cabo un análisis apropiado; especialmente cuando el caudal de entrada se transforma en un caudal de salida diferente durante el proceso de laminación en un embalse o llanura de inundación. Los análisis de frecuencia de avenidas multivariados han sido tradicionalmente llevados a cabo mediante el uso de distribuciones bivariadas estándar con el fin de modelar variables correlacionadas. Sin embargo, su uso conlleva limitaciones como la necesidad de usar el mismo tipo de distribuciones marginales para todas las variables y la existencia de una relación de dependencia lineal entre ellas. Recientemente, el uso de cópulas se ha extendido en hidrología debido a sus beneficios en relación al contexto multivariado, permitiendo superar los inconvenientes de las técnicas tradicionales. Una copula es una función que representa la estructura de dependencia de las variables de estudio, y permite obtener la distribución de frecuencia multivariada de dichas variables mediante sus distribuciones marginales, sin importar el tipo de distribución marginal utilizada. La estimación de periodos de retorno multivariados, y por lo tanto, de cuantiles multivariados, también se facilita debido a la manera en la que las cópulas están formuladas. La presente tesis doctoral busca proporcionar metodologías que mejoren las técnicas tradicionales usadas por profesionales para estimar cuantiles de avenida más adecuados para el diseño y la gestión de presas, así como para la evaluación del riesgo de avenida, mediante análisis de frecuencia de avenidas bivariados basados en cópulas. Las variables consideradas para ello son el caudal punta y el volumen del hidrograma. Con el objetivo de llevar a cabo un estudio completo, la presente investigación abarca: (i) el análisis de frecuencia de avenidas local bivariado centrado en examinar y comparar los periodos de retorno teóricos basados en la probabilidad natural de ocurrencia de una avenida, con el periodo de retorno asociado al riesgo de sobrevertido de la presa bajo análisis, con el fin de proporcionar cuantiles en una estación de aforo determinada; (ii) la extensión del enfoque local al regional, proporcionando un procedimiento completo para llevar a cabo un análisis de frecuencia de avenidas regional bivariado para proporcionar cuantiles en estaciones sin aforar o para mejorar la estimación de dichos cuantiles en estaciones aforadas; (iii) el uso de cópulas para investigar tendencias bivariadas en avenidas debido al aumento de los niveles de urbanización en una cuenca; y (iv) la extensión de series de avenida observadas mediante la combinación de los beneficios de un modelo basado en cópulas y de un modelo hidrometeorológico. Accurate design flood estimates associated with high return periods are necessary to design and manage hydraulic structures such as dams. In practice, the estimate of such quantiles is usually done via univariate flood frequency analyses, mostly based on the study of peak flows. Nevertheless, the nature of floods is multivariate, being essential to consider representative flood characteristics, such as flood peak, hydrograph volume and hydrograph duration to carry out an appropriate analysis; especially when the inflow peak is transformed into a different outflow peak during the routing process in a reservoir or floodplain. Multivariate flood frequency analyses have been traditionally performed by using standard bivariate distributions to model correlated variables, yet they entail some shortcomings such as the need of using the same kind of marginal distribution for all variables and the assumption of a linear dependence relation between them. Recently, the use of copulas has been extended in hydrology because of their benefits regarding dealing with the multivariate context, as they overcome the drawbacks of the traditional approach. A copula is a function that represents the dependence structure of the studied variables, and allows obtaining the multivariate frequency distribution of them by using their marginal distributions, regardless of the kind of marginal distributions considered. The estimate of multivariate return periods, and therefore multivariate quantiles, is also facilitated by the way in which copulas are formulated. The present doctoral thesis seeks to provide methodologies that improve traditional techniques used by practitioners, in order to estimate more appropriate flood quantiles for dam design, dam management and flood risk assessment, through bivariate flood frequency analyses based on the copula approach. The flood variables considered for that goal are peak flow and hydrograph volume. In order to accomplish a complete study, the present research addresses: (i) a bivariate local flood frequency analysis focused on examining and comparing theoretical return periods based on the natural probability of occurrence of a flood, with the return period associated with the risk of dam overtopping, to estimate quantiles at a given gauged site; (ii) the extension of the local to the regional approach, supplying a complete procedure for performing a bivariate regional flood frequency analysis to either estimate quantiles at ungauged sites or improve at-site estimates at gauged sites; (iii) the use of copulas to investigate bivariate flood trends due to increasing urbanisation levels in a catchment; and (iv) the extension of observed flood series by combining the benefits of a copula-based model and a hydro-meteorological model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Hoy en día, por primera vez en la historia, la mayor parte de la población podrá vivir hasta los sesenta años y más (United Nations, 2015). Sin embargo, todavía existe poca evidencia que demuestre que las personas mayores, estén viviendo con mejor salud que sus padres, a la misma edad, ya que la mayoría de los problemas de salud en edades avanzadas están asociados a las enfermedades crónicas (WHO, 2015). Los sistemas sanitarios de los países desarrollados funcionan adecuadamente cuando se trata del cuidado de enfermedades agudas, pero no son lo suficientemente eficaces en la gestión de las enfermedades crónicas. Durante la última década, se han realizado esfuerzos para mejorar esta gestión, por medio de la utilización de estrategias de prevención y de reenfoque de la provisión de los servicios de atención para la salud (Kane et al. 2005). Según una revisión sistemática de modelos de cuidado de salud, comisionada por el sistema nacional de salud Británico, pocos modelos han conceptualizado cuáles son los componentes que hay que utilizar para proporcionar un cuidado crónico efectivo, y estos componentes no han sido suficientemente estructurados y articulados. Por lo tanto, no hay suficiente evidencia sobre el impacto real de cualquier modelo existente en la actualidad (Ham, 2006). Las innovaciones podrían ayudar a conseguir mejores diagnósticos, tratamientos y gestión de pacientes crónicos, así como a dar soporte a los profesionales y a los pacientes en el cuidado. Sin embargo, la forma en las que estas innovaciones se proporcionan no es lo suficientemente eficiente, efectiva y amigable para el usuario. Para mejorar esto, hace falta crear equipos de trabajo y estrategias multidisciplinares. En conclusión, hacen falta actividades que permitan conseguir que las innovaciones sean utilizadas en los sistemas de salud que quieren mejorar la gestión del cuidado crónico, para que sea posible: 1) traducir la “atención sanitaria basada en la evidencia” en “conocimiento factible”; 2) hacer frente a la complejidad de la atención sanitaria a través de una investigación multidisciplinaria; 3) identificar una aproximación sistemática para que se establezcan intervenciones innovadoras en el cuidado de salud. El marco de referencia desarrollado en este trabajo de investigación es un intento de aportar estas mejoras. Las siguientes hipótesis han sido propuestas: Hipótesis 1: es posible definir un proceso de traducción que convierta un modelo de cuidado crónico en una descripción estructurada de objetivos, requisitos e indicadores clave de rendimiento. Hipótesis 2: el proceso de traducción, si se ejecuta a través de elementos basados en la evidencia, multidisciplinares y de orientación económica, puede convertir un modelo de cuidado crónico en un marco descriptivo, que define el ciclo de vida de soluciones innovadoras para el cuidado de enfermedades crónicas. Hipótesis 3: es posible definir un método para evaluar procesos, resultados y capacidad de desarrollar habilidades, y asistir equipos multidisciplinares en la creación de soluciones innovadoras para el cuidado crónico. Hipótesis 4: es posible dar soporte al desarrollo de soluciones innovadoras para el cuidado crónico a través de un marco de referencia y conseguir efectos positivos, medidos en indicadores clave de rendimiento. Para verificar las hipótesis, se ha definido una aproximación metodológica compuesta de cuatro Fases, cada una asociada a una hipótesis. Antes de esto, se ha llevado a cabo una “Fase 0”, donde se han analizado los antecedentes sobre el problema (i.e. adopción sistemática de la innovación en el cuidado crónico) desde una perspectiva multi-dominio y multi-disciplinar. Durante la fase 1, se ha desarrollado un Proceso de Traducción del Conocimiento, elaborado a partir del JBI Joanna Briggs Institute (JBI) model of evidence-based healthcare (Pearson, 2005), y sobre el cual se han definido cuatro Bloques de Innovación. Estos bloques consisten en una descripción de elementos innovadores, definidos en la fase 0, que han sido añadidos a los cuatros elementos que componen el modelo JBI. El trabajo llevado a cabo en esta fase ha servido también para definir los materiales que el proceso de traducción tiene que ejecutar. La traducción que se ha llevado a cabo en la fase 2, y que traduce la mejor evidencia disponible de cuidado crónico en acción: resultado de este proceso de traducción es la parte descriptiva del marco de referencia, que consiste en una descripción de un modelo de cuidado crónico (se ha elegido el Chronic Care Model, Wagner, 1996) en términos de objetivos, especificaciones e indicadores clave de rendimiento y organizada en tres ciclos de innovación (diseño, implementación y evaluación). Este resultado ha permitido verificar la segunda hipótesis. Durante la fase 3, para demostrar la tercera hipótesis, se ha desarrollado un método-mixto de evaluación de equipos multidisciplinares que trabajan en innovaciones para el cuidado crónico. Este método se ha creado a partir del método mixto usado para la evaluación de equipo multidisciplinares translacionales (Wooden, 2013). El método creado añade una dimensión procedural al marco. El resultado de esta fase consiste, por lo tanto, en una primera versión del marco de referencia, lista para ser experimentada. En la fase 4, se ha validado el marco a través de un caso de estudio multinivel y con técnicas de observación-participante como método de recolección de datos. Como caso de estudio se han elegido las actividades de investigación que el grupo de investigación LifeStech ha desarrollado desde el 2008 para mejorar la gestión de la diabetes, actividades realizadas en un contexto internacional. Los resultados demuestran que el marco ha permitido mejorar las actividades de trabajo en distintos niveles: 1) la calidad y cantidad de las publicaciones; 2) se han conseguido dos contratos de investigación sobre diabetes: el primero es un proyecto de investigación aplicada, el segundo es un proyecto financiado para acelerar las innovaciones en el mercado; 3) a través de los indicadores claves de rendimiento propuestos en el marco, una prueba de concepto de un prototipo desarrollado en un proyecto de investigación ha sido transformada en una evaluación temprana de una intervención eHealth para el manejo de la diabetes, que ha sido recientemente incluida en Repositorio de prácticas innovadoras del Partenariado de Innovación Europeo en Envejecimiento saludable y activo. La verificación de las 4 hipótesis ha permitido demonstrar la hipótesis principal de este trabajo de investigación: es posible contribuir a crear un puente entre la atención sanitaria y la innovación y, por lo tanto, mejorar la manera en que el cuidado crónico sea procurado en los sistemas sanitarios. ABSTRACT Nowadays, for the first time in history, most people can expect to live into their sixties and beyond (United Nations, 2015). However, little evidence suggests that older people are experiencing better health than their parents, and most of the health problems of older age are linked to Chronic Diseases (WHO, 2015). The established health care systems in developed countries are well suited to the treatment of acute diseases but are mostly inadequate for dealing with CDs. Healthcare systems are challenging the burden of chronic diseases by putting more emphasis on the prevention of disease and by looking for new ways to reorient the provision of care (Kane et al., 2005). According to an evidence-based review commissioned by the British NHS Institute, few models have conceptualized effective components of care for CDs and these components have been not structured and articulated. “Consequently, there is limited evidence about the real impact of any of the existing models” (Ham, 2006). Innovations could support to achieve better diagnosis, treatment and management for patients across the continuum of care, by supporting health professionals and empowering patients to take responsibility. However, the way they are delivered is not sufficiently efficient, effective and consumer friendly. The improvement of innovation delivery, involves the creation of multidisciplinary research teams and taskforces, rather than just working teams. There are several actions to improve the adoption of innovations from healthcare systems that are tackling the epidemics of CDs: 1) Translate Evidence-Based Healthcare (EBH) into actionable knowledge; 2) Face the complexity of healthcare through multidisciplinary research; 3) Identify a systematic approach to support effective implementation of healthcare interventions through innovation. The framework proposed in this research work is an attempt to provide these improvements. The following hypotheses have been drafted: Hypothesis 1: it is possible to define a translation process to convert a model of chronic care into a structured description of goals, requirements and key performance indicators. Hypothesis 2: a translation process, if executed through evidence-based, multidisciplinary, holistic and business-oriented elements, can convert a model of chronic care in a descriptive framework, which defines the whole development cycle of innovative solutions for chronic disease management. Hypothesis 3: it is possible to design a method to evaluate processes, outcomes and skill acquisition capacities, and assist multidisciplinary research teams in the creation of innovative solutions for chronic disease management. Hypothesis 4: it is possible to assist the development of innovative solutions for chronic disease management through a reference framework and produce positive effects, measured through key performance indicators. In order to verify the hypotheses, a methodological approach, composed of four Phases that correspond to each one of the stated hypothesis, was defined. Prior to this, a “Phase 0”, consisting in a multi-domain and multi-disciplinary background analysis of the problem (i.e.: systematic adoption of innovation to chronic care), was carried out. During phase 1, in order to verify the first hypothesis, a Knowledge Translation Process (KTP) was developed, starting from the JBI Joanna Briggs Institute (JBI) model of evidence-based healthcare was used (Pearson, 2005) and adding Four Innovation Blocks. These blocks represent an enriched description, added to the JBI model, to accelerate the transformation of evidence-healthcare through innovation; the innovation blocks are built on top of the conclusions drawn after Phase 0. The background analysis gave also indication on the materials and methods to be used for the execution of the KTP, carried out during phase 2, that translates the actual best available evidence for chronic care into action: this resulted in a descriptive Framework, which is a description of a model of chronic care (the Chronic Care Model was chosen, Wagner, 1996) in terms of goals, specified requirements and Key Performance Indicators, and articulated in the three development cycles of innovation (i.e. design, implementation and evaluation). Thanks to this result the second hypothesis was verified. During phase 3, in order to verify the third hypothesis, a mixed-method to evaluate multidisciplinary teams working on innovations for chronic care, was created, based on a mixed-method used for the evaluation of Multidisciplinary Translational Teams (Wooden, 2013). This method adds a procedural dimension to the descriptive component of the Framework, The result of this phase consisted in a draft version of the framework, ready to be tested in a real scenario. During phase 4, a single and multilevel case study, with participant-observation data collection, was carried out, in order to have a complete but at the same time multi-sectorial evaluation of the framework. The activities that the LifeStech research group carried out since 2008 to improve the management of diabetes have been selected as case study. The results achieved showed that the framework allowed to improve the research activities in different directions: the quality and quantity of the research publications that LifeStech has issued, have increased substantially; 2 project grants to improve the management of diabetes, have been assigned: the first is a grant funding applied research while the second is about accelerating innovations into the market; by using the assessment KPIs of the framework, the proof of concept validation of a prototype developed in a research project was transformed into an early stage assessment of innovative eHealth intervention for Diabetes Management, which has been recently included in the repository of innovative practice of the European Innovation Partnership on Active and Health Ageing initiative. The verification of the 4 hypotheses lead to verify the main hypothesis of this research work: it is possible to contribute to bridge the gap between healthcare and innovation and, in turn, improve the way chronic care is delivered by healthcare systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cover title.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cover title.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVES: The objective of this research was to design a clinical decision support system (CDSS) that supports heterogeneous clinical decision problems and runs on multiple computing platforms. Meeting this objective required a novel design to create an extendable and easy to maintain clinical CDSS for point of care support. The proposed solution was evaluated in a proof of concept implementation. METHODS: Based on our earlier research with the design of a mobile CDSS for emergency triage we used ontology-driven design to represent essential components of a CDSS. Models of clinical decision problems were derived from the ontology and they were processed into executable applications during runtime. This allowed scaling applications' functionality to the capabilities of computing platforms. A prototype of the system was implemented using the extended client-server architecture and Web services to distribute the functions of the system and to make it operational in limited connectivity conditions. RESULTS: The proposed design provided a common framework that facilitated development of diversified clinical applications running seamlessly on a variety of computing platforms. It was prototyped for two clinical decision problems and settings (triage of acute pain in the emergency department and postoperative management of radical prostatectomy on the hospital ward) and implemented on two computing platforms-desktop and handheld computers. CONCLUSIONS: The requirement of the CDSS heterogeneity was satisfied with ontology-driven design. Processing of application models described with the help of ontological models allowed having a complex system running on multiple computing platforms with different capabilities. Finally, separation of models and runtime components contributed to improved extensibility and maintainability of the system.