897 resultados para Simulation-based methods


Relevância:

90.00% 90.00%

Publicador:

Resumo:

BACKGROUND: HIV surveillance requires monitoring of new HIV diagnoses and differentiation of incident and older infections. In 2008, Switzerland implemented a system for monitoring incident HIV infections based on the results of a line immunoassay (Inno-Lia) mandatorily conducted for HIV confirmation and type differentiation (HIV-1, HIV-2) of all newly diagnosed patients. Based on this system, we assessed the proportion of incident HIV infection among newly diagnosed cases in Switzerland during 2008-2013. METHODS AND RESULTS: Inno-Lia antibody reaction patterns recorded in anonymous HIV notifications to the federal health authority were classified by 10 published algorithms into incident (up to 12 months) or older infections. Utilizing these data, annual incident infection estimates were obtained in two ways, (i) based on the diagnostic performance of the algorithms and utilizing the relationship 'incident = true incident + false incident', (ii) based on the window-periods of the algorithms and utilizing the relationship 'Prevalence = Incidence x Duration'. From 2008-2013, 3'851 HIV notifications were received. Adult HIV-1 infections amounted to 3'809 cases, and 3'636 of them (95.5%) contained Inno-Lia data. Incident infection totals calculated were similar for the performance- and window-based methods, amounting on average to 1'755 (95% confidence interval, 1588-1923) and 1'790 cases (95% CI, 1679-1900), respectively. More than half of these were among men who had sex with men. Both methods showed a continuous decline of annual incident infections 2008-2013, totaling -59.5% and -50.2%, respectively. The decline of incident infections continued even in 2012, when a 15% increase in HIV notifications had been observed. This increase was entirely due to older infections. Overall declines 2008-2013 were of similar extent among the major transmission groups. CONCLUSIONS: Inno-Lia based incident HIV-1 infection surveillance proved useful and reliable. It represents a free, additional public health benefit of the use of this relatively costly test for HIV confirmation and type differentiation.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

BACKGROUND: Available methods to simulate nucleotide or amino acid data typically use Markov models to simulate each position independently. These approaches are not appropriate to assess the performance of combinatorial and probabilistic methods that look for coevolving positions in nucleotide or amino acid sequences. RESULTS: We have developed a web-based platform that gives a user-friendly access to two phylogenetic-based methods implementing the Coev model: the evaluation of coevolving scores and the simulation of coevolving positions. We have also extended the capabilities of the Coev model to allow for the generalization of the alphabet used in the Markov model, which can now analyse both nucleotide and amino acid data sets. The simulation of coevolving positions is novel and builds upon the developments of the Coev model. It allows user to simulate pairs of dependent nucleotide or amino acid positions. CONCLUSIONS: The main focus of our paper is the new simulation method we present for coevolving positions. The implementation of this method is embedded within the web platform Coev-web that is freely accessible at http://coev.vital-it.ch/, and was tested in most modern web browsers.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The objective of this study is to show that bone strains due to dynamic mechanical loading during physical activity can be analysed using the flexible multibody simulation approach. Strains within the bone tissue play a major role in bone (re)modeling. Based on previous studies, it has been shown that dynamic loading seems to be more important for bone (re)modeling than static loading. The finite element method has been used previously to assess bone strains. However, the finite element method may be limited to static analysis of bone strains due to the expensive computation required for dynamic analysis, especially for a biomechanical system consisting of several bodies. Further, in vivo implementation of strain gauges on the surfaces of bone has been used previously in order to quantify the mechanical loading environment of the skeleton. However, in vivo strain measurement requires invasive methodology, which is challenging and limited to certain regions of superficial bones only, such as the anterior surface of the tibia. In this study, an alternative numerical approach to analyzing in vivo strains, based on the flexible multibody simulation approach, is proposed. In order to investigate the reliability of the proposed approach, three 3-dimensional musculoskeletal models where the right tibia is assumed to be flexible, are used as demonstration examples. The models are employed in a forward dynamics simulation in order to predict the tibial strains during walking on a level exercise. The flexible tibial model is developed using the actual geometry of the subject’s tibia, which is obtained from 3 dimensional reconstruction of Magnetic Resonance Images. Inverse dynamics simulation based on motion capture data obtained from walking at a constant velocity is used to calculate the desired contraction trajectory for each muscle. In the forward dynamics simulation, a proportional derivative servo controller is used to calculate each muscle force required to reproduce the motion, based on the desired muscle contraction trajectory obtained from the inverse dynamics simulation. Experimental measurements are used to verify the models and check the accuracy of the models in replicating the realistic mechanical loading environment measured from the walking test. The predicted strain results by the models show consistency with literature-based in vivo strain measurements. In conclusion, the non-invasive flexible multibody simulation approach may be used as a surrogate for experimental bone strain measurement, and thus be of use in detailed strain estimation of bones in different applications. Consequently, the information obtained from the present approach might be useful in clinical applications, including optimizing implant design and devising exercises to prevent bone fragility, accelerate fracture healing and reduce osteoporotic bone loss.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Hypothesis: The quality of care for chronic patients depends on the collaborative skills of the healthcare providers.1,2 The literature lacks reports of the use of simulation to teach collaborative skills in non-acute care settings. We posit that simulation offers benefits for supporting the development of collaborative practice in non-acute settings. We explored the benefits and challenges of using an Interprofessional Team - Objective Structured Clinical Examination (IT-OSCE) as a formative assessment tool. IT-OSCE is an intervention which involves an interprofessional team of trainees interacting with a simulated patient (SP) enabling them to practice collaborative skills in non-acute care settings.5 A simulated patient are people trained to portray patients in a simulated scenario for educational purposes.6,7 Since interprofessional education (IPE) ultimately aims to provide collaborative patient-centered care.8,9 We sought to promote patient-centeredness in the learning process. Methods: The IT-OSCE was conducted with four trios of students from different professions. The debriefing was co-facilitated by the SP with a faculty. The participants were final-year students in nursing, physiotherapy and medicine. Our research question focused on the introduction of co-facilitated (SP and faculty) debriefing after an IT-OSCE: 1) What are the benefits and challenges of involving the SP during the debriefing? and 2) To evaluate the IT-OSCE, an exploratory case study was used to provide fine grained data 10, 11. Three focus groups were conducted - two with students (n=6; n=5), one with SPs (n=3) and one with faculty (n=4). Audiotapes were transcribed for thematic analysis performed by three researchers, who found a consensus on the final set of themes. Results: The thematic analysis showed little differentiation between SPs, student and faculty perspectives. The analysis of transcripts revealed more particularly, that the SP's co-facilitation during the debriefing of an IT-OSCE proved to be feasible. It was appreciated by all the participants and appeared to value and to promote patient-centeredness in the learning process. The main challenge consisted in SPs feedback, more particularly in how they could report accurate observations to a students' group rather than individual students. Conclusion: In conclusion, SP methodology using an IT-OSCE seems to be a useful and promising way to train collaborative skills, aligning IPE, simulation-based team training in a non-acute care setting and patient-centeredness. We acknowledge the limitations of the study, especially the small sample and consider the exploration of SP-based IPE in non-acute care settings as strength. Future studies could consider the preparation of SPs and faculty as co-facilitators. References: 1. Borrill CS, Carletta J, Carter AJ, et al. The effectiveness of health care teams in the National Health Service. Aston centre for Health Service Organisational Research. 2001. 2. Reeves S, Lewin S, Espin S, Zwarenstein M. Interprofessional teamwork for health and social care. Oxford: Wiley-Blackwell; 2010. 3. Issenberg S, McGaghie WC, Petrusa ER, Gordon DL, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning - a BEME systematic review. Medical Teacher. 2005;27(1):10-28. 4. McGaghie W, Petrusa ER, Gordon DL, Scalese RJ. A critical review of simulation-based medical education research: 2003-2009. Medical Education. 2010;44(1):50-63. 5. Simmons B, Egan-Lee E, Wagner SJ, Esdaile M, Baker L, Reeves S. Assessment of interprofessional learning: the design of an interprofessional objective structured clinical examination (iOSCE) approach. Journal of Interprofessional Care. 2011;25(1):73-74. 6. Nestel D, Layat Burn C, Pritchard SA, Glastonbury R, Tabak D. The use of simulated patients in medical education: Guide Supplement 42.1 - Viewpoint. Medical teacher. 2011;33(12):1027-1029. Disclosures: None (C) 2014 by Lippincott Williams & Wilkins, Inc.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Teaching the measurement of blood pressure for both nursing and public health nursing students The purpose of this two-phase study was to develop the teaching of blood pressure measurement within the nursing degree programmes of the Universities of Applied Sciences. The first survey phase described what and how blood pressure measurement was taught within nursing degree programmes. The second intervention phase (2004-2005) evaluated first academic year nursing and public health nursing students’ knowledge and skills results for blood pressure measurement. Additionally, the effect on the Taitoviikko experimental group students’ blood pressure measurement knowledge and skills level. A further objective was to construct models for an instrument (RRmittTest) to evaluate nursing students measurement of blood pressure (2003-2009). The research data for the survey phase were collected from teachers (total sampling, N=107, response rate 77%) using a specially developed RRmittopetus-questionnaire. Quasi-experimental study data on the RRmittTest-instrument was collected from students (purposive sampling, experimental group, n=29, control group, n=44). The RRmittTest consisted of a test of knowledge (Tietotesti) and simulation-based test (TaitoSimkäsi and Taitovideo) of skills. Measurements were made immediately after the teaching and in clinical practice. Statistical methods were used to analyse the results and responses to open-ended questions were organised and classified. Due to the small amount of materials involved and the results of distribution tests of the variables, non-parametric analytic methods were mainly used. Experimental group and control group similar knowledge and skills teaching was based on the results of the national survey phase (RRmittopetus) questionnaire results. Experimental group teaching includes the supervised Taitoviikko teaching method. During Taitoviikko students studied blood pressure measurement at the municipal hospital in a real nursing environment, guided by a teacher and a clinical nursing professional. In order to evaluate both learning and teaching the processes and components of blood pressure measurement were clearly defined as follows: the reliability of measurement instruments, activities preceding blood pressure measurement, technical execution of the measurement, recording, lifestyle guidance and measurement at home (self-monitoring). According to the survey study, blood pressure measurement is most often taught at Universities of Applied Sciences, separately, as knowledge (teaching of theory, 2 hours) and skills (classroom practice, 4 hours). The teaching was implemented largely in a classroom and was based mainly on a textbook. In the intervention phase the students had good knowledge of blood pressure measurement. However, their blood pressure measurement skills were deficient and the control group students, in particular, were highly deficient. Following in clinical practice the experimental group and control group students’ blood pressure measurement recording knowledge improve and experimental groups declined lifestyle guidance. Skills did not improve within any of the components analysed. The control groups` skills on the whole, declined statistically.There was a significant decline amongst the experimental group although only in one component measured. The results describe the learning results for first academic year students and no parallel conclusions should be drawn when considering any learning results for graduating students. The results support the use and further development of the Taitoviiko teaching method. The RRmittTest developed for the study should be assessed and the results seen from a negative perspective. This evaluation tool needs to be developed and retested.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The amount of installed wind power has been growing exponentially during the past ten years. As wind turbines have become a significant source of electrical energy, the interactions between the turbines and the electric power network need to be studied more thoroughly than before. Especially, the behavior of the turbines in fault situations is of prime importance; simply disconnecting all wind turbines from the network during a voltage drop is no longer acceptable, since this would contribute to a total network collapse. These requirements have been a contributor to the increased role of simulations in the study and design of the electric drive train of a wind turbine. When planning a wind power investment, the selection of the site and the turbine are crucial for the economic feasibility of the installation. Economic feasibility, on the other hand, is the factor that determines whether or not investment in wind power will continue, contributing to green electricity production and reduction of emissions. In the selection of the installation site and the turbine (siting and site matching), the properties of the electric drive train of the planned turbine have so far been generally not been taken into account. Additionally, although the loss minimization of some of the individual components of the drive train has been studied, the drive train as a whole has received less attention. Furthermore, as a wind turbine will typically operate at a power level lower than the nominal most of the time, efficiency analysis in the nominal operating point is not sufficient. This doctoral dissertation attempts to combine the two aforementioned areas of interest by studying the applicability of time domain simulations in the analysis of the economicfeasibility of a wind turbine. The utilization of a general-purpose time domain simulator, otherwise applied to the study of network interactions and control systems, in the economic analysis of the wind energy conversion system is studied. The main benefits of the simulation-based method over traditional methods based on analytic calculation of losses include the ability to reuse and recombine existing models, the ability to analyze interactions between the components and subsystems in the electric drive train (something which is impossible when considering different subsystems as independent blocks, as is commonly done in theanalytical calculation of efficiencies), the ability to analyze in a rather straightforward manner the effect of selections other than physical components, for example control algorithms, and the ability to verify assumptions of the effects of a particular design change on the efficiency of the whole system. Based on the work, it can be concluded that differences between two configurations can be seen in the economic performance with only minor modifications to the simulation models used in the network interaction and control method study. This eliminates the need ofdeveloping analytic expressions for losses and enables the study of the system as a whole instead of modeling it as series connection of independent blocks with no lossinterdependencies. Three example cases (site matching, component selection, control principle selection) are provided to illustrate the usage of the approach and analyze its performance.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper, we review the advances of monocular model-based tracking for last ten years period until 2014. In 2005, Lepetit, et. al, [19] reviewed the status of monocular model based rigid body tracking. Since then, direct 3D tracking has become quite popular research area, but monocular model-based tracking should still not be forgotten. We mainly focus on tracking, which could be applied to aug- mented reality, but also some other applications are covered. Given the wide subject area this paper tries to give a broad view on the research that has been conducted, giving the reader an introduction to the different disciplines that are tightly related to model-based tracking. The work has been conducted by searching through well known academic search databases in a systematic manner, and by selecting certain publications for closer examination. We analyze the results by dividing the found papers into different categories by their way of implementation. The issues which have not yet been solved are discussed. We also discuss on emerging model-based methods such as fusing different types of features and region-based pose estimation which could show the way for future research in this subject.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Tannins, typically segregated into two major groups, the hydrolyzable tannins (HTs) and the proanthocyanidins (PAs), are plant polyphenolic secondary metabolites found throughout the plant kingdom. On one hand, tannins may cause harmful nutritional effects on herbivores, for example insects, and hence they work as plants’ defense against plant-eating animals. On the other hand, they may affect positively some herbivores, such as mammals, for example by their antioxidant, antimicrobial, anti-inflammatory or anticarcinogenic activities. This thesis focuses on understanding the bioactivity of plant tannins, their anthelmintic properties and the tools used for the qualitative and quantitative analysis of this endless source of structural diversity. The first part of the experimental work focused on the development of ultra-high performance liquid chromatography−tandem mass spectrometry (UHPLC-MS/MS) based methods for the rapid fingerprint analysis of bioactive polyphenols, especially tannins. In the second part of the experimental work the in vitro activity of isolated and purified HTs and their hydrolysis product, gallic acid, was tested against egg hatching and larval motility of two larval developmental stages, L1 and L2, of a common ruminant gastrointestinal parasite, Haemonchus contortus. The results indicated clear relationships between the HT structure and the anthelmintic activity. The activity of the studied compounds depended on many structural features, including size, functional groups present in the structure, and the structural rigidness. To further understand tannin bioactivity on a molecular level, the interaction between bovine serum albumin (BSA), and seven HTs and epigallocatechin gallate was examined. The objective was to define the effect of pH on the formation on tannin–protein complexes and to evaluate the stability of the formed complexes by gel electrophoresis and MALDI-TOF-MS. The results indicated that more basic pH values had a stabilizing effect on the tannin–protein complexes and that the tannin oxidative activity was directly linked with their tendency to form covalently stabilized complexes with BSA at increased pH.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper proposes finite-sample procedures for testing the SURE specification in multi-equation regression models, i.e. whether the disturbances in different equations are contemporaneously uncorrelated or not. We apply the technique of Monte Carlo (MC) tests [Dwass (1957), Barnard (1963)] to obtain exact tests based on standard LR and LM zero correlation tests. We also suggest a MC quasi-LR (QLR) test based on feasible generalized least squares (FGLS). We show that the latter statistics are pivotal under the null, which provides the justification for applying MC tests. Furthermore, we extend the exact independence test proposed by Harvey and Phillips (1982) to the multi-equation framework. Specifically, we introduce several induced tests based on a set of simultaneous Harvey/Phillips-type tests and suggest a simulation-based solution to the associated combination problem. The properties of the proposed tests are studied in a Monte Carlo experiment which shows that standard asymptotic tests exhibit important size distortions, while MC tests achieve complete size control and display good power. Moreover, MC-QLR tests performed best in terms of power, a result of interest from the point of view of simulation-based tests. The power of the MC induced tests improves appreciably in comparison to standard Bonferroni tests and, in certain cases, outperforms the likelihood-based MC tests. The tests are applied to data used by Fischer (1993) to analyze the macroeconomic determinants of growth.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We study the problem of testing the error distribution in a multivariate linear regression (MLR) model. The tests are functions of appropriately standardized multivariate least squares residuals whose distribution is invariant to the unknown cross-equation error covariance matrix. Empirical multivariate skewness and kurtosis criteria are then compared to simulation-based estimate of their expected value under the hypothesized distribution. Special cases considered include testing multivariate normal, Student t; normal mixtures and stable error models. In the Gaussian case, finite-sample versions of the standard multivariate skewness and kurtosis tests are derived. To do this, we exploit simple, double and multi-stage Monte Carlo test methods. For non-Gaussian distribution families involving nuisance parameters, confidence sets are derived for the the nuisance parameters and the error distribution. The procedures considered are evaluated in a small simulation experi-ment. Finally, the tests are applied to an asset pricing model with observable risk-free rates, using monthly returns on New York Stock Exchange (NYSE) portfolios over five-year subperiods from 1926-1995.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

De plus en plus de recherches sur les Interactions Humain-Machine (IHM) tentent d’effectuer des analyses fines de l’interaction afin de faire ressortir ce qui influence les comportements des utilisateurs. Tant au niveau de l’évaluation de la performance que de l’expérience des utilisateurs, on note qu’une attention particulière est maintenant portée aux réactions émotionnelles et cognitives lors de l’interaction. Les approches qualitatives standards sont limitées, car elles se fondent sur l’observation et des entrevues après l’interaction, limitant ainsi la précision du diagnostic. L’expérience utilisateur et les réactions émotionnelles étant de nature hautement dynamique et contextualisée, les approches d’évaluation doivent l’être de même afin de permettre un diagnostic précis de l’interaction. Cette thèse présente une approche d’évaluation quantitative et dynamique qui permet de contextualiser les réactions des utilisateurs afin d’en identifier les antécédents dans l’interaction avec un système. Pour ce faire, ce travail s’articule autour de trois axes. 1) La reconnaissance automatique des buts et de la structure de tâches de l’utilisateur, à l’aide de mesures oculométriques et d’activité dans l’environnement par apprentissage machine. 2) L’inférence de construits psychologiques (activation, valence émotionnelle et charge cognitive) via l’analyse des signaux physiologiques. 3) Le diagnostic de l‘interaction reposant sur le couplage dynamique des deux précédentes opérations. Les idées et le développement de notre approche sont illustrés par leur application dans deux contextes expérimentaux : le commerce électronique et l’apprentissage par simulation. Nous présentons aussi l’outil informatique complet qui a été implémenté afin de permettre à des professionnels en évaluation (ex. : ergonomes, concepteurs de jeux, formateurs) d’utiliser l’approche proposée pour l’évaluation d’IHM. Celui-ci est conçu de manière à faciliter la triangulation des appareils de mesure impliqués dans ce travail et à s’intégrer aux méthodes classiques d’évaluation de l’interaction (ex. : questionnaires et codage des observations).

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Les centres d’appels sont des éléments clés de presque n’importe quelle grande organisation. Le problème de gestion du travail a reçu beaucoup d’attention dans la littérature. Une formulation typique se base sur des mesures de performance sur un horizon infini, et le problème d’affectation d’agents est habituellement résolu en combinant des méthodes d’optimisation et de simulation. Dans cette thèse, nous considérons un problème d’affection d’agents pour des centres d’appels soumis a des contraintes en probabilité. Nous introduisons une formulation qui exige que les contraintes de qualité de service (QoS) soient satisfaites avec une forte probabilité, et définissons une approximation de ce problème par moyenne échantillonnale dans un cadre de compétences multiples. Nous établissons la convergence de la solution du problème approximatif vers celle du problème initial quand la taille de l’échantillon croit. Pour le cas particulier où tous les agents ont toutes les compétences (un seul groupe d’agents), nous concevons trois méthodes d’optimisation basées sur la simulation pour le problème de moyenne échantillonnale. Étant donné un niveau initial de personnel, nous augmentons le nombre d’agents pour les périodes où les contraintes sont violées, et nous diminuons le nombre d’agents pour les périodes telles que les contraintes soient toujours satisfaites après cette réduction. Des expériences numériques sont menées sur plusieurs modèles de centre d’appels à faible occupation, au cours desquelles les algorithmes donnent de bonnes solutions, i.e. la plupart des contraintes en probabilité sont satisfaites, et nous ne pouvons pas réduire le personnel dans une période donnée sont introduire de violation de contraintes. Un avantage de ces algorithmes, par rapport à d’autres méthodes, est la facilité d’implémentation.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Objectifs: Les patients hospitalisés aux soins intensifs (SI) sont souvent victimes d’erreurs médicales. La nature interprofessionnelle des équipes de SI les rend vulnérables aux erreurs de communication. L’objectif primaire du projet est d’améliorer la communication dans une équipe interprofessionnelle de soins intensifs par une formation en simulation à haute fidélité. Méthodologie Une étude prospective randomisée contrôlée à double insu a été réalisée. Dix équipes de six professionnels de SI ont complété trois scénarios de simulations de réanimation. Le groupe intervention était débreffé sur des aspects de communication alors que le groupe contrôle était débreffé sur des aspects techniques de réanimation. Trois mois plus tard, les équipes réalisaient une quatrième simulation sans débreffage. Les simulations étaient toutes évaluées pour la qualité, l’efficacité de la communication et le partage des informations critiques par quatre évaluateurs. Résultats Pour l’issue primaire, il n’y a pas eu d’amélioration plus grande de la communication dans le groupe intervention en comparaison avec le groupe contrôle. Une amélioration de 16% de l’efficacité des communications a été notée dans les équipes de soins intensifs indépendamment du groupe étudié. Les infirmiers et les inhalothérapeutes ont amélioré significativement l’efficacité de la communication après trois sessions. L’effet observé ne s’est pas maintenu à trois mois. Conclusion Une formation sur simulateur à haute fidélité couplée à un débreffage peut améliorer à court terme l’efficacité des communications dans une équipe interprofessionnelle de SI.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Die in dieser Arbeit durchgeführten Untersuchungen zeigen, daß es möglich ist, komplexe thermische Systeme, unter Verwendung der thermisch-, elektrischen Analogien, mit PSpice zu simulieren. Im Mittelpunkt der Untersuchungen standen hierbei Strangkühlkörper zur Kühlung von elektronischen Bauelementen. Es konnte gezeigt werden,daß alle Wärmeübertragungsarten, (Wärmeleitung, Konvektion und Wärmestrahlung) in der Simulation berücksichtigt werden können. Für die Berechnung der Konvektion wurden verschiedene Methoden hergeleitet. Diese gelten zum einen für verschiedene Kühlkörpergeometrien, wie z.B. ebene Flächen und Kühlrippenzwischenräume, andererseits unterscheiden sie sich, je nachdem, ob freie oder erzwungene Konvektion betrachtet wird. Für die Wärmestrahlung zwischen den Kühlrippen wurden verschiedenen Berechnungsmethoden entwickelt. Für die Simulation mit PSpice wurde die Berechnung der Wärmestrahlung zwischen den Kühlrippen vereinfacht. Es konnte gezeigt werden, daß die Fehler, die durch die Vereinfachung entstehen, vernachlässigbar klein sind. Für das thermische Verhalten einer zu kühlenden Wärmequelle wurde ein allgemeines Modell entworfen. Zur Bestimmung der Modellparameter wurden verschiedene Meßverfahren entwickelt. Für eine im Fachgebiet Elektromechanik entwickelte Wärmequelle zum Test von Kühlvorrichtungen wurde mit Hilfe dieser Meßverfahren eine Parameterbestimmung durchgeführt. Die Erstellung des thermischen Modells eines Kühlkörpers für die Simulation in PSpice erfordert die Analyse der Kühlkörpergeometrie. Damit diese Analyse weitestgehend automatisiert werden kann, wurden verschiedene Algorithmen unter Matlab entwickelt. Es wurde ein Algorithmus entwickelt, der es ermöglicht, den Kühlkörper in Elementarzellen zu zerlegen, die für die Erstellung des Simulationsmodells benötigt werden. Desweiteren ist es für die Simulation notwendig zu wissen, welche der Elementarzellen am Rand des Kühlkörpers liegen, welche der Elementarzellen an einem Kühlrippenzwischenraum liegen und welche Kühlkörperkanten schräg verlaufen. Auch zur Lösung dieser Aufgaben wurden verschiedene Algorithmen entwickelt. Diese Algorithmen wurden zu einem Programm zusammengefaßt, das es gestattet, unterschiedliche Strangkühlkörper zu simulieren und die Simulationsergebnisse in Form der Temperaturverteilung auf der Montagefläche des Kühlkörpers grafisch darzustellen. Es können stationäre und transiente Simulationen durchgeführt werden. Desweiteren kann der thermische Widerstand des Kühlkörpers RthK als Funktion der Verlustleistung der Wärmequelle dargestellt werden. Zur Verifikation der Simulationsergebnisse wurden Temperaturmessungen an Kühlkörpern durchgeführt und mit den Simulationsergebnissen verglichen. Diese Vergleiche zeigen, daß die Abweichungen im Bereich der Streuung der Temperaturmessung liegen. Das hier entwickelte Verfahren zur thermischen Simulation von Strangkühlkörpern kann somit als gut bewertet werden.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Mit aktiven Magnetlagern ist es möglich, rotierende Körper durch magnetische Felder berührungsfrei zu lagern. Systembedingt sind bei aktiv magnetgelagerten Maschinen wesentliche Signale ohne zusätzlichen Aufwand an Messtechnik für Diagnoseaufgaben verfügbar. In der Arbeit wird ein Konzept entwickelt, das durch Verwendung der systeminhärenten Signale eine Diagnose magnetgelagerter rotierender Maschinen ermöglicht und somit neben einer kontinuierlichen Anlagenüberwachung eine schnelle Bewertung des Anlagenzustandes gestattet. Fehler können rechtzeitig und ursächlich in Art und Größe erkannt und entsprechende Gegenmaßnahmen eingeleitet werden. Anhand der erfassten Signale geschieht die Gewinnung von Merkmalen mit signal- und modellgestützten Verfahren. Für den Magnetlagerregelkreis erfolgen Untersuchungen zum Einsatz modellgestützter Parameteridentifikationsverfahren, deren Verwendbarkeit wird bei der Diagnose am Regler und Leistungsverstärker nachgewiesen. Unter Nutzung von Simulationsmodellen sowie durch Experimente an Versuchsständen werden die Merkmalsverläufe im normalen Referenzzustand und bei auftretenden Fehlern aufgenommen und die Ergebnisse in einer Wissensbasis abgelegt. Diese dient als Grundlage zur Festlegung von Grenzwerten und Regeln für die Überwachung des Systems und zur Erstellung wissensbasierter Diagnosemodelle. Bei der Überwachung werden die Merkmalsausprägungen auf das Überschreiten von Grenzwerten überprüft, Informationen über erkannte Fehler und Betriebszustände gebildet sowie gegebenenfalls Alarmmeldungen ausgegeben. Sich langsam anbahnende Fehler können durch die Berechnung der Merkmalstrends mit Hilfe der Regressionsanalyse erkannt werden. Über die bisher bei aktiven Magnetlagern übliche Überwachung von Grenzwerten hinaus erfolgt bei der Fehlerdiagnose eine Verknüpfung der extrahierten Merkmale zur Identifizierung und Lokalisierung auftretender Fehler. Die Diagnose geschieht mittels regelbasierter Fuzzy-Logik, dies gestattet die Einbeziehung von linguistischen Aussagen in Form von Expertenwissen sowie die Berücksichtigung von Unbestimmtheiten und ermöglicht damit eine Diagnose komplexer Systeme. Für Aktor-, Sensor- und Reglerfehler im Magnetlagerregelkreis sowie Fehler durch externe Kräfte und Unwuchten werden Diagnosemodelle erstellt und verifiziert. Es erfolgt der Nachweis, dass das entwickelte Diagnosekonzept mit beherrschbarem Rechenaufwand korrekte Diagnoseaussagen liefert. Durch Kaskadierung von Fuzzy-Logik-Modulen wird die Transparenz des Regelwerks gewahrt und die Abarbeitung der Regeln optimiert. Endresultat ist ein neuartiges hybrides Diagnosekonzept, welches signal- und modellgestützte Verfahren der Merkmalsgewinnung mit wissensbasierten Methoden der Fehlerdiagnose kombiniert. Das entwickelte Diagnosekonzept ist für die Anpassung an unterschiedliche Anforderungen und Anwendungen bei rotierenden Maschinen konzipiert.