896 resultados para sensor grid database system
Resumo:
Ireland’s remote position on the tip of Europe ensures that the country is vulnerable to uncertainty of supply. The reliance on conventional sources of electricity has ensured that escalated prices and high carbon emissions have been witnessed whilst opportunities that inherent resources provide, such as the wind, have not been capitalised upon. The intermittent nature of the wind make it difficult to maximise its potential as in many cases the highest wind speeds are highest when demand is low. The West of Ireland’s combination of wind speeds and unique topography makes it suitable for and innovative wind powered pumped storage system, which can essentially regulate the wind generated electricity and integrate further penetration of renewable energy. In addition, its location along the Atlantic Ocean provides further scope for innovation as seawater can be integrated into the system design. The construction of such an unprecedented project in combination with increased interconnectors has the potential to make Ireland a rechargeable battery for Europe. However, such ambitious plans are at the very early stages and are in direct contrast to current events in the Irish energy market. This study focuses on the feasibility of West of Ireland pumped storage systems. Entailed within this is an extensive desk study, a detailed site selection process and a feasibility study of grid connection. To increase opportunities to identify the best possible site, the feasibility study was focused on the Galway and Mayo areas solely.
Resumo:
The purpose of this study was to evaluate the determinism of the AS-lnterface network and the 3 main families of control systems, which may use it, namely PLC, PC and RTOS. During the course of this study the PROFIBUS and Ethernet field level networks were also considered in order to ensure that they would not introduce unacceptable latencies into the overall control system. This research demonstrated that an incorrectly configured Ethernet network introduces unacceptable variable duration latencies into the control system, thus care must be exercised if the determinism of a control system is not to be compromised. This study introduces a new concept of using statistics and process capability metrics in the form of CPk values, to specify how suitable a control system is for a given control task. The PLC systems, which were tested, demonstrated extremely deterministic responses, but when a large number of iterations were introduced in the user program, the mean control system latency was much too great for an AS-I network. Thus the PLC was found to be unsuitable for an AS-I network if a large, complex user program Is required. The PC systems, which were tested were non-deterministic and had latencies of variable duration. These latencies became extremely exaggerated when a graphing ActiveX was included in the control application. These PC systems also exhibited a non-normal frequency distribution of control system latencies, and as such are unsuitable for implementation with an AS-I network. The RTOS system, which was tested, overcame the problems identified with the PLC systems and produced an extremely deterministic response, even when a large number of iterations were introduced in the user program. The RTOS system, which was tested, is capable of providing a suitable deterministic control system response, even when an extremely large, complex user program is required.
Resumo:
Magdeburg, Univ., Fak. für Informatik, Diss., 2011
Resumo:
Magdeburg, Univ., Fak. für Elektrotechnik und Informationstechnik, Diss., 2009
Resumo:
El Port d'Informació Científica és un centre de Computació Grid de referència que dona suport a comunitats científiques, com el LHC (CERN). Al PIC, trobem una gran varietat de tecnologies que proporcionen serveis al centre. Des de l'arquitectura i elements de la xarxa, fins a recursos informàtics de computació, sistemes d'emmagatzematge a disc i cinta magnètica, bases de dades (ORACLE/PostgreSQL). El projecte consisteix en el disseny i implementació d'una base de dades col·lectora de tota la informació rellevant dels diferents sistemes del centre, i un portal web on mostrar tots els valors i gràfiques, tot basat en programari lliure.
Resumo:
Activation of the hepatoportal glucose sensors by portal glucose infusion leads to increased glucose clearance and induction of hypoglycemia. Here, we investigated whether glucagon-like peptide-1 (GLP-1) could modulate the activity of these sensors. Mice were therefore infused with saline (S-mice) or glucose (P-mice) through the portal vein at a rate of 25 mg/kg. min. In P-mice, glucose clearance increased to 67.5 +/- 3.7 mg/kg. min as compared with 24.1 +/- 1.5 mg/kg. min in S-mice, and glycemia decreased from 5.0 +/- 0.1 to 3.3 +/- 0.1 mmol/l at the end of the 3-h infusion period. Coinfusion of GLP-1 with glucose into the portal vein at a rate of 5 pmol/kg. min (P-GLP-1 mice) did not increase the glucose clearance rate (57.4 +/- 5.0 ml/kg. min) and hypoglycemia (3.8 +/- 0.1 mmol/l) observed in P-mice. In contrast, coinfusion of glucose and the GLP-1 receptor antagonist exendin-(9-39) into the portal vein at a rate of 0.5 pmol/kg. min (P-Ex mice) reduced glucose clearance to 36.1 +/- 2.6 ml/kg. min and transiently increased glycemia to 9.2 +/- 0.3 mmol/l at 60 min of infusion before it returned to the fasting level (5.6 +/- 0.3 mmol/l) at 3 h. When glucose and exendin-(9-39) were infused through the portal and femoral veins, respectively, glucose clearance increased to 70.0 +/- 4.6 ml/kg. min and glycemia decreased to 3.1 +/- 0.1 mmol/l, indicating that exendin-(9-39) has an effect only when infused into the portal vein. Finally, portal vein infusion of glucose in GLP-1 receptor(-/-) mice failed to increase the glucose clearance rate (26.7 +/- 2.9 ml/kg. min). Glycemia increased to 8.5 +/- 0.5 mmol/l at 60 min and remained elevated until the end of the glucose infusion (8.2 +/- 0.4 mmol/l). Together, our data show that the GLP-1 receptor is part of the hepatoportal glucose sensor and that basal fasting levels of GLP-1 sufficiently activate the receptor to confer maximum glucose competence to the sensor. These data demonstrate an important extrapancreatic effect of GLP-1 in the control of glucose homeostasis.
Resumo:
BACKGROUND: DNA sequence integrity, mRNA concentrations and protein-DNA interactions have been subject to genome-wide analyses based on microarrays with ever increasing efficiency and reliability over the past fifteen years. However, very recently novel technologies for Ultra High-Throughput DNA Sequencing (UHTS) have been harnessed to study these phenomena with unprecedented precision. As a consequence, the extensive bioinformatics environment available for array data management, analysis, interpretation and publication must be extended to include these novel sequencing data types. DESCRIPTION: MIMAS was originally conceived as a simple, convenient and local Microarray Information Management and Annotation System focused on GeneChips for expression profiling studies. MIMAS 3.0 enables users to manage data from high-density oligonucleotide SNP Chips, expression arrays (both 3'UTR and tiling) and promoter arrays, BeadArrays as well as UHTS data using MIAME-compliant standardized vocabulary. Importantly, researchers can export data in MAGE-TAB format and upload them to the EBI's ArrayExpress certified data repository using a one-step procedure. CONCLUSION: We have vastly extended the capability of the system such that it processes the data output of six types of GeneChips (Affymetrix), two different BeadArrays for mRNA and miRNA (Illumina) and the Genome Analyzer (a popular Ultra-High Throughput DNA Sequencer, Illumina), without compromising on its flexibility and user-friendliness. MIMAS, appropriately renamed into Multiomics Information Management and Annotation System, is currently used by scientists working in approximately 50 academic laboratories and genomics platforms in Switzerland and France. MIMAS 3.0 is freely available via http://multiomics.sourceforge.net/.
Resumo:
Abstract : In the subject of fingerprints, the rise of computers tools made it possible to create powerful automated search algorithms. These algorithms allow, inter alia, to compare a fingermark to a fingerprint database and therefore to establish a link between the mark and a known source. With the growth of the capacities of these systems and of data storage, as well as increasing collaboration between police services on the international level, the size of these databases increases. The current challenge for the field of fingerprint identification consists of the growth of these databases, which makes it possible to find impressions that are very similar but coming from distinct fingers. However and simultaneously, this data and these systems allow a description of the variability between different impressions from a same finger and between impressions from different fingers. This statistical description of the withinand between-finger variabilities computed on the basis of minutiae and their relative positions can then be utilized in a statistical approach to interpretation. The computation of a likelihood ratio, employing simultaneously the comparison between the mark and the print of the case, the within-variability of the suspects' finger and the between-variability of the mark with respect to a database, can then be based on representative data. Thus, these data allow an evaluation which may be more detailed than that obtained by the application of rules established long before the advent of these large databases or by the specialists experience. The goal of the present thesis is to evaluate likelihood ratios, computed based on the scores of an automated fingerprint identification system when the source of the tested and compared marks is known. These ratios must support the hypothesis which it is known to be true. Moreover, they should support this hypothesis more and more strongly with the addition of information in the form of additional minutiae. For the modeling of within- and between-variability, the necessary data were defined, and acquired for one finger of a first donor, and two fingers of a second donor. The database used for between-variability includes approximately 600000 inked prints. The minimal number of observations necessary for a robust estimation was determined for the two distributions used. Factors which influence these distributions were also analyzed: the number of minutiae included in the configuration and the configuration as such for both distributions, as well as the finger number and the general pattern for between-variability, and the orientation of the minutiae for within-variability. In the present study, the only factor for which no influence has been shown is the orientation of minutiae The results show that the likelihood ratios resulting from the use of the scores of an AFIS can be used for evaluation. Relatively low rates of likelihood ratios supporting the hypothesis known to be false have been obtained. The maximum rate of likelihood ratios supporting the hypothesis that the two impressions were left by the same finger when the impressions came from different fingers obtained is of 5.2 %, for a configuration of 6 minutiae. When a 7th then an 8th minutia are added, this rate lowers to 3.2 %, then to 0.8 %. In parallel, for these same configurations, the likelihood ratios obtained are on average of the order of 100,1000, and 10000 for 6,7 and 8 minutiae when the two impressions come from the same finger. These likelihood ratios can therefore be an important aid for decision making. Both positive evolutions linked to the addition of minutiae (a drop in the rates of likelihood ratios which can lead to an erroneous decision and an increase in the value of the likelihood ratio) were observed in a systematic way within the framework of the study. Approximations based on 3 scores for within-variability and on 10 scores for between-variability were found, and showed satisfactory results. Résumé : Dans le domaine des empreintes digitales, l'essor des outils informatisés a permis de créer de puissants algorithmes de recherche automatique. Ces algorithmes permettent, entre autres, de comparer une trace à une banque de données d'empreintes digitales de source connue. Ainsi, le lien entre la trace et l'une de ces sources peut être établi. Avec la croissance des capacités de ces systèmes, des potentiels de stockage de données, ainsi qu'avec une collaboration accrue au niveau international entre les services de police, la taille des banques de données augmente. Le défi actuel pour le domaine de l'identification par empreintes digitales consiste en la croissance de ces banques de données, qui peut permettre de trouver des impressions très similaires mais provenant de doigts distincts. Toutefois et simultanément, ces données et ces systèmes permettent une description des variabilités entre différentes appositions d'un même doigt, et entre les appositions de différents doigts, basées sur des larges quantités de données. Cette description statistique de l'intra- et de l'intervariabilité calculée à partir des minuties et de leurs positions relatives va s'insérer dans une approche d'interprétation probabiliste. Le calcul d'un rapport de vraisemblance, qui fait intervenir simultanément la comparaison entre la trace et l'empreinte du cas, ainsi que l'intravariabilité du doigt du suspect et l'intervariabilité de la trace par rapport à une banque de données, peut alors se baser sur des jeux de données représentatifs. Ainsi, ces données permettent d'aboutir à une évaluation beaucoup plus fine que celle obtenue par l'application de règles établies bien avant l'avènement de ces grandes banques ou par la seule expérience du spécialiste. L'objectif de la présente thèse est d'évaluer des rapports de vraisemblance calcul és à partir des scores d'un système automatique lorsqu'on connaît la source des traces testées et comparées. Ces rapports doivent soutenir l'hypothèse dont il est connu qu'elle est vraie. De plus, ils devraient soutenir de plus en plus fortement cette hypothèse avec l'ajout d'information sous la forme de minuties additionnelles. Pour la modélisation de l'intra- et l'intervariabilité, les données nécessaires ont été définies, et acquises pour un doigt d'un premier donneur, et deux doigts d'un second donneur. La banque de données utilisée pour l'intervariabilité inclut environ 600000 empreintes encrées. Le nombre minimal d'observations nécessaire pour une estimation robuste a été déterminé pour les deux distributions utilisées. Des facteurs qui influencent ces distributions ont, par la suite, été analysés: le nombre de minuties inclus dans la configuration et la configuration en tant que telle pour les deux distributions, ainsi que le numéro du doigt et le dessin général pour l'intervariabilité, et la orientation des minuties pour l'intravariabilité. Parmi tous ces facteurs, l'orientation des minuties est le seul dont une influence n'a pas été démontrée dans la présente étude. Les résultats montrent que les rapports de vraisemblance issus de l'utilisation des scores de l'AFIS peuvent être utilisés à des fins évaluatifs. Des taux de rapports de vraisemblance relativement bas soutiennent l'hypothèse que l'on sait fausse. Le taux maximal de rapports de vraisemblance soutenant l'hypothèse que les deux impressions aient été laissées par le même doigt alors qu'en réalité les impressions viennent de doigts différents obtenu est de 5.2%, pour une configuration de 6 minuties. Lorsqu'une 7ème puis une 8ème minutie sont ajoutées, ce taux baisse d'abord à 3.2%, puis à 0.8%. Parallèlement, pour ces mêmes configurations, les rapports de vraisemblance sont en moyenne de l'ordre de 100, 1000, et 10000 pour 6, 7 et 8 minuties lorsque les deux impressions proviennent du même doigt. Ces rapports de vraisemblance peuvent donc apporter un soutien important à la prise de décision. Les deux évolutions positives liées à l'ajout de minuties (baisse des taux qui peuvent amener à une décision erronée et augmentation de la valeur du rapport de vraisemblance) ont été observées de façon systématique dans le cadre de l'étude. Des approximations basées sur 3 scores pour l'intravariabilité et sur 10 scores pour l'intervariabilité ont été trouvées, et ont montré des résultats satisfaisants.
Resumo:
Grid is a hardware and software infrastructure that provides dependable, consistent, pervasive, and inexpensive access to high-end computational resources. Grid enables access to the resources but it does not guarantee any quality of service. Moreover, Grid does not provide performance isolation; job of one user can influence the performance of other user’s job. The other problem with Grid is that the users of Grid belong to scientific community and the jobs require specific and customized software environment. Providing the perfect environment to the user is very difficult in Grid for its dispersed and heterogeneous nature. Though, Cloud computing provide full customization and control, but there is no simple procedure available to submit user jobs as in Grid. The Grid computing can provide customized resources and performance to the user using virtualization. A virtual machine can join the Grid as an execution node. The virtual machine can also be submitted as a job with user jobs inside. Where the first method gives quality of service and performance isolation, the second method also provides customization and administration in addition. In this thesis, a solution is proposed to enable virtual machine reuse which will provide performance isolation with customization and administration. The same virtual machine can be used for several jobs. In the proposed solution customized virtual machines join the Grid pool on user request. Proposed solution describes two scenarios to achieve this goal. In first scenario, user submits their customized virtual machine as a job. The virtual machine joins the Grid pool when it is powered on. In the second scenario, user customized virtual machines are preconfigured in the execution system. These virtual machines join the Grid pool on user request. Condor and VMware server is used to deploy and test the scenarios. Condor supports virtual machine jobs. The scenario 1 is deployed using Condor VM universe. The second scenario uses VMware-VIX API for scripting powering on and powering off of the remote virtual machines. The experimental results shows that as scenario 2 does not need to transfer the virtual machine image, the virtual machine image becomes live on pool more faster. In scenario 1, the virtual machine runs as a condor job, so it easy to administrate the virtual machine. The only pitfall in scenario 1 is the network traffic.
Resumo:
Background: Disease management, a system of coordinated health care interventions for populations with chronic diseases in which patient self-care is a key aspect, has been shown to be effective for several conditions. Little is known on the supply of disease management programs in Switzerland. Objectives: To systematically search, record and evaluate data on existing disease management programs in Switzerland. Methods: Programs met our operational definition of disease management if their interventions targeted a chronic disease, included a multidisciplinary team and lasted at least 6 months. To find existing programs, we searched Swiss official websites, Swiss web-pages using Google, medical electronic database (Medline), and checked references from selected documents. We also contacted personally known individuals, those identified as possibly working in the field, individuals working in major Swiss health insurance companies and people recommended by previously contacted persons (snow ball strategy). We developed an extraction grid and collected information pertaining to the following 8 domains: patient population, intervention recipient, intervention content, delivery personnel, method of communication, intensity and complexity, environment and clinical outcomes (measures?). Results: We identified 8 programs fulfilling our operational definition of disease management. Programs targeted patients with diabetes, hypertension, heart failure, obesity, alcohol dependence, psychiatric disorders or breast cancer, and were mainly directed towards patients. The interventions were multifaceted and included education in almost all cases. Half of the programs included regularly scheduled follow-up, by phone in 3 instances. Healthcare professionals involved were physicians, nurses, case managers, social workers, psychologists and dietitians. None fulfilled the 6 criteria established by the Disease Management Association of America. Conclusions: Our study shows that disease management programs, in a country with universal health insurance coverage and little incentive to develop new healthcare strategies, are scarce, although we may have missed existing programs. Nonetheless, those already implemented are very interesting and rather comprehensive. Appropriate evaluation of these programs should be performed in order to build upon them and try to design a generic disease management framework suited to the Swiss healthcare system.
Resumo:
Background. A software based tool has been developed (Optem) to allow automatize the recommendations of the Canadian Multiple Sclerosis Working Group for optimizing MS treatment in order to avoid subjective interpretation. METHODS: Treatment Optimization Recommendations (TORs) were applied to our database of patients treated with IFN beta1a IM. Patient data were assessed during year 1 for disease activity, and patients were assigned to 2 groups according to TOR: "change treatment" (CH) and "no change treatment" (NCH). These assessments were then compared to observed clinical outcomes for disease activity over the following years. RESULTS: We have data on 55 patients. The "change treatment" status was assigned to 22 patients, and "no change treatment" to 33 patients. The estimated sensitivity and specificity according to last visit status were 73.9% and 84.4%. During the following years, the Relapse Rate was always higher in the "change treatment" group than in the "no change treatment" group (5 y; CH: 0.7, NCH: 0.07; p < 0.001, 12 m - last visit; CH: 0.536, NCH: 0.34). We obtained the same results with the EDSS (4 y; CH: 3.53, NCH: 2.55, annual progression rate in 12 m - last visit; CH: 0.29, NCH: 0.13). CONCLUSION: Applying TOR at the first year of therapy allowed accurate prediction of continued disease activity in relapses and disability progression.
Resumo:
We present a computer vision system that associates omnidirectional vision with structured light with the aim of obtaining depth information for a 360 degrees field of view. The approach proposed in this article combines an omnidirectional camera with a panoramic laser projector. The article shows how the sensor is modelled and its accuracy is proved by means of experimental results. The proposed sensor provides useful information for robot navigation applications, pipe inspection, 3D scene modelling etc
Resumo:
Path planning and control strategies applied to autonomous mobile robots should fulfil safety rules as well as achieve final goals. Trajectory planning applications should be fast and flexible to allow real time implementations as well as environment interactions. The methodology presented uses the on robot information as the meaningful data necessary to plan a narrow passage by using a corridor based on attraction potential fields that approaches the mobile robot to the final desired configuration. It employs local and dense occupancy grid perception to avoid collisions. The key goals of this research project are computational simplicity as well as the possibility of integrating this method with other methods reported by the research community. Another important aspect of this work consist in testing the proposed method by using a mobile robot with a perception system composed of a monocular camera and odometers placed on the two wheels of the differential driven motion system. Hence, visual data are used as a local horizon of perception in which trajectories without collisions are computed by satisfying final goal approaches and safety criteria
Resumo:
Colour image segmentation based on the hue component presents some problems due to the physical process of image formation. One of that problems is colour clipping, which appear when at least one of the sensor components is saturated. We have designed a system, that works for a trained set of colours, to recover the chromatic information of those pixels on which colour has been clipped. The chromatic correction method is based on the fact that hue and saturation are invariant to the uniform scaling of the three RGB components. The proposed method has been validated by means of a specific colour image processing board that has allowed its execution in real time. We show experimental results of the application of our method