931 resultados para Database schema
Resumo:
Poster at Open Repositories 2014, Helsinki, Finland, June 9-13, 2014
Resumo:
Large-headed total hip arthroplasty (THA) and hip resurfacing arthroplasty (HRA) with metal-on-metal (MoM) bearings became popular during the last decade. Recently, it has become evident that the large-head MoM hip implants are associated with increased revision rates despite their theoretical advantages. The purpose of this study was to evaluate the early results of primary MoM hip replacements and of acetabular revisions. I analyzed retrospectively the results of four MoM implant designs and the survival rate of acetabular revisions with impaction bone grafting, as documented in the Turku University Hospital database. Further, I evaluated the correlation between femoral head size and dislocation rate, and used the Finnish Arthroplasty Register data to compare the survival of three large-head MoM THAs to analogous HRAs. The early results for the Magnum M2A–ReCap THA were good. A larger head size decreased the risk of dislocation. Articular surface replacement (ASR) THA yielded inferior results compared to analogous HRA. For two other designs the results were similar. The R3–Synergy THA yielded inferior results compared to the reference implants. The survival of acetabular reconstructions with impaction bone grafting was inferior compared to previous reports. In conclusion, the early results of the Biomet ReCap–Magnum design were promising, and large head sizes decreased the dislocation rate. The survival of different MoM hip implant designs varied. The survival of new designs and techniques may be inferior to those reported by the clinics where implants are developed. An important caveat is that early promising results of new devices may rapidly worsen. New implants need to be introduced in a controlled fashion to the market; here, arthroplasty registers are a valuable tool that needs to be used.
Resumo:
Data management consists of collecting, storing, and processing the data into the format which provides value-adding information for decision-making process. The development of data management has enabled of designing increasingly effective database management systems to support business needs. Therefore as well as advanced systems are designed for reporting purposes, also operational systems allow reporting and data analyzing. The used research method in the theory part is qualitative research and the research type in the empirical part is case study. Objective of this paper is to examine database management system requirements from reporting managements and data managements perspectives. In the theory part these requirements are identified and the appropriateness of the relational data model is evaluated. In addition key performance indicators applied to the operational monitoring of production are studied. The study has revealed that the appropriate operational key performance indicators of production takes into account time, quality, flexibility and cost aspects. Especially manufacturing efficiency has been highlighted. In this paper, reporting management is defined as a continuous monitoring of given performance measures. According to the literature review, the data management tool should cover performance, usability, reliability, scalability, and data privacy aspects in order to fulfill reporting managements demands. A framework is created for the system development phase based on requirements, and is used in the empirical part of the thesis where such a system is designed and created for reporting management purposes for a company which operates in the manufacturing industry. Relational data modeling and database architectures are utilized when the system is built for relational database platform.
Resumo:
1:1000000.
Resumo:
Ontology matching is an important task when data from multiple data sources is integrated. Problems of ontology matching have been studied widely in the researchliterature and many different solutions and approaches have been proposed alsoin commercial software tools. In this survey, well-known approaches of ontologymatching, and its subtype schema matching, are reviewed and compared. The aimof this report is to summarize the knowledge about the state-of-the-art solutionsfrom the research literature, discuss how the methods work on different application domains, and analyze pros and cons of different open source and academic tools inthe commercial world.
Resumo:
With the growth in new technologies, using online tools have become an everyday lifestyle. It has a greater impact on researchers as the data obtained from various experiments needs to be analyzed and knowledge of programming has become mandatory even for pure biologists. Hence, VTT came up with a new tool, R Executables (REX) which is a web application designed to provide a graphical interface for biological data functions like Image analysis, Gene expression data analysis, plotting, disease and control studies etc., which employs R functions to provide results. REX provides a user interactive application for the biologists to directly enter the values and run the required analysis with a single click. The program processes the given data in the background and prints results rapidly. Due to growth of data and load on server, the interface has gained problems concerning time consumption, poor GUI, data storage issues, security, minimal user interactive experience and crashes with large amount of data. This thesis handles the methods by which these problems were resolved and made REX a better application for the future. The old REX was developed using Python Django and now, a new programming language, Vaadin has been implemented. Vaadin is a Java framework for developing web applications and the programming language is extremely similar to Java with new rich components. Vaadin provides better security, better speed, good and interactive interface. In this thesis, subset functionalities of REX was selected which includes IST bulk plotting and image segmentation and implemented those using Vaadin. A code of 662 lines was programmed by me which included Vaadin as the front-end handler while R language was used for back-end data retrieval, computing and plotting. The application is optimized to allow further functionalities to be migrated with ease from old REX. Future development is focused on including Hight throughput screening functions along with gene expression database handling
Resumo:
Music archives and composition manuscripts from the Viola database.
Resumo:
A cognitively based instructional program for narrative writing was developed. The effects of using cognitively based schematic planning organizers at the pre-writing stage were evaluated using subjects from the Primary, Junior and Intermediate divisions. Results indicate that the use of organizers based on problem solving significantly improved the organization and the overall quality of narrative writing for students in grades 3, 6 and 7. The magnitude of the improvement of the treatment group over the control group performance in Organization ranged from 10.7% to 22.9%. Statistical and observational data indicate many implications for further research into the cognitive basis for writing and reading; for the improvement and evaluation of school writing programs; for the design of school curricula; and for the inservice education for teachers of writing.
Resumo:
A review of the literature reveals that there are a number of children in the educational system who are characterized by Attention Deficit Disorder. Further review of the literature reveals that there are information processing programs which have had some success in increasing the learning of these children. Currently, an information processing program which is based on schema theory is being implemented in Lincoln County. Since schema theory based programs build structural, conditional, factual, and procedural schemata which assist the learner in attending to salient factors, learning should be increased. Thirty-four children were selected from a random sampling of Grade Seven classes in Lincoln County. Seventeen of these children were identified by the researcher and classroom teacher as being characterized by Attention Deficit Disorder. From the remaining population, 17 children who were not characterized by Attention Deficit Disorder were randomly selected. The data collected were compared using independent t-tests, paired t-tests, and correlation analysis. Significant differences were found in all cases. The Non-Attention Deficit Disorder children scored significantly higher on all the tests but the Attention Defici t Disorder children had a significantly higher ratio of gain between the pretests and posttests.
Resumo:
Rough Set Data Analysis (RSDA) is a non-invasive data analysis approach that solely relies on the data to find patterns and decision rules. Despite its noninvasive approach and ability to generate human readable rules, classical RSDA has not been successfully used in commercial data mining and rule generating engines. The reason is its scalability. Classical RSDA slows down a great deal with the larger data sets and takes much longer times to generate the rules. This research is aimed to address the issue of scalability in rough sets by improving the performance of the attribute reduction step of the classical RSDA - which is the root cause of its slow performance. We propose to move the entire attribute reduction process into the database. We defined a new schema to store the initial data set. We then defined SOL queries on this new schema to find the attribute reducts correctly and faster than the traditional RSDA approach. We tested our technique on two typical data sets and compared our results with the traditional RSDA approach for attribute reduction. In the end we also highlighted some of the issues with our proposed approach which could lead to future research.
Resumo:
The 19th Century Tombstone Database project was funded by the program Federal Summer Youth Employment scheme in the summer of 1982 and led by Dr. David W. Rupp, a Professor at the Classics Department, Brock University. The main goal of the project was to collect information related to various cemeteries in Niagara region and burials that took place from 1790-1890. Data was collected and presented in the form of data summary forms of persons, tombstone sketches, photographs of tombstones, maps, and computer printouts. The materials created as a result of a research completed for the 19th Century Tombstone Database project are important as a number of the tombstones have been damaged or gone missing since the research was finished. Before Dr. Rupp retired from Brock University, he donated project materials to the Brock University Special Collections and Archives.
Resumo:
A qualitative research study that asked international students how they thought of words to enter into a library database to see if language learning was also involved.
Resumo:
Introduction: Avec l’abondance d’information gratuite disponible en ligne, la tâche de trouver, de trier et d’acheminer de l’information pertinente à l’auditoire approprié peut s’avérer laborieuse. En décembre 2010, la Bibliothèque virtuelle canadienne de santé / Canadian Virtual Health Library (BVCS) a formé un comité d’experts afin d’identifier, d’évaluer, de sélectionner et d’organiser des ressources d’intérêt pour les professionnels de la santé. Méthodes: Cette affiche identifiera les décisions techniques du comité d’experts, incluant le système de gestion de contenus retenu, l’utilisation des éléments Dublin Core et des descripteurs Medical Subject Headings pour la description des ressources, et le développement et l’adaptation de taxonomies à partir de la classification MeSH. La traduction française des descripteurs MeSH à l’aide du portail CISMeF sera également abordée. Résultats: Au mois de mai 2011, le comité a lancé la base de données BVCS de ressources en ligne gratuites sur la santé, regroupant plus de 1600 sites web et ressources. Une variété de types de contenus sont représentés, incluant des articles et rapports, des bases de données interactives et des outils de pratique clinique. Discussion: Les bénéfices et défis d’une collaboration pancanadienne virtuelle seront présentés, ainsi que l’inclusion cruciale d’un membre francophone pour composer avec la nature bilingue de la base de données. En lien avec cet aspect du projet, l’affiche sera présentée en français et en anglais. Introduction: With the abundance of freely available online information, the task of finding, filtering and fitting relevant information to the appropriate audience, is daunting. In December 2010 the Canadian Virtual Health Library / Bibliothèque virtuelle canadienne de santé (CVHL) formed an expert committee to identify, evaluate, select and organize resources relevant to health professionals. Methods: This poster will identify the key technical decisions of the expert committee including the content management system used to manage the data, the use of Dublin Core elements and Medical Subject Headings to describe the resources, and the development and adaptation of taxonomies from MeSH classification to catalog resources. The translation of MeSH terms to French using the CiSMeF portal will also be discussed. Results: In May 2010, the committee launched the CVHL database of free web-based health resources. Content ranged from online articles and reports to videos, interactive databases and clinical practice tools, and included more than 1,600 websites and resources. Discussion: The benefits and challenges of a virtual, pan-Canadian collaboration, and the critical inclusion of a Francophone member to address the bilingual nature of the database, will be presented. In keeping with the nature of the project, the poster will be presented in French and English.