19 resultados para Database application, Biologia cellulare, Image retrieval
em Universidad Politécnica de Madrid
Resumo:
ImageCLEF is a pilot experiment run at CLEF 2003 for cross language image retrieval using textual captions related to image contents. In this paper, we describe the participation of the MIRACLE research team (Multilingual Information RetrievAl at CLEF), detailing the different experiments and discussing their preliminary results.
Resumo:
The emergence of cloud datacenters enhances the capability of online data storage. Since massive data is stored in datacenters, it is necessary to effectively locate and access interest data in such a distributed system. However, traditional search techniques only allow users to search images over exact-match keywords through a centralized index. These techniques cannot satisfy the requirements of content based image retrieval (CBIR). In this paper, we propose a scalable image retrieval framework which can efficiently support content similarity search and semantic search in the distributed environment. Its key idea is to integrate image feature vectors into distributed hash tables (DHTs) by exploiting the property of locality sensitive hashing (LSH). Thus, images with similar content are most likely gathered into the same node without the knowledge of any global information. For searching semantically close images, the relevance feedback is adopted in our system to overcome the gap between low-level features and high-level features. We show that our approach yields high recall rate with good load balance and only requires a few number of hops.
Resumo:
Digital image correlation (DIC) is applied to analyzing the deformation mechanisms under transverse compression in a fiber-reinforced composite. To this end, compression tests in a direction perpendicular to the fibers were carried out inside a scanning electron microscope and secondary electron images obtained at different magnifications during the test. Optimum DIC parameters to resolve the displacement and strain field were computed from numerical simulations of a model composite and they were applied to micrographs obtained at different magnifications (250_, 2000_, and 6000_). It is shown that DIC of low-magnification micrographs was able to capture the long range fluctuations in strain due to the presence of matrix-rich and fiber-rich zones, responsible for the onset of damage. At higher magnification, the strain fields obtained with DIC qualitatively reproduce the non-homogeneous deformation pattern due to the presence of stiff fibers dispersed in a compliant matrix and provide accurate results of the average composite strain. However, comparison with finite element simulations revealed that DIC was not able to accurately capture the average strain in each phase.
Resumo:
Managing large medical image collections is an increasingly demanding important issue in many hospitals and other medical settings. A huge amount of this information is daily generated, which requires robust and agile systems. In this paper we present a distributed multi-agent system capable of managing very large medical image datasets. In this approach, agents extract low-level information from images and store them in a data structure implemented in a relational database. The data structure can also store semantic information related to images and particular regions. A distinctive aspect of our work is that a single image can be divided so that the resultant sub-images can be stored and managed separately by different agents to improve performance in data accessing and processing. The system also offers the possibility of applying some region-based operations and filters on images, facilitating image classification. These operations can be performed directly on data structures in the database.
Resumo:
Moment invariants have been thoroughly studied and repeatedly proposed as one of the most powerful tools for 2D shape identification. In this paper a set of such descriptors is proposed, being the basis functions discontinuous in a finite number of points. The goal of using discontinuous functions is to avoid the Gibbs phenomenon, and therefore to yield a better approximation capability for discontinuous signals, as images. Moreover, the proposed set of moments allows the definition of rotation invariants, being this the other main design concern. Translation and scale invariance are achieved by means of standard image normalization. Tests are conducted to evaluate the behavior of these descriptors in noisy environments, where images are corrupted with Gaussian noise up to different SNR values. Results are compared to those obtained using Zernike moments, showing that the proposed descriptor has the same performance in image retrieval tasks in noisy environments, but demanding much less computational power for every stage in the query chain.
Resumo:
A methodology for downscaling solar irradiation from satellite-derived databases is described using R software. Different packages such as raster, parallel, solaR, gstat, sp and rasterVis are considered in this study for improving solar resource estimation in areas with complex topography, in which downscaling is a very useful tool for reducing inherent deviations in satellite-derived irradiation databases, which lack of high global spatial resolution. A topographical analysis of horizon blocking and sky-view is developed with a digital elevation model to determine what fraction of hourly solar irradiation reaches the Earth's surface. Eventually, kriging with external drift is applied for a better estimation of solar irradiation throughout the region analyzed. This methodology has been implemented as an example within the region of La Rioja in northern Spain, and the mean absolute error found is a striking 25.5% lower than with the original database.
Resumo:
A “Collaborative Agreement” involving the collective participation of our students in their last year of our “Nuclear Engineering Master Degree Programme” for: “the review and capturing of selected spent fuel isotopic assay data sets to be included in the new SFCOMPO database"
Resumo:
Este Trabajo de Fin de Grado (TFG) consiste en el diseño y el desarrollo de una base de datos para almacenar datos de secuenciación genética. Además, también será necesario poder utilizar la herramienta BLAST, que está formada por un conjunto de programas para buscar por similitud y alinear secuencias, con los datos que se encuentran almacenados en dicha base de datos.---ABSTRACT---The aim of this Bachelor’s Thesis is to design and develop a database to store data of genetic sequences. Furthermore, it will be necessary to use BLAST, which is a suite of programs to search and match similarities sequences into a database.
Resumo:
El presente trabajo propone un procedimiento a través del cual un razonador evalúa la información de una base de datos y la clasifica de forma automática en conceptos, relaciones, roles y atributos. Esta clasificación se desarrolla mediante un procedimiento dividido en dos métodos: primero, un Algoritmo de migración el cual genera una ontología con los elementos del esquema relacional de la base de datos. El segundo método es la Estrategia de clasificación de la información, esta consiste en una serie de consultas SPARQL mediante las que se clasifica la información de la base de datos.---ABSTRACT---This paper proposes a method by which a reasoner evaluates information from a database and automatically classifies in concepts, relationships, roles and attributes. This classification is developed through a procedure divided into two methods: first, a migration algorithm which generates an ontology with elements of relational schema database. The second method is the strategy classification of information, this is a series of SPARQL queries through that classified using the information the database.
Resumo:
NoSQL data stores are becoming more and more popular. Graph databases are one of this kind of data stores. In this paper we present an overview of the implementation of snapshot isolation for Neo4j, a very popular graph database.
Resumo:
El objetivo principal de este proyecto es la realización de un sistema, que permita a través de alguna herramienta accesible a cualquier usuario, poder interactuar con una base de datos que contenga un catálogo bien documentado de los objetos disponibles en el Museo Histórico de la Informática (MHI) perteneciente a la Escuela Técnica Superior de Ingenieros Informáticos (ETSIINF) de la Universidad Politécnica de Madrid (UPM). Hasta el momento, no existía inventario alguno, por lo que la contribución del trabajo que aquí se presenta, supone un gran avance en la organización de los fondos del Museo. Ello contribuirá al desarrollo del objetivo principal del MHI, que es la difusión de la historia de la informática, mediante un medio de los más usados hoy en día, internet. El trabajo realizado se presenta a lo largo de 10 capítulos. En los que se muestran, el análisis del problema, los requisitos y las distintas alternativas posibles de solución, así como la solución adoptada y su desarrollo, tanto en el diseño de la base de datos como de sitio Web que hace posible la visualización e interacción de la información. En el primer capítulo, se puede encontrar una breve introducción del proyecto. Se indican los objetivos, la motivación y el alcance del mismo. En el segundo capítulo, se muestran los requisitos del problema, se analizan las tecnologías, herramientas y lenguajes disponibles para diseñar bases de datos, y se propone la elección de una de las tecnologías, teniendo en cuenta las limitaciones del entorno en el cual se va a implantar la solución. En el tercer capítulo, se diseña la solución propuesta para el sistema. Primero se muestra el diseño de bajo nivel, que serán los cimientos y posteriormente se explica el diseño de alto nivel. Finalmente, se introduce el conjunto de pruebas que el sistema tendrá que pasar para garantizar su correcto funcionamiento. El cuarto capítulo, muestra todas las tecnologías, herramientas, lenguajes y plantillas utilizadas para la implementación de la WEB. Mientras que en el capítulo cinco, se pueden ver los resultados de las pruebas realizadas. En el capítulo seis, se evalúan los costes económicos de realización de proyecto y se presenta la agenda de actividades y tareas llevadas a cabo para su desarrollo. El séptimo capítulo, resume las contribuciones técnicas del proyecto tratadas en los capítulos anteriores, así como las conclusiones personales. Mientras que, el capítulo ocho, apunta una serie de trabajos futuros que se podrían realizarse utilizando como base este proyecto. El capítulo nueve contiene las referencias de la información que se han consultado y que se citan en el texto, y el décimo complementa este proceso de información, incluyendo un glosario de términos técnicos. El contenido de la memoria concluye con el manual de usuario para la administración de la base de datos, que se incluye en forma de anexo.---ABSTRACT---The main goal of this project is the development of a system that would allow through some accessible tool for any user to interact with a database that contains a well-documented objects available in the Computer History Museum's (MHI) catalog, which belongs to the School of Computer Engineers (ETSIINF) of the Polytechnic University of Madrid (UPM). So far, there was no inventory, so the contribution of the work presented here, is a breakthrough in the organization of the Museum's collections. This will contribute to the development of the main goal of the MHI, which is the diffusion of computer history, by means of the most used today, internet. The work is presented along 10 chapters. Which show the analysis of the problem, requirements, the different possible solutions and the solution adopted and its development, both in the design of the database and Web site, which enables the visualization and interaction of the information. In the first chapter, a brief introduction of the project is found. Objectives, motivation and scope of the project are specified. In the second chapter, the requirements of the problem are shown. Technologies, tools and languages available to design databases are analysed, and the choice of a technology is proposed, taking into account the limitations of the environment in which it will to implement the solution. In the third chapter, the proposed system solution is designed. First, low-level design, which will be the foundation of the project, is shown, and then the high-level design is explained. Finally, test suite, which the system will have to past to ensure their proper functioning, are introduced. The fourth chapter shows all technologies, tools, languages and templates used to implement the WEB. While in chapter five, the results of the tests are shown. The economic costs of development the project are evaluated in chapter six, and the schedule of activities and tasks carried out for this development are shown. The seventh chapter summarizes the technical contributions of the project discussed in previous chapters, as well as personal conclusions. While the eighth chapter, suggests future works that could be made, based on this project. Ninth chapter contains references to information that have been consulted and cited in the text, and the tenth chapter includes a glossary of technical terms, to complement that process of information. Finally an annex includes a user manual for managing the database.
Resumo:
Desde el inicio de la globalización, el aprendizaje de la lengua inglesa se ha instaurado como una necesidad. Hoy en día, con la adopción del Espacio Europeo de Educación Superior este lenguaje no sólo se impone como un requisito para los estudiantes sino que se exige un nivel B2, lo cual significa un esfuerzo mayor tanto como para el alumno como para el profesor a la hora de hacer de este ejercicio un hábito y lograr la evaluación continua de los mismos. Este proyecto intenta extender las funcionalidades de una aplicación existente llamada Illlab con ejercicios que se adapten al nivel B2 y permitan la interacción entre alumnos durante la realización de estos ejercicios. El objetivo de esta aplicación es el de desarrollar ejercicios extra en la aplicación Illlab que añadan complejidad para el aprendizaje de inglés de un nivel B2 y que además se puedan realizar actividades entre los alumnos. La idea es hacer una aplicación de preguntas y respuestas “multiple choice” con cuatro opciones por pregunta. El fuerte de este juego está en presentar material variado sobre uso de la lengua y además permitir el juego entre varios alumnos. La extensión de ILLLab se plantea como un proyecto para desarrollar interfaces y funcionalidades adicionales en la antigua aplicación. La principal funcionalidad que se añade es un juego de preguntas y respuestas con opciones múltiples para un nivel B2 y las interfaces responden a necesidades de intercambio y manejo de contenido por Internet mediante estándares aceptados en el mundo del aprendizaje digital tales como Common Cartridge o SCORM. Este proyecto simplemente adapta la aplicación para su uso en un entorno de evaluación de actividades en el cual el profesor tiene acceso a las actividades que realizan los alumnos de un curso para su posterior evaluación. Antiguamente ILLLab sólo contenía ejercicios que se llevaban a cabo en el dispositivo móvil por lo que el control de estas actividades no era posible. La mejora se presenta como una interfaz Common Cartridge para el manejo del contenido, una interfaz de comunicación sobre servicios web tipo REST y el manejo de base de datos mediante Hibernate que agrupa una serie de librerías Java para la persistencia de objetos de la base de datos. ABSTRACT. Since the onset of globalization, the learning of the English language has become as a necessity. Today, with the adoption of the European Higher Education Area this language is not only imposed as a requirement for students but a B2 level is required, which means a greater effort both to the student and teacher when it comes to make the learning exercise a habit and achieve continuous evaluation of students. This project aims to extend the functionality of an existing application called Illlab with an exercise that suits the B2 level and allow interaction between students while performing these exercises. The purpose of this application is to develop an additional exercise in the application Illlab that adds complexity for learning English at B2 level and also enables the interaction among students. The main idea is to make an application in multiple choices style with four options. The strength of this game is to present varied material on use of Enlgish and also allow play between two students. ILLLab extension is conceived as a project to develop interfaces and additional functionalities in the old application. The main functionalities added are a game of questions and answers with multiple choices for a B2 level and interfaces that meet information exchange requirements and content management over the Internet using standards adopted in the world of digital learning such as Common Cartridge or SCORM. This project simply adapts the application for its use in an activities evaluation environment in which the teacher has access to the activities performed by students in a course for further evaluation. The former versión of ILLLab contained only exercises that were carried out on the mobile device so that the evaluation of these activities was not possible. The improvement comes as a Common Cartridge interface for content management, a communication interface with REST web services and a database access using Hibernate which groups a number of Java libraries for object persistence in the database.
Resumo:
This paper describes the participation of DAEDALUS at ImageCLEF 2011 Medical Retrieval task. We have focused on multimodal (or mixed) experiments that combine textual and visual retrieval. The main objective of our research has been to evaluate the effect on the medical retrieval process of the existence of an extended corpus that is annotated with the image type, associated to both the image itself and also to its textual description. For this purpose, an image classifier has been developed to tag each document with its class (1st level of the hierarchy: Radiology, Microscopy, Photograph, Graphic, Other) and subclass (2nd level: AN, CT, MR, etc.). For the textual-based experiments, several runs using different semantic expansion techniques have been performed. For the visual-based retrieval, different runs are defined by the corpus used in the retrieval process and the strategy for obtaining the class and/or subclass. The best results are achieved in runs that make use of the image subclass based on the classification of the sample images. Although different multimodal strategies have been submitted, none of them has shown to be able to provide results that are at least comparable to the ones achieved by the textual retrieval alone. We believe that we have been unable to find a metric for the assessment of the relevance of the results provided by the visual and textual processes
Resumo:
This paper presents a study on the effect of blurred images in hand biometrics. Blurred images simulates out-of-focus effects in hand image acquisition, a common consequence of unconstrained, contact-less and platform-free hand biometrics in mobile devices. The proposed biometric system presents a hand image segmentation based on multiscale aggregation, a segmentation method invariant to different changes like noise or blurriness, together with an innovative feature extraction and a template creation, oriented to obtain an invariant performance against blurring effects. The results highlight that the proposed system is invariant to some low degrees of blurriness, requiring an image quality control to detect and correct those images with a high degree of blurriness. The evaluation has considered a synthetic database created based on a publicly available database with 120 individuals. In addition, several biometric techniques could benefit from the approach proposed in this paper, since blurriness is a very common effect in biometric techniques involving image acquisition.
Resumo:
Current development platforms for designing spoken dialog services feature different kinds of strategies to help designers build, test, and deploy their applications. In general, these platforms are made up of several assistants that handle the different design stages (e.g. definition of the dialog flow, prompt and grammar definition, database connection, or to debug and test the running of the application). In spite of all the advances in this area, in general the process of designing spoken-based dialog services is a time consuming task that needs to be accelerated. In this paper we describe a complete development platform that reduces the design time by using different types of acceleration strategies based on using information from the data model structure and database contents, as well as cumulative information obtained throughout the successive steps in the design. Thanks to these accelerations, the interaction with the platform is simplified and the design is reduced, in most cases, to simple confirmations to the “proposals” that the platform automatically provides at each stage. Different kinds of proposals are available to complete the application flow such as the possibility of selecting which information slots should be requested to the user together, predefined templates for common dialogs, the most probable actions that make up each state defined in the flow, different solutions to solve specific speech-modality problems such as the presentation of the lists of retrieved results after querying the backend database. The platform also includes accelerations for creating speech grammars and prompts, and the SQL queries for accessing the database at runtime. Finally, we will describe the setup and results obtained in a simultaneous summative, subjective and objective evaluations with different designers used to test the usability of the proposed accelerations as well as their contribution to reducing the design time and interaction.