817 resultados para web based intervention
Resumo:
Background: Hypertension and Diabetes is a public health and economic concern in the United States. The utilization of medical home concepts increases the receipt of preventive services, however, do they also increase adherence to treatments? This study examined the effect of patient-centered medical home technologies such as the electronic health record, clinical support system, and web-based care management in improving health outcomes related to hypertension and diabetes. Methods: A systematic review of the literature used a best evidence synthesis approach to address the general question " Do patient-centered medical home technologies have an effect of diabetes and hypertension treatment?" This was followed by an evaluation of specific examples of the technologies utilized such as computer-assisted recommendations and web-based care management provided by the patient's electronic health record. Ebsco host, Ovid host, and Google Scholar were the databases used to conduct the literature search. Results: The initial search identified over 25 studies based on content and quality that implemented technology interventions to improve communication between provider and patient. After further assessing the articles for risk of bias and study design, 13 randomized controlled studies were chosen. All of the studies chosen were conducted in various primary care settings in both private practices and hospitals between the years 2000 and 2007. The sample sizes of the studies ranged from 42 to 2924 participants. The mean age for all of the studies ranged from 56 to 71 years. The percent women in the studies ranged from one to 78 percent. Over one-third of the studies did not provide the racial composition of the participants. For the seven studies that did provide information about the ethnic composition, 64% of the intervention participants were White. All of the studies utilized some type of web-based or computer-based communication to manage hypertension or diabetes care. Findings on outcomes were mixed, with nine out of 13 studies showing no significant effect on outcomes examined, and four of the studies showing significant and positive impact on health outcomes related to hypertension or diabetes Conclusion: Although the technologies improved patient and provider satisfaction, the outcomes measures such as blood pressure control and glucose control were inconclusive. Further research is needed with diverse ethnic and SES population to investigate the role of patient-centered technologies on hypertension and diabetes control. Also, further research is needed to investigate the effects of innovative medical home technologies that can be used by both patients and providers to increase quality of communication concerning adherence to treatments.^
Resumo:
Specialized search engines such as PubMed, MedScape or Cochrane have increased dramatically the visibility of biomedical scientific results. These web-based tools allow physicians to access scientific papers instantly. However, this decisive improvement had not a proportional impact in clinical practice due to the lack of advanced search methods. Even queries highly specified for a concrete pathology frequently retrieve too many information, with publications related to patients treated by the physician beyond the scope of the results examined. In this work we present a new method to improve scientific article search using patient information. Two pathologies have been used within the project to retrieve relevant literature to patient data and to be integrated with other sources. Promising results suggest the suitability of the approach, highlighting publications dealing with patient features and facilitating literature search to physicians.
Resumo:
The uptake of Linked Data (LD) has promoted the proliferation of datasets and their associated ontologies for describing different domains. Par-ticular LD development characteristics such as agility and web-based architec-ture necessitate the revision, adaption, and lightening of existing methodologies for ontology development. This thesis proposes a lightweight method for ontol-ogy development in an LD context which will be based in data-driven agile de-velopments, existing resources to be reused, and the evaluation of the obtained products considering both classical ontological engineering principles and LD characteristics.
Resumo:
Lately, videoconference applications have experienced an evolution towards the World Wide Web. New technologies have given browsers real-time communications capabilities. In this context, WebRTC aims to provide this functionality by following and defining standards. Being a new effort, WebRTC still lacks advanced videoconferencing services such as session recording, media mixing and adjusting to varying network conditions. This paper analyzes these challenges and proposes an architecture based on a traditional communications entity, the Multipoint Control Unit or MCU as a solution.
Resumo:
P2P applications are increasingly present on the web. We have identified a gap in current proposals when it comes to the use of traditional P2P overlays for real-time multimedia streaming. We analyze the possibilities and challenges to extend WebRTC in order to implement JavaScript APIs for P2P streaming algorithms.
Resumo:
Aiming to address requirements concerning integration of services in the context of ?big data?, this paper presents an innovative approach that (i) ensures a flexible, adaptable and scalable information and computation infrastructure, and (ii) exploits the competences of stakeholders and information workers to meaningfully confront information management issues such as information characterization, classification and interpretation, thus incorporating the underlying collective intelligence. Our approach pays much attention to the issues of usability and ease-of-use, not requiring any particular programming expertise from the end users. We report on a series of technical issues concerning the desired flexibility of the proposed integration framework and we provide related recommendations to developers of such solutions. Evaluation results are also discussed.
Resumo:
This paper reports on an innovative approach that aims to reduce information management costs in data-intensive and cognitively-complex biomedical environments. Recognizing the importance of prominent high-performance computing paradigms and large data processing technologies as well as collaboration support systems to remedy data-intensive issues, it adopts a hybrid approach by building on the synergy of these technologies. The proposed approach provides innovative Web-based workbenches that integrate and orchestrate a set of interoperable services that reduce the data-intensiveness and complexity overload at critical decision points to a manageable level, thus permitting stakeholders to be more productive and concentrate on creative activities.
Resumo:
Este Proyecto Fin de Grado tiene como objetivo fundamental el perfeccionamiento y puesta en explotación de un sistema de ayuda a la decisión que evalúa el desarrollo del lenguaje en niños de 0 a 6 años de edad. Este sistema está formado fundamentalmente por una aplicación diseñada y construida mediante una arquitectura de componentes de software modular y reutilizable. La aplicación será usada por los pediatras para realizar evaluaciones del desarrollo del lenguaje infantil y además por los neuropediatras, logopedas y miembros de equipos de Atención Temprana para consultar las evaluaciones y validar las decisiones propuestas por el sistema. El sistema es accesible vía web y almacena toda la información que maneja en una base de datos. Asimismo, el sistema se apoya en un modelo conceptual u ontología desarrollado previamente para inferir las decisiones adecuadas para las evaluaciones del lenguaje. El sistema incorpora las funciones de gestión de los usuarios del mismo. ABSTRACT This Grade End Project has as fundamental objective the improvement and deployment of a decision support system for evaluating children language development from 0 to 6 years of age. This system is mainly formed by an application designed and built using a modular and reusable software component architecture. The application will be used by pediatricians for evaluating children´s speech development and also by neuro-pediatricians, speech therapists and early childhood intervention team members, for consulting previous evaluations and for validating system´s proposed decision. The system is web based and stores its information in a database. Likewise, the system is supported by a conceptual model or ontology previously developed to infer the appropriate decision for language evaluation. The system also includes user management functions.
Resumo:
This paper argues about the utility of advanced knowledge-based techniques to develop web-based applications that help consumers in finding products within marketplaces in e-commerce. In particular, we describe the idea of model-based approach to develop a shopping agent that dynamically configures a product according to the needs and preferences of customers. Finally, the paper summarizes the advantages provided by this approach.
Resumo:
Carbon (C) and nitrogen (N) process-based models are important tools for estimating and reporting greenhouse gas emissions and changes in soil C stocks. There is a need for continuous evaluation, development and adaptation of these models to improve scientific understanding, national inventories and assessment of mitigation options across the world. To date, much of the information needed to describe different processes like transpiration, photosynthesis, plant growth and maintenance, above and below ground carbon dynamics, decomposition and nitrogen mineralization. In ecosystem models remains inaccessible to the wider community, being stored within model computer source code, or held internally by modelling teams. Here we describe the Global Research Alliance Modelling Platform (GRAMP), a web-based modelling platform to link researchers with appropriate datasets, models and training material. It will provide access to model source code and an interactive platform for researchers to form a consensus on existing methods, and to synthesize new ideas, which will help to advance progress in this area. The platform will eventually support a variety of models, but to trial the platform and test the architecture and functionality, it was piloted with variants of the DNDC model. The intention is to form a worldwide collaborative network (a virtual laboratory) via an interactive website with access to models and best practice guidelines; appropriate datasets for testing, calibrating and evaluating models; on-line tutorials and links to modelling and data provider research groups, and their associated publications. A graphical user interface has been designed to view the model development tree and access all of the above functions.
Resumo:
Systematic evaluation of Learning Objects is essential to make high quality Web-based education possible. For this reason, several educational repositories and e-Learning systems have developed their own evaluation models and tools. However, the differences of the context in which Learning Objects are produced and consumed suggest that no single evaluation model is sufficient for all scenarios. Besides, no much effort has been put in developing open tools to facilitate Learning Object evaluation and use the quality information for the benefit of end users. This paper presents LOEP, an open source web platform that aims to facilitate Learning Object evaluation in different scenarios and educational settings by supporting and integrating several evaluation models and quality metrics. The work exposed in this paper shows that LOEP is capable of providing Learning Object evaluation to e-Learning systems in an open, low cost, reliable and effective way. Possible scenarios where LOEP could be used to implement quality control policies and to enhance search engines are also described. Finally, we report the results of a survey conducted among reviewers that used LOEP, showing that they perceived LOEP as a powerful and easy to use tool for evaluating Learning Objects.
Resumo:
Evaluating and measuring the pedagogical quality of Learning Objects is essential for achieving a successful web-based education. On one hand, teachers need some assurance of quality of the teaching resources before making them part of the curriculum. On the other hand, Learning Object Repositories need to include quality information into the ranking metrics used by the search engines in order to save users time when searching. For these reasons, several models such as LORI (Learning Object Review Instrument) have been proposed to evaluate Learning Object quality from a pedagogical perspective. However, no much effort has been put in defining and evaluating quality metrics based on those models. This paper proposes and evaluates a set of pedagogical quality metrics based on LORI. The work exposed in this paper shows that these metrics can be effectively and reliably used to provide quality-based sorting of search results. Besides, it strongly evidences that the evaluation of Learning Objects from a pedagogical perspective can notably enhance Learning Object search if suitable evaluations models and quality metrics are used. An evaluation of the LORI model is also described. Finally, all the presented metrics are compared and a discussion on their weaknesses and strengths is provided.
Resumo:
Language resources, such as multilingual lexica and multilingual electronic dictionaries, contain collections of lexical entries in several languages. Having access to the corresponding explicit or implicit translation relations between such entries might be of great interest for many NLP-based applications. By using Semantic Web-based techniques, translations can be available on the Web to be consumed by other (semantic enabled) resources in a direct manner, not relying on application-specific formats. To that end, in this paper we propose a model for representing translations as linked data, as an extension of the lemon model. Our translation module represents some core information associated to term translations and does not commit to specific views or translation theories. As a proof of concept, we have extracted the translations of the terms contained in Terminesp, a multilingual terminological database, and represented them as linked data. We have made them accessible on the Web both for humans (via a Web interface) and software agents (with a SPARQL endpoint).
MEDLINEplus: building and maintaining the National Library of Medicine's consumer health Web service
Resumo:
MEDLINEplus is a Web-based consumer health information resource, made available by the National Library of Medicine (NLM). MEDLINEplus has been designed to provide consumers with a well-organized, selective Web site facilitating access to reliable full-text health information. In addition to full-text resources, MEDLINEplus directs consumers to dictionaries, organizations, directories, libraries, and clearinghouses for answers to health questions. For each health topic, MEDLINEplus includes a preformulated MEDLINE search created by librarians. The site has been designed to match consumer language to medical terminology. NLM has used advances in database and Web technologies to build and maintain MEDLINEplus, allowing health sciences librarians to contribute remotely to the resource. This article describes the development and implementation of MEDLINEplus, its supporting technology, and plans for future development.
Resumo:
OBJETIVO: Analizar la opinión que los usuarios tienen sobre alimentos genéticamente modificados y su información en el etiquetado. MÉTODOS: Realizada revisión sistemática de la literatura científica sobre los alimentos transgénicos y el etiquetado a partir de la consulta de las bases de datos bibliográficas: Medline (vía PubMed), EMBASE, ISIWeb of Knowledge, Cochrane Library Plus, FSTA, LILACS, CINAHL y AGRICOLA. Los descriptores seleccionados fueron: «organisms, genetically modified» y «food labeling». La búsqueda se realizó desde la primera fecha disponible hasta junio de 2012, seleccionando los artículos pertinentes escritos en inglés, portugués y castellano. RESULTADOS: Se seleccionaron 40 artículos. En todos ellos, se debía haber realizado una intervención poblacional enfocada al conocimiento de los consumidores sobre los alimentos genéticamente modificados y su necesidad, o no, de incluir información en el etiquetado. El consumidor expresa su preferencia por el producto no-genéticamente modificado, y apunta que está dispuesto a pagar algo más por él, pero, en definitiva compra el artículo que está a mejor precio en un mercado que acoge las nuevas tecnologías. En 18 artículos la población se mostraba favorable a su etiquetado obligatorio y seis al etiquetado voluntario; siete trabajos demostraban el poco conocimiento de la población sobre los transgénicos y, en tres, la población subestimó la cantidad que consumía. En todo caso, se observó la influencia del precio del producto genéticamente modificado. CONCLUSIONES: La etiqueta debe ser homogénea y aclarar el grado de tolerancia en humanos de alimentos genéticamente modificados en comparación con los no modificados. Asimismo, debe dejar claro su composición, o no, de alimento genéticamente modificado y la forma de producción de estos artículos de consumo. La etiqueta también debe ir acompañada de un sello de certificación de una agencia del estado y datos para contacto. El consumidor expresa su preferencia por el producto no-genéticamente modificado pero señaló que acaba comprando el artículo que está a mejor precio en un mercado que acoge las nuevas tecnologías.