862 resultados para Web, Html 5, JavaScript, Dart, Structured Web Programming


Relevância:

50.00% 50.00%

Publicador:

Resumo:

Las bibliotecas universitarias recopilan, de manera rutinaria estadísticas sobre el uso de sus colecciones impresas y de la actividad in situ. Paralelamente y de manera sostenida, han ido incorporando recursos y servicios electrónicos, lo que ha motivado la elaboración de normas internacionales que definen indicadores que permiten medir su uso, no obstante contar con un software estándar es aún un asunto pendiente. Por otro lado, para medir la actividad de un sitio web existen varios programas gratuitos y de código abierto. Este trabajo tiene como objetivo determinar si los softwares de analítica web gratuitos para sitios web AWStats, Google Analytics y Piwik, pueden utilizarse para evaluar el uso de recursos y servicios electrónicos, conforme a los indicadores propuestos por las normas ANSI/NISO Z39.7-2013, ISO 2789:2003, ISO 20983:2003, BS ISO 11620:2008, EMIS, Counter e ICOLC. Para tales efectos, fueron utilizados para realizar el análisis de esta investigación sitio web y el catálogo en línea de la Biblioteca Florentino Ameghino, Biblioteca Central de la Facultad de Ciencias Naturales y Museo de la Universidad Nacional de la Plata, Argentina. Los resultados reflejan las características de los indicadores, el software y el caso de estudio. Estas características son abordadas en las conclusiones con el fin de darle contexto y perspectiva a la respuesta de la pregunta de si es viable medir el uso de recursos y servicios electrónicos de una biblioteca universitaria por medio de programas estadísticos para sitios web

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Las bibliotecas universitarias recopilan, de manera rutinaria estadísticas sobre el uso de sus colecciones impresas y de la actividad in situ. Paralelamente y de manera sostenida, han ido incorporando recursos y servicios electrónicos, lo que ha motivado la elaboración de normas internacionales que definen indicadores que permiten medir su uso, no obstante contar con un software estándar es aún un asunto pendiente. Por otro lado, para medir la actividad de un sitio web existen varios programas gratuitos y de código abierto. Este trabajo tiene como objetivo determinar si los softwares de analítica web gratuitos para sitios web AWStats, Google Analytics y Piwik, pueden utilizarse para evaluar el uso de recursos y servicios electrónicos, conforme a los indicadores propuestos por las normas ANSI/NISO Z39.7-2013, ISO 2789:2003, ISO 20983:2003, BS ISO 11620:2008, EMIS, Counter e ICOLC. Para tales efectos, fueron utilizados para realizar el análisis de esta investigación sitio web y el catálogo en línea de la Biblioteca Florentino Ameghino, Biblioteca Central de la Facultad de Ciencias Naturales y Museo de la Universidad Nacional de la Plata, Argentina. Los resultados reflejan las características de los indicadores, el software y el caso de estudio. Estas características son abordadas en las conclusiones con el fin de darle contexto y perspectiva a la respuesta de la pregunta de si es viable medir el uso de recursos y servicios electrónicos de una biblioteca universitaria por medio de programas estadísticos para sitios web

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Este trabajo se propone el diseño de una grilla exhaustiva de características específicas de naturaleza formal, informativa y ergonómica para hacer que un sitio Web de bibliotecas universitarias sea de calidad, accesible y garantice el acceso a la información. Para ello, se efectúa un relevamiento de la bibliografía que analiza el tema y se elabora un modelo que pondera dichas características seleccionadas. Las conclusiones demuestran que es posible diseñar un modelo específico de atributos, estableciendo cuáles son los que mejor representan la calidad de una biblioteca universitaria en el entorno Web

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Thesis (Master's)--University of Washington, 2016-06

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Background Strong evidence exists for the efficacy of screening and brief intervention for reducing hazardous drinking. However, problems have been highlighted with respect to its implementation in health-care systems, not least of which is a reluctance of some doctors to discuss alcohol proactively with their patients. Aims To determine the efficacy of a novel web-based screening and brief intervention (e-SBI) to reduce hazardous drinking. Design A double-blind randomized controlled trial. Setting A university student health service. Participants A total of 16 7 students (17-26 years) were recruited in the reception area and completed a 3-minute web-based screen including the Alcohol Use Disorder Identifiation Test (AUDIT) questionnaire. Of these, 112 tested positive, and 104 (52 females) who consented to follow-up were included in the trial. Measurements Drinking frequency, typical occasion quantity, total volume, heavy episode frequency (females > 80 g ethanol, males > 120 g ethanol), number of personal problems, an academic problems score. Intervention Participants were randomized to 10-15 minutes of web-based assessment and personalized feedback on their drinking (intervention, n = 5 1) or to a leaflet-only control group (n = 5 3). Findings Mean baseline AUDIT scores for control and intervention groups were 16.6 (SD = 6.0) and 16.6 (SD = 5.7). At 6 weeks, participants receiving e-SBI reported significantly lower total consumption (geometric mean ratio = 0.74; 9 5 % confidence interval: 0.56-0.96), lower heavy episode frequency (0.63; 0.42-0.92) and fewer personal problems (0.70; 0.54-0.91). At 6 months personal problems remained lower (0.76; 0.60-0.97), although consumption did not differ significantly. At 6 months, academic problems were lower in the intervention group relative to controls (0.72; 0.51-1.02). Conclusions e-SBI reduced hazardous drinking among university students, to an extent similar to that found for practitioner-delivered brief interventions in the general population. e-SBI offers promise as a strategy to reduce alcohol-related harm in a way that is non-intrusive, appealing to the target group, and capable of being incorporated into primary care. Research is required to replicate the findings, to determine the duration of intervention effects, and to investigate the mechanisms by which the intervention operates.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Information and content integration are believed to be a possible solution to the problem of information overload in the Internet. The article is an overview of a simple solution for integration of information and content on the Web. Previous approaches to content extraction and integration are discussed, followed by introduction of a novel technology to deal with the problems, based on XML processing. The article includes lessons learned from solving issues of changing webpage layout, incompatibility with HTML standards and multiplicity of the results returned. The method adopting relative XPath queries over DOM tree proves to be more robust than previous approaches to Web information integration. Furthermore, the prototype implementation demonstrates the simplicity that enables non-professional users to easily adopt this approach in their day-to-day information management routines.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Web wrapper extracts data from HTML document. The accuracy and quality of the information extracted by web wrapper relies on the structure of the HTML document. If an HTML document is changed, the web wrapper may or may not function correctly. This paper presents an Adjacency-Weight method to be used in the web wrapper extraction process or in a wrapper self-maintenance mechanism to validate web wrappers. The algorithm and data structures are illustrated by some intuitive examples.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Models are central tools for modern scientists and decision makers, and there are many existing frameworks to support their creation, execution and composition. Many frameworks are based on proprietary interfaces, and do not lend themselves to the integration of models from diverse disciplines. Web based systems, or systems based on web services, such as Taverna and Kepler, allow composition of models based on standard web service technologies. At the same time the Open Geospatial Consortium has been developing their own service stack, which includes the Web Processing Service, designed to facilitate the executing of geospatial processing - including complex environmental models. The current Open Geospatial Consortium service stack employs Extensible Markup Language as a default data exchange standard, and widely-used encodings such as JavaScript Object Notation can often only be used when incorporated with Extensible Markup Language. Similarly, no successful engagement of the Web Processing Service standard with the well-supported technologies of Simple Object Access Protocol and Web Services Description Language has been seen. In this paper we propose a pure Simple Object Access Protocol/Web Services Description Language processing service which addresses some of the issues with the Web Processing Service specication and brings us closer to achieving a degree of interoperability between geospatial models, and thus realising the vision of a useful 'model web'.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

With the recent rapid growth of the Semantic Web (SW), the processes of searching and querying content that is both massive in scale and heterogeneous have become increasingly challenging. User-friendly interfaces, which can support end users in querying and exploring this novel and diverse, structured information space, are needed to make the vision of the SW a reality. We present a survey on ontology-based Question Answering (QA), which has emerged in recent years to exploit the opportunities offered by structured semantic information on the Web. First, we provide a comprehensive perspective by analyzing the general background and history of the QA research field, from influential works from the artificial intelligence and database communities developed in the 70s and later decades, through open domain QA stimulated by the QA track in TREC since 1999, to the latest commercial semantic QA solutions, before tacking the current state of the art in open user-friendly interfaces for the SW. Second, we examine the potential of this technology to go beyond the current state of the art to support end-users in reusing and querying the SW content. We conclude our review with an outlook for this novel research area, focusing in particular on the R&D directions that need to be pursued to realize the goal of efficient and competent retrieval and integration of answers from large scale, heterogeneous, and continuously evolving semantic sources.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

The Semantic Web relies on carefully structured, well defined, data to allow machines to communicate and understand one another. In many domains (e.g. geospatial) the data being described contains some uncertainty, often due to incomplete knowledge; meaningful processing of this data requires these uncertainties to be carefully analysed and integrated into the process chain. Currently, within the SemanticWeb there is no standard mechanism for interoperable description and exchange of uncertain information, which renders the automated processing of such information implausible, particularly where error must be considered and captured as it propagates through a processing sequence. In particular we adopt a Bayesian perspective and focus on the case where the inputs / outputs are naturally treated as random variables. This paper discusses a solution to the problem in the form of the Uncertainty Markup Language (UncertML). UncertML is a conceptual model, realised as an XML schema, that allows uncertainty to be quantified in a variety of ways i.e. realisations, statistics and probability distributions. UncertML is based upon a soft-typed XML schema design that provides a generic framework from which any statistic or distribution may be created. Making extensive use of Geography Markup Language (GML) dictionaries, UncertML provides a collection of definitions for common uncertainty types. Containing both written descriptions and mathematical functions, encoded as MathML, the definitions within these dictionaries provide a robust mechanism for defining any statistic or distribution and can be easily extended. Universal Resource Identifiers (URIs) are used to introduce semantics to the soft-typed elements by linking to these dictionary definitions. The INTAMAP (INTeroperability and Automated MAPping) project provides a use case for UncertML. This paper demonstrates how observation errors can be quantified using UncertML and wrapped within an Observations & Measurements (O&M) Observation. The interpolation service uses the information within these observations to influence the prediction outcome. The output uncertainties may be encoded in a variety of UncertML types, e.g. a series of marginal Gaussian distributions, a set of statistics, such as the first three marginal moments, or a set of realisations from a Monte Carlo treatment. Quantifying and propagating uncertainty in this way allows such interpolation results to be consumed by other services. This could form part of a risk management chain or a decision support system, and ultimately paves the way for complex data processing chains in the Semantic Web.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

This thesis provides a set of tools for managing uncertainty in Web-based models and workflows.To support the use of these tools, this thesis firstly provides a framework for exposing models through Web services. An introduction to uncertainty management, Web service interfaces,and workflow standards and technologies is given, with a particular focus on the geospatial domain.An existing specification for exposing geospatial models and processes, theWeb Processing Service (WPS), is critically reviewed. A processing service framework is presented as a solutionto usability issues with the WPS standard. The framework implements support for Simple ObjectAccess Protocol (SOAP), Web Service Description Language (WSDL) and JavaScript Object Notation (JSON), allowing models to be consumed by a variety of tools and software. Strategies for communicating with models from Web service interfaces are discussed, demonstrating the difficultly of exposing existing models on the Web. This thesis then reviews existing mechanisms for uncertainty management, with an emphasis on emulator methods for building efficient statistical surrogate models. A tool is developed to solve accessibility issues with such methods, by providing a Web-based user interface and backend to ease the process of building and integrating emulators. These tools, plus the processing service framework, are applied to a real case study as part of the UncertWeb project. The usability of the framework is proved with the implementation of aWeb-based workflow for predicting future crop yields in the UK, also demonstrating the abilities of the tools for emulator building and integration. Future directions for the development of the tools are discussed.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Bacterial lipoproteins have many important functions and represent a class of possible vaccine candidates. The prediction of lipoproteins from sequence is thus an important task for computational vaccinology. Naïve-Bayesian networks were trained to identify SpaseII cleavage sites and their preceding signal sequences using a set of 199 distinct lipoprotein sequences. A comprehensive range of sequence models was used to identify the best model for lipoprotein signal sequences. The best performing sequence model was found to be 10-residues in length, including the conserved cysteine lipid attachment site and the nine residues prior to it. The sensitivity of prediction for LipPred was 0.979, while the specificity was 0.742. Here, we describe LipPred, a web server for lipoprotein prediction; available at the URL: http://www.jenner.ac.uk/LipPred/. LipPred is the most accurate method available for the detection of SpaseIIcleaved lipoprotein signal sequences and the prediction of their cleavage sites.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

The Protein pKa Database (PPD) v1.0 provides a compendium of protein residue-specific ionization equilibria (pKa values), as collated from the primary literature, in the form of a web-accessible postgreSQL relational database. Ionizable residues play key roles in the molecular mechanisms that underlie many biological phenomena, including protein folding and enzyme catalysis. The PPD serves as a general protein pKa archive and as a source of data that allows for the development and improvement of pKa prediction systems. The database is accessed through an HTML interface, which offers two fast, efficient search methods: an amino acid-based query and a Basic Local Alignment Search Tool search. Entries also give details of experimental techniques and links to other key databases, such as National Center for Biotechnology Information and the Protein Data Bank, providing the user with considerable background information.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

In this paper we consider two computer systems and the dynamic Web technologies they are using. Different contemporary dynamic web technologies are described in details and their advantages and disadvantages have been shown. Specific applications are developed, clinic and studying systems, and their programming models are described. Finally we implement these two applications in the students education process: Online studying has been tested in the Technical University – Varna, Web based clinic system has been used for practical education of the students in the Medical College - Sofia, branch V. Tarnovo