862 resultados para Web, Html 5, JavaScript, Dart, Structured Web Programming


Relevância:

50.00% 50.00%

Publicador:

Resumo:

With the emergence of Internet, the global connectivity of computers has become a reality. Internet has progressed to provide many user-friendly tools like Gopher, WAIS, WWW etc. for information publishing and access. The WWW, which integrates all other access tools, also provides a very convenient means for publishing and accessing multimedia and hypertext linked documents stored in computers spread across the world. With the emergence of WWW technology, most of the information activities are becoming Web-centric. Once the information is published on the Web, a user can access this information from any part of the world. A Web browser like Netscape or Internet Explorer is used as a common user interface for accessing information/databases. This will greatly relieve a user from learning the search syntax of individual information systems. Libraries are taking advantage of these developments to provide access to their resources on the Web. CDS/ISIS is a very popular bibliographic information management software used in India. In this tutorial we present details of integrating CDS/ISIS with the WWW. A number of tools are now available for making CDS/ISIS database accessible on the Internet/Web. Some of these are 1) the WAIS_ISIS Server. 2) the WWWISIS Server 3) the IQUERY Server. In this tutorial, we have explained in detail the steps involved in providing Web access to an existing CDS/ISIS database using the freely available software, WWWISIS. This software is developed, maintained and distributed by BIREME, the Latin American & Caribbean Centre on Health Sciences Information. WWWISIS acts as a server for CDS/ISIS databases in a WWW client/server environment. It supports functions for searching, formatting and data entry operations over CDS/ISIS databases. WWWISIS is available for various operating systems. We have tested this software on Windows '95, Windows NT and Red Hat Linux release 5.2 (Appolo) Kernel 2. 0. 36 on an i686. The testing was carried out using IISc's main library's OPAC containing more than 80,000 records and Current Contents issues (bibliographic data) containing more than 25,000 records. WWWISIS is fully compatible with CDS/ISIS 3.07 file structure. However, on a system running Unix or its variant, there is no guarantee of this compatibility. It is therefore safe to recreate the master and the inverted files, using utilities provided by BIREME, under Unix environment.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

This paper presents an intelligent procurement marketplace for finding the best mix of web services to dynamically compose the business process desired by a web service requester. We develop a combinatorial auction approach that leads to an integer programming formulation for the web services composition problem. The model takes into account the Quality of Service (QoS) and Service Level Agreements (SLA) for differentiating among multiple service providers who are capable of fulfilling a functionality. An important feature of the model is interface aware composition.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Background: Haemophilus influenzae (H. Influenzae) is the causative agent of pneumonia, bacteraemia and meningitis. The organism is responsible for large number of deaths in both developed and developing countries. Even-though the first bacterial genome to be sequenced was that of H. Influenzae, there is no exclusive database dedicated for H. Influenzae. This prompted us to develop the Haemophilus influenzae Genome Database (HIGDB). Methods: All data of HIGDB are stored and managed in MySQL database. The HIGDB is hosted on Solaris server and developed using PERL modules. Ajax and JavaScript are used for the interface development. Results: The HIGDB contains detailed information on 42,741 proteins, 18,077 genes including 10 whole genome sequences and also 284 three dimensional structures of proteins of H. influenzae. In addition, the database provides ``Motif search'' and ``GBrowse''. The HIGDB is freely accessible through the URL:http://bioserverl.physicslisc.ernetin/HIGDB/. Discussion: The HIGDB will be a single point access for bacteriological, clinical, genomic and proteomic information of H. influenzae. The database can also be used to identify DNA motifs within H. influenzae genomes and to compare gene or protein sequences of a particular strain with other strains of H. influenzae. (C) 2014 Elsevier Ltd. All rights reserved.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

This presentation describes how the Galveston Bay Bibliography and Galveston Bay Information Center projects transitioned from a paper world to electronic format. The mission of the Galveston Bay Information Center (GBIC) is to serve as a repository for information and a resource for research on Galveston Bay, its watershed, and the coastal region for all members of the Galveston Bay Community. GBIC was created in response to critical losses of data and information that were identified in the early stages of Galveston Bay National Estuary Program and includes a physical collection as well as a bibliography. With the loss of funding imminent, the administrators of GBIC recognized the need to preserve the collection and its digital bibliography in perpetuity. Enter Texas Digital Library. The University’s participation in TDL enabled GBIC to make the Galveston Bay Bibliography more robust and visible to the entire world.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Background: Two distinct trends are emerging with respect to how data is shared, collected, and analyzed within the bioinformatics community. First, Linked Data, exposed as SPARQL endpoints, promises to make data easier to collect and integrate by moving towards the harmonization of data syntax, descriptive vocabularies, and identifiers, as well as providing a standardized mechanism for data access. Second, Web Services, often linked together into workflows, normalize data access and create transparent, reproducible scientific methodologies that can, in principle, be re-used and customized to suit new scientific questions. Constructing queries that traverse semantically-rich Linked Data requires substantial expertise, yet traditional RESTful or SOAP Web Services cannot adequately describe the content of a SPARQL endpoint. We propose that content-driven Semantic Web Services can enable facile discovery of Linked Data, independent of their location. Results: We use a well-curated Linked Dataset - OpenLifeData - and utilize its descriptive metadata to automatically configure a series of more than 22,000 Semantic Web Services that expose all of its content via the SADI set of design principles. The OpenLifeData SADI services are discoverable via queries to the SHARE registry and easy to integrate into new or existing bioinformatics workflows and analytical pipelines. We demonstrate the utility of this system through comparison of Web Service-mediated data access with traditional SPARQL, and note that this approach not only simplifies data retrieval, but simultaneously provides protection against resource-intensive queries. Conclusions: We show, through a variety of different clients and examples of varying complexity, that data from the myriad OpenLifeData can be recovered without any need for prior-knowledge of the content or structure of the SPARQL endpoints. We also demonstrate that, via clients such as SHARE, the complexity of federated SPARQL queries is dramatically reduced.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Tradicionalmente la enseñanza de los autómatas y lenguajes formales basa su principal aplicación práctica en la construcción de compiladores. Sin embargo, las tareas de diseño y programación necesarias son excesivamente complejas como para que los estudiantes, que están cursando el tercer cuatrimestre de la Ingeniería, puedan abordarlas con el rigor necesario. Es posible incorporar otro enfoque práctico, real y más actual de las expresiones regulares en estas asignaturas, aprovechando su frecuente uso como herramienta de especificación de patrones a la hora de diseñar formularios de entrada de datos en diferentes contextos y, particularmente, en aplicaciones web de tres capas. El hecho de trabajar esta competencia junto con el desarrollo teórico de las expresiones regulares permite a los estudiantes ser conscientes de la importante utilidad práctica de este concepto, sin restringirlo a otros usos más clásicos relacionados con el diseño de procesadores de textos o analizadores léxicos. Durante el curso 2006-07 se ha propuesto a los estudiantes de Ingeniería Técnica en Informática de Sistemas de la Universidad del País Vasco desarrollar fragmentos de código basados en una notación formal para resolver problemas de reconocimiento de patrones. La experiencia se ha llevado a cabo utilizando concretamente la notación, inspirada en las expresiones regulares, de JavaScript, resultando viable, efectiva y bien valorada por parte de los estudiantes.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

The primary objective of this project, “the Assessment of Existing Information on Atlantic Coastal Fish Habitat”, is to inform conservation planning for the Atlantic Coastal Fish Habitat Partnership (ACFHP). ACFHP is recognized as a Partnership by the National Fish Habitat Action Plan (NFHAP), whose overall mission is to protect, restore, and enhance the nation’s fish and aquatic communities through partnerships that foster fish habitat conservation. This project is a cooperative effort of NOAA/NOS Center for Coastal Monitoring and Assessment (CCMA) Biogeography Branch and ACFHP. The Assessment includes three components; 1. a representative bibliographic and assessment database, 2. a Geographical Information System (GIS) spatial framework, and 3. a summary document with description of methods, analyses of habitat assessment information, and recommendations for further work. The spatial bibliography was created by linking the bibliographic table developed in Microsoft Excel and exported to SQL Server, with the spatial framework developed in ArcGIS and exported to GoogleMaps. The bibliography is a comprehensive, searchable database of over 500 selected documents and data sources on Atlantic coastal fish species and habitats. Key information captured for each entry includes basic bibliographic data, spatial footprint (e.g. waterbody or watershed), species and habitats covered, and electronic availability. Information on habitat condition indicators, threats, and conservation recommendations are extracted from each entry and recorded in a separate linked table. The spatial framework is a functional digital map based on polygon layers of watersheds, estuarine and marine waterbodies derived from NOAA’s Coastal Assessment Framework, MMS/NOAA’s Multipurpose Marine Cadastre, and other sources, providing spatial reference for all of the documents cited in the bibliography. Together, the bibliography and assessment tables and their spatial framework provide a powerful tool to query and assess available information through a publicly available web interface. They were designed to support the development of priorities for ACFHP’s conservation efforts within a geographic area extending from Maine to Florida, and from coastal watersheds seaward to the edge of the continental shelf. The Atlantic Coastal Fish Habitat Partnership has made initial use of the Assessment of Existing Information. Though it has not yet applied the AEI in a systematic or structured manner, it expects to find further uses as the draft conservation strategic plan is refined, and as regional action plans are developed. It also provides a means to move beyond an “assessment of existing information” towards an “assessment of fish habitat”, and is being applied towards the National Fish Habitat Action Plan (NFHAP) 2010 Assessment. Beyond the scope of the current project, there may be application to broader initiatives such as Integrated Ecosystem Assessments (IEAs), Ecosystem Based Management (EBM), and Marine Spatial Planning (MSP).

Relevância:

50.00% 50.00%

Publicador:

Resumo:

基于Java EE技术的Web容器已经成为面向Web计算环境的主流中间件平台,它为创建、部署、运行、集成和管理企业Web组件提供了基础设施支持。然而,企业用户规模的不断增长与企业Web应用程序的日趋复杂使得Web容器的性能面临严峻的威胁。 影响服务器性能的主要因素是服务器的资源与负载。在传统服务器设计领域,研究者已经取得了一系列的研究成果,如资源管理、准入控制、动态缓存、差分服务等。然而,上述传统服务器领域的研究成果难以满足开放、动态的Web计算环境需求。首先,Web容器无法预先获知企业Web应用程序的资源需求与负载特征,这使得它不得不采用一种“尽力而为”的服务策略。其次,传统资源管理技术难以解决不同资源之间的管理冲突,影响了资源的利用率。最后,传统的负载管理技术侧重于负载规模的变化,很少考虑负载的组成特征与行为特征。为解决上述问题,论文提出了一种面向资源的分阶段体系结构ROSA,并研究了基于ROSA的动态缓存技术与差分服务技术,具体研究内容如下: 论文首先提出了面向资源的分阶段体系结构ROSA,它将请求处理所需的处理资源(主要指线程)与共享资源加以区分,并且通过分阶段体系结构设计来有效解决处理资源阻塞。同时,论文扩展Petri网模型给出了ROSA体系结构的描述模型RDM与性能模型RPM,前者用于验证ROSA体系结构设计方案的结构性质与行为性质,后者用于定量的计算ROSA体系结构设计方案的平均响应时间、吞吐率等性能指标。最后,论文提出了相关设计模式用于ROSA体系结构的开发。 论文提出了一种访问模式驱动的动态缓存机制,消除了Web容器、Web组件与动态缓存机制的耦合,并能够利用客户访问模式提高Web容器的缓存效率。首先,论文设计了Web容器动态缓存框架来缓存Web组件动态生成的Web内容,减少Web组件的重复执行。然后,论文设计了模式图描述客户的访问行为,有效的消除了传统马尔可夫模型引入的无效访问行为。最后,论文在模式图基础上分别基于会话和缓存对象设计了两个预测概率函数以及访问模式驱动的缓存替换算法,有效的提高了动态缓存机制的效率。 论文提出了一种自适应差分服务机制以满足Web容器剧烈变化的负载,同时为企业Web应用程序的差分服务需求提供支持。首先,论文基于客户行为特征设计了动态的会话优先级,弥补了传统静态优先级定义(客户优先级、请求优先级)无法描述客户行为优先级的缺点。然后,通过扩展ROSA体系结构设计,提供了准入控制、优先级调度、资源预留等多种差分服务策略,从而适应Web容器的不同负载。最后,论文设计了自适应差分服务策略选择算法,根据Web容器负载状况自适应选择差分服务策略,消除了单一差分服务策略在某些情况下出现的资源浪费、请求“饿死”、不必要的请求丢失现象,保障了关键请求的性能指标。 最后,以上研究成果已经应用在一个高性能的Web容器系统Once Web容器中。在Java EE性能测试基准TPC-W的测试中取得了较好的效果。

Relevância:

50.00% 50.00%

Publicador:

Resumo:

简化多语言Web网站服务的管理与开发。在实践中,管理与开发多语言网站的大部分工作是保持网站的各种信息之间相互独立。在开发与管理多语言网站的过程中有许多与人相关的角色,如设计人员、实施人员(如程序员)、系统管理员、翻译人员与用户等角色。按照这些不同的角色对网站的各种信息进行严格分类,并保持在同一个网站中这些分类后的信息相互独立,也就是说负责翻译的人员不需要看到脚本语言,如JavaScript。同样,图形设计人员也不需要精通多种语言,也不必在多种语言环境中工作。从以上方面论述如何设计及实现多语言网站.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

形式化定义了Web服务组合过程中的5种基本逻辑结构,并采用有色Petri网表示,然后将其抽象为服务的代数运算;在此基础上,提出了经过服务运算后得到的服务的性质及组合服务的构造方法;最后通过实例分析,说明该建模方法可以保证组合的服务是正确且可终止的。

Relevância:

50.00% 50.00%

Publicador:

Resumo:

With long-term marine surveys and research, and especially with the development of new marine environment monitoring technologies, prodigious amounts of complex marine environmental data are generated, and continuously increase rapidly. Features of these data include massive volume, widespread distribution, multiple-sources, heterogeneous, multi-dimensional and dynamic in structure and time. The present study recommends an integrative visualization solution for these data, to enhance the visual display of data and data archives, and to develop a joint use of these data distributed among different organizations or communities. This study also analyses the web services technologies and defines the concept of the marine information gird, then focuses on the spatiotemporal visualization method and proposes a process-oriented spatiotemporal visualization method. We discuss how marine environmental data can be organized based on the spatiotemporal visualization method, and how organized data are represented for use with web services and stored in a reusable fashion. In addition, we provide an original visualization architecture that is integrative and based on the explored technologies. In the end, we propose a prototype system of marine environmental data of the South China Sea for visualizations of Argo floats, sea surface temperature fields, sea current fields, salinity, in-situ investigation data, and ocean stations. An integration visualization architecture is illustrated on the prototype system, which highlights the process-oriented temporal visualization method and demonstrates the benefit of the architecture and the methods described in this study.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

本文针对多用户访问Web数据库的过程进行了分析并提出改进思路,然后利用带抑止弧的扩充Petri网对改进后的访问过程进行建模。

Relevância:

50.00% 50.00%

Publicador:

Resumo:

introdução. Por onde começar?. Alov Mapa. Interface do Alov mapa. Criar arquivo HTML. Definir o arquivo de configuração do projeto. Usando a opção de busca. Usando o botão de informações. Usando o botão de ligação Web. Usando o botão de seleção (SELECT). Conclusões.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

O objetivo deste trabalho é mostrar um sistema de MTI pela web, implementado na Embrapa Informática Agropecuária e pode, com algumas adaptações, servir para outras empresas.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Li, Longzhuang, Liu, Yonghuai, Obregon, A., Weatherston, M. Visual Segmentation-Based Data Record Extraction From Web Documents. Proceedings of IEEE International Conference on Information Reuse and Integration, 2007, pp. 502-507. Sponsorship: IEEE