803 resultados para composite web services


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Urquhart, C. (editor for JUSTEIS team), Spink, S., Thomas, R., Yeoman, A., Durbin, J., Turner, J., Armstrong, A., Lonsdale, R. & Fenton, R. (2003). JUSTEIS (JISC Usage Surveys: Trends in Electronic Information Services) Strand A: survey of end users of all electronic information services (HE and FE), with Action research report. Final report 2002/2003 Cycle Four. Aberystwyth: Department of Information Studies, University of Wales Aberystwyth with Information Automation Ltd (CIQM). Sponsorship: JISC

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Poster pokazuje metody komunikacji z czytelnikiem stosowane w Bibliotece Uniwersyteckiej w Poznaniu w technologii mediów cyfrowych. Cyfrowe narzędzia komunikacji stały się bardzo pomocne, niemal niezbędne w pozyskiwaniu nowych czytelników, podtrzymywaniu i rozwijaniu współpracy w społeczności w sieci Web.2.0, zarówno tej globalnej, jak i lokalnej akademickiej. Strona WWW jako statyczna komunikacyjnie jest wspierana przez fora dyskusyjne, chaty, wideokonferencje, warsztaty informacyjne, które są prowadzone w czasie rzeczywistym. Twórczą siłę relacji społecznych z biblioteką rozwinęły interaktywne serwisy społecznościowe (Facebook) oraz komunikatory internetowe integrowane na platformie Ask a Librarian. Biblioteka stała się Biblioteką 2.0 ukierunkowaną na komunikację z czytelnikiem. Aktywne uczestnictwo i udział czytelników przy tworzeniu zasobów naukowych wdrożyliśmy w projekcie instytucjonalnego repozytorium - Adam Mickiewicz Repository (AMUR). Biblioteka zmienia się dla czytelników i z czytelnikami. Wykorzystywane platformy i serwisy społecznościowe dostarczają unikatowych danych o nowych potrzebach informacyjnych i oczekiwaniach docelowego Patrona 2.0, co skutkuje doskonaleniu usług istniejących i tworzeniu nowych. Biblioteka monitoruje usługi i potrzeby czytelników przez prowadzone badania społeczne. Technologie cyfrowe stosowane w komunikacji sprawiają, iż biblioteka staje się bliższa, bardziej dostępna, aby stać się w rezultacie partnerem dla stałych i nowych czytelników. Biblioteka Uniwersytecka w Poznaniu bierze udział w programach europejskich w zakresie katalogowania i digitalizacji zasobu biblioteki cyfrowej WBC, w zakresie wdrożenia nowych technologii i rozwiązań podnoszących jakość usług bibliotecznych, działalności kulturotwórczej (Poznańska Dyskusyjna Akademia Kominksu, deBiUty) i edukacji informacyjnej. Biblioteka Uniwersytecka w Poznaniu jest członkiem organizacji międzynarodowych: LIBER (Liga Europejskich Bibliotek Naukowych), IAML (Stowarzyszenie Bibliotek Muzycznych, Archiwów i Ośrodków Dokumentacji), CERL - Europejskie Konsorcjum Bibliotek Naukowych.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The exploding demand for services like the World Wide Web reflects the potential that is presented by globally distributed information systems. The number of WWW servers world-wide has doubled every 3 to 5 months since 1993, outstripping even the growth of the Internet. At each of these self-managed sites, the Common Gateway Interface (CGI) and Hypertext Transfer Protocol (HTTP) already constitute a rudimentary basis for contributing local resources to remote collaborations. However, the Web has serious deficiencies that make it unsuited for use as a true medium for metacomputing --- the process of bringing hardware, software, and expertise from many geographically dispersed sources to bear on large scale problems. These deficiencies are, paradoxically, the direct result of the very simple design principles that enabled its exponential growth. There are many symptoms of the problems exhibited by the Web: disk and network resources are consumed extravagantly; information search and discovery are difficult; protocols are aimed at data movement rather than task migration, and ignore the potential for distributing computation. However, all of these can be seen as aspects of a single problem: as a distributed system for metacomputing, the Web offers unpredictable performance and unreliable results. The goal of our project is to use the Web as a medium (within either the global Internet or an enterprise intranet) for metacomputing in a reliable way with performance guarantees. We attack this problem one four levels: (1) Resource Management Services: Globally distributed computing allows novel approaches to the old problems of performance guarantees and reliability. Our first set of ideas involve setting up a family of real-time resource management models organized by the Web Computing Framework with a standard Resource Management Interface (RMI), a Resource Registry, a Task Registry, and resource management protocols to allow resource needs and availability information be collected and disseminated so that a family of algorithms with varying computational precision and accuracy of representations can be chosen to meet realtime and reliability constraints. (2) Middleware Services: Complementary to techniques for allocating and scheduling available resources to serve application needs under realtime and reliability constraints, the second set of ideas aim at reduce communication latency, traffic congestion, server work load, etc. We develop customizable middleware services to exploit application characteristics in traffic analysis to drive new server/browser design strategies (e.g., exploit self-similarity of Web traffic), derive document access patterns via multiserver cooperation, and use them in speculative prefetching, document caching, and aggressive replication to reduce server load and bandwidth requirements. (3) Communication Infrastructure: Finally, to achieve any guarantee of quality of service or performance, one must get at the network layer that can provide the basic guarantees of bandwidth, latency, and reliability. Therefore, the third area is a set of new techniques in network service and protocol designs. (4) Object-Oriented Web Computing Framework A useful resource management system must deal with job priority, fault-tolerance, quality of service, complex resources such as ATM channels, probabilistic models, etc., and models must be tailored to represent the best tradeoff for a particular setting. This requires a family of models, organized within an object-oriented framework, because no one-size-fits-all approach is appropriate. This presents a software engineering challenge requiring integration of solutions at all levels: algorithms, models, protocols, and profiling and monitoring tools. The framework captures the abstract class interfaces of the collection of cooperating components, but allows the concretization of each component to be driven by the requirements of a specific approach and environment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With web caching and cache-related services like CDNs and edge services playing an increasingly significant role in the modern internet, the problem of the weak consistency and coherence provisions in current web protocols is becoming increasingly significant and drawing the attention of the standards community [LCD01]. Toward this end, we present definitions of consistency and coherence for web-like environments, that is, distributed client-server information systems where the semantics of interactions with resource are more general than the read/write operations found in memory hierarchies and distributed file systems. We then present a brief review of proposed mechanisms which strengthen the consistency of caches in the web, focusing upon their conceptual contributions and their weaknesses in real-world practice. These insights motivate a new mechanism, which we call "Basis Token Consistency" or BTC; when implemented at the server, this mechanism allows any client (independent of the presence and conformity of any intermediaries) to maintain a self-consistent view of the server's state. This is accomplished by annotating responses with additional per-resource application information which allows client caches to recognize the obsolescence of currently cached entities and identify responses from other caches which are already stale in light of what has already been seen. The mechanism requires no deviation from the existing client-server communication model, and does not require servers to maintain any additional per-client state. We discuss how our mechanism could be integrated into a fragment-assembling Content Management System (CMS), and present a simulation-driven performance comparison between the BTC algorithm and the use of the Time-To-Live (TTL) heuristic.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Global biodiversity is eroding at an alarming rate, through a combination of anthropogenic disturbance and environmental change. Ecological communities are bewildering in their complexity. Experimental ecologists strive to understand the mechanisms that drive the stability and structure of these complex communities in a bid to inform nature conservation and management. Two fields of research have had high profile success at developing theories related to these stabilising structures and testing them through controlled experimentation. Biodiversity-ecosystem functioning (BEF) research has explored the likely consequences of biodiversity loss on the functioning of natural systems and the provision of important ecosystem services. Empirical tests of BEF theory often consist of simplified laboratory and field experiments, carried out on subsets of ecological communities. Such experiments often overlook key information relating to patterns of interactions, important relationships, and fundamental ecosystem properties. The study of multi-species predator-prey interactions has also contributed much to our understanding of how complex systems are structured, particularly through the importance of indirect effects and predator suppression of prey populations. A growing number of studies describe these complex interactions in detailed food webs, which encompass all the interactions in a community. This has led to recent calls for an integration of BEF research with the comprehensive study of food web properties and patterns, to help elucidate the mechanisms that allow complex communities to persist in nature. This thesis adopts such an approach, through experimentation at Lough Hyne marine reserve, in southwest Ireland. Complex communities were allowed to develop naturally in exclusion cages, with only the diversity of top trophic levels controlled. Species removals were carried out and the resulting changes to predator-prey interactions, ecosystem functioning, food web properties, and stability were studied in detail. The findings of these experiments contribute greatly to our understanding of the stability and structure of complex natural communities.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The mobile cloud computing paradigm can offer relevant and useful services to the users of smart mobile devices. Such public services already exist on the web and in cloud deployments, by implementing common web service standards. However, these services are described by mark-up languages, such as XML, that cannot be comprehended by non-specialists. Furthermore, the lack of common interfaces for related services makes discovery and consumption difficult for both users and software. The problem of service description, discovery, and consumption for the mobile cloud must be addressed to allow users to benefit from these services on mobile devices. This paper introduces our work on a mobile cloud service discovery solution, which is utilised by our mobile cloud middleware, Context Aware Mobile Cloud Services (CAMCS). The aim of our approach is to remove complex mark-up languages from the description and discovery process. By means of the Cloud Personal Assistant (CPA) assigned to each user of CAMCS, relevant mobile cloud services can be discovered and consumed easily by the end user from the mobile device. We present the discovery process, the architecture of our own service registry, and service description structure. CAMCS allows services to be used from the mobile device through a user's CPA, by means of user defined tasks. We present the task model of the CPA enabled by our solution, including automatic tasks, which can perform work for the user without an explicit request.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE: The research studied the status of hospital librarians and library services to better inform the Medical Library Association's advocacy activities. METHODS: The Vital Pathways Survey Subcommittee of the Task Force on Vital Pathways for Hospital Librarians distributed a web-based survey to hospital librarians and academic health sciences library directors. The survey results were compared to data collected in a 1989 survey of hospital libraries by the American Hospital Association in order to identify any trends in hospital libraries, roles of librarians, and library services. A web-based hospital library report form based on the survey questions was also developed to more quickly identify changes in the status of hospital libraries on an ongoing basis. RESULTS: The greatest change in library services between 1989 and 2005/06 was in the area of access to information, with 40% more of the respondents providing access to commercial online services, 100% more providing access to Internet resources, and 28% more providing training in database searching and use of information resources. Twenty-nine percent (n = 587) of the 2005/06 respondents reported a decrease in staff over the last 5 years. CONCLUSIONS: Survey data support reported trends of consolidation of hospitals and hospital libraries and additions of new services. These services have likely required librarians to acquire new skills. It is hoped that future surveys will be undertaken to continue to study these trends.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

PURPOSE: Risk-stratified guidelines can improve quality of care and cost-effectiveness, but their uptake in primary care has been limited. MeTree, a Web-based, patient-facing risk-assessment and clinical decision support tool, is designed to facilitate uptake of risk-stratified guidelines. METHODS: A hybrid implementation-effectiveness trial of three clinics (two intervention, one control). PARTICIPANTS: consentable nonadopted adults with upcoming appointments. PRIMARY OUTCOME: agreement between patient risk level and risk management for those meeting evidence-based criteria for increased-risk risk-management strategies (increased risk) and those who do not (average risk) before MeTree and after. MEASURES: chart abstraction was used to identify risk management related to colon, breast, and ovarian cancer, hereditary cancer, and thrombosis. RESULTS: Participants = 488, female = 284 (58.2%), white = 411 (85.7%), mean age = 58.7 (SD = 12.3). Agreement between risk management and risk level for all conditions for each participant, except for colon cancer, which was limited to those <50 years of age, was (i) 1.1% (N = 2/174) for the increased-risk group before MeTree and 16.1% (N = 28/174) after and (ii) 99.2% (N = 2,125/2,142) for the average-risk group before MeTree and 99.5% (N = 2,131/2,142) after. Of those receiving increased-risk risk-management strategies at baseline, 10.5% (N = 2/19) met criteria for increased risk. After MeTree, 80.7% (N = 46/57) met criteria. CONCLUSION: MeTree integration into primary care can improve uptake of risk-stratified guidelines and potentially reduce "overuse" and "underuse" of increased-risk services.Genet Med 18 10, 1020-1028.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this work is to improve retrieval and navigation services on bibliographic data held in digital libraries. This paper presents the design and implementation of OntoBib¸ an ontology-based bibliographic database system that adopts ontology-driven search in its retrieval. The presented work exemplifies how a digital library of bibliographic data can be managed using Semantic Web technologies and how utilizing the domain specific knowledge improves both search efficiency and navigation of web information and document retrieval.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Se analizan y describen las principales líneas de trabajo de la Web Semántica en el ámbito de los archivos de televisión. Para ello, se analiza y contextualiza la web semántica desde una perspectiva general para posteriormente analizar las principales iniciativas que trabajan con lo audiovisual: Proyecto MuNCH, Proyecto S5T, Semantic Television y VideoActive.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, the support for legacy application, which is one of the important advantages of Grid computing, is presented. The ability to reuse existing codes/applications in combination with other Web/Internet technologies, such as Java, makes Grid computing a good choice for developers to wrap existing applications behind Intranet or the Internet. The approach developed can be used for migrating legacy applications into Grid Services, which speeds up the popularization of Grid technology. The approach is illustrated using a case study with detailed description of its implementation step by step. Globus Toolkit is utilized to develop the system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cold-formed steel sections are often used as wall studs or floor joists; such sections often include web holes for ease of installation of the services. Cold-formed steel design codes, however, do not consider the effect of such web holes. In this paper, a combination of experimental tests and non-linear elasto-plastic finite element analyses are used to investigate the effect of such holes on web crippling under interior-two-flange (ITF) loading conditions; the cases of both flange fastened and flange unfastened are considered. A good agreement between the experimental tests and finite element analyses was obtained. The finite element model was then used for the purposes of a parametric study on the effect of different sizes and position of holes in the web. It was demonstrated that the main factors influencing the web crippling strength are the ratio of the hole depth to the depth of the web, and the ratio of the distance from the edge of the bearing to the flat depth of web. Design recommendations in the form of web crippling strength reduction factors are proposed, that are conservative to both the experimental and finite element results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The postbuckling behaviour of a panel with blade-stiffeners incorporating tapered flanges was experimentally investigated. A new failure mechanism was identified for this particular type of stiffener. Failure was initiated by mid-plane delamination at the free edge of the postbuckled stiffener web at a node-line. This was consistent with an interlaminar shear stress failure and was calculated from strain gauge measurements using an approximate analysis based on lamination theory and incorporating edge effects. The critical shear stress was found to agree well with the shear strength obtained from a three-point bending test of the web laminate.