924 resultados para SOAP web service


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a hierarchical clustering method for semantic Web service discovery. This method aims to improve the accuracy and efficiency of the traditional service discovery using vector space model. The Web service is converted into a standard vector format through the Web service description document. With the help of WordNet, a semantic analysis is conducted to reduce the dimension of the term vector and to make semantic expansion to meet the user’s service request. The process and algorithm of hierarchical clustering based semantic Web service discovery is discussed. Validation is carried out on the dataset.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Detta examensarbete har utförts på SSAB- Tunnplåt i Borlänge under vårterminen 2004 och omfattar 10 veckors arbete.SSAB sköter idag sin kommunikation med distributionslagren via fax, telefon eller e-post. Eftersom detta är ett ganska tidskrävande kommunikationssätt, vill SSAB ha en smidigare och snabbare kommunikationslösning. Den lösning som SSAB vill ha är en extern Web-service-lösning för att upprätta en säker kommunikation med sina distributionslager.Parallellt med byggandet av Web-service-lösningen arbetades en förvaltningsmodell fram. Den beskriver hur förvaltningsorganisationen med dess rutiner kan se ut vid implementering av lösningen.För att skapa en säker förbindelse med Web-servicen skall en webbklient användas som i sin tur anropar en COM+ komponent. Detta för att kunna skicka med certifikatet ifrån webbklienten till webbservern där Web-servicen ligger. COM+ komponenten måste få tillgång till en användarprofil när den kommunicerar med Web-servicen. Detta för att kunna upprätta en SSL-förbindelse i det inledande skedet. SSL-förbindelsen skall läggas i den VPN-tunnel som mVPN tillhandahåller via WSSAL.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Architectural description languages (ADLs) are used to specify a high-level, compositional view of a software application, specifying how a system is to be composed from coarse-grain components. ADLs usually come equipped with a formal dynamic semantics, facilitating specification and analysis of distributed and event-based systems. In this paper, we describe the TrustME, an ADL framework that provides both a process and a structural view of web service-based systems. We use Petri-net descriptions to give a dynamic view of business workflow for web service collaboration. We adapt the approach of Schmidt to define a form of Meyer's design-by-contract for configuring workflow architectures. This serves as a configuration-level means of constructing safer, more robust systems.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Context is an important factor for the success of dynamic service composition. Although many contextbased AI or workflow approaches have been proposed to support dynamic service composition, there is still an unaddressed issue of the support of fine-granularity context management. In this paper, we propose a granularity-based context model together with an approach to supporting the intelligent context-aware service composing problem. The corresponding case study is provided to show the validity of our approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

With the emergence of cloud computing, the need for flexible and detailed publication and selection of services that expose cloud resources is greatly stressed. While dynamic attributes have improved the publication and selection of resources in distributed systems, the use of dynamic attributes is yet to be tried in Web services: a key element that makes cloud computing possible. We propose a new approach to Web service publication and selection using dynamic attributes shown in Web service WSDL documents, the most commonly accessed and used elements of Web services.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In order to achieve automatic and more intelligent service composition, dynamic description logic (DDL) is proposed and utilized as one emerging logic-level solution. However, reasoning optimization and utilization in such DDL-related solutions is still an open problem. In this paper, we propose the context-aware reasoning-based service agent model (CARSA) which exploits the relationships among different service consumers and providers, together with the corresponding optimization approach to strengthen the effectiveness of Web service composition. Through the model, two reasoning optimization methods are proposed based on the substitute relationship and the dependency relationship, respectively, so irrelevant actions can be filtered out of the reasoning space before the DDL reasoning process is carried out. The case study and experimental analysis demonstrates the capability of the proposed approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Wireless sensor networks (WSNs) are attractive for monitoring and gathering physical information (e.g. temperature) via lots of deployed sensors. For the applications in WSNs, Web service is one of the recommended frameworks to publish, invoke, and manage services. However, the standard Web service description language (WSDL), defines only the service input and output while ignoring the corresponding input-to-output mapping relationships. This presents a serious challenge in distinguishing services with similar input and output interface. In this paper, we address this challenge by embedding the service policy into the traditional WSDL2.0 schema to describe the input-to-output mapping relationships. The service policy is then transformed into a policy binary tree so that the similarity between different Web services can be quantitatively evaluated. Furthermore, a new service redundancy detection approach is proposed based on this similarity. Finally, the case study and experimental analysis illustrate the applicability and capability of the proposed service redundancy detection approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The ubiquity of the Internet and Web has led to the emergency of several Web search engines with varying capabilities. A weakness of existing search engines is the very extensive amount of hits that they can produce. Moreover, only a small number of web users actually know how to utilize the true power of Web search engines. Therefore, there is a need for searching infrastructure to help ease and guide the searching efforts of web users toward their desired objectives. In this paper, we propose a context-based meta-search engine and discuss its implementation on top of the actual Google.com search engine. The proposed meta-search engine benefits the user the most when the user does not know what exact document he or she is looking for. Comparison of the context-based meta-search engine with both Google and Guided Google shows that the results returned by context-based meta-search engine is much more intuitive and accurate than the results returned by both Google and Guided Google.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the past decade there has been massive growth of data on the internet. Many people rely on XML based RSS feeds to receive updates from websites. In this paper, we propose a method for managing the RSS feeds from various news websites. A web service is developed to deliver filtered news items from RSS feeds to a mobile client. Each news item is indexed, subsequently, the indexes are used for filtering news items. Indexing is done in two steps. First, classical text categorization algorithms are used to assign a category to each news item, second, geoparsing is used to assign geolocation data to each news item. An android application is developed to access filtered news items by consuming the proposed web service. A prototype is implemented using Rapid miner 5.0 as the data mining tool and SVM as the classification algorithm. Geoparsing and geocoding web services, and Android API are used to implement location-based access to news items. Experimental results prove that the proposed approach is effective and saves a significant amount of information overload processing time.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The control of industrial processes has become increasingly complex due to variety of factory devices, quality requirement and market competition. Such complexity requires a large amount of data to be treated by the three levels of process control: field devices, control systems and management softwares. To use data effectively in each one of these levels is extremely important to industry. Many of today s industrial computer systems consist of distributed software systems written in a wide variety of programming languages and developed for specific platforms, so, even more companies apply a significant investment to maintain or even re-write their systems for different platforms. Furthermore, it is rare that a software system works in complete isolation. In industrial automation is common that, software had to interact with other systems on different machines and even written in different languages. Thus, interoperability is not just a long-term challenge, but also a current context requirement of industrial software production. This work aims to propose a middleware solution for communication over web service and presents an user case applying the solution developed to an integrated system for industrial data capture , allowing such data to be available simplified and platformindependent across the network

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Service provisioning is a challenging research area for the design and implementation of autonomic service-oriented software systems. It includes automated QoS management for such systems and their applications. Monitoring, Diagnosis and Repair are three key features of QoS management. This work presents a self-healing Web service-based framework that manages QoS degradation at runtime. Our approach is based on proxies. Proxies act on meta-level communications and extend the HTTP envelope of the exchanged messages with QoS-related parameter values. QoS Data are filtered over time and analysed using statistical functions and the Hidden Markov Model. Detected QoS degradations are handled with proxies. We experienced our framework using an orchestrated electronic shop application (FoodShop).