965 resultados para grid service job


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This is an extended version of an article presented at the Second International Conference on Software, Services and Semantic Technologies, Sofia, Bulgaria, 11–12 September 2010.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The authors take a broad view that ultimately Grid- or Web-services must be located via personalised, semantic-rich discovery processes. They argue that such processes must rely on the storage of arbitrary metadata about services that originates from both service providers and service users. Examples of such metadata are reliability metrics, quality of service data, or semantic service description markup. This paper presents UDDI-MT, an extension to the standard UDDI service directory approach that supports the storage of such metadata via a tunnelling technique that ties the metadata store to the original UDDI directory. They also discuss the use of a rich, graph-based RDF query language for syntactic queries on this data. Finally, they analyse the performance of each of these contributions in our implementation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We take a broad view that ultimately Grid- or Web-services must be located via personalised, semantic-rich discovery processes. We argue that such processes must rely on the storage of arbitrary metadata about services that originates from both service providers and service users. Examples of such metadata are reliability metrics, quality of service data, or semantic service description markup. This paper presents UDDI-MT, an extension to the standard UDDI service directory approach that supports the storage of such metadata via a tunnelling technique that ties the metadata store to the original UDDI directory. We also discuss the use of a rich, graph-based RDF query language for syntactic queries on this data. Finally, we analyse the performance of each of these contributions in our implementation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Grid is a large-scale computer system that is capable of coordinating resources that are not subject to centralised control, whilst using standard, open, general-purpose protocols and interfaces, and delivering non-trivial qualities of service. In this chapter, we argue that Grid applications very strongly suggest the use of agent-based computing, and we review key uses of agent technologies in Grids: user agents, able to customize and personalise data; agent communication languages offering a generic and portable communication medium; and negotiation allowing multiple distributed entities to reach service level agreements. In the second part of the chapter, we focus on Grid service discovery, which we have identified as a prime candidate for use of agent technologies: we show that Grid-services need to be located via personalised, semantic-rich discovery processes, which must rely on the storage of arbitrary metadata about services that originates from both service providers and service users. We present UDDI-MT, an extension to the standard UDDI service directory approach that supports the storage of such metadata via a tunnelling technique that ties the metadata store to the original UDDI directory. The outcome is a flexible service registry which is compatible with existing standards and also provides metadata-enhanced service discovery.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Postprint

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Compute grids are used widely in many areas of environmental science, but there has been limited uptake of grid computing by the climate modelling community, partly because the characteristics of many climate models make them difficult to use with popular grid middleware systems. In particular, climate models usually produce large volumes of output data, and running them usually involves complicated workflows implemented as shell scripts. For example, NEMO (Smith et al. 2008) is a state-of-the-art ocean model that is used currently for operational ocean forecasting in France, and will soon be used in the UK for both ocean forecasting and climate modelling. On a typical modern cluster, a particular one year global ocean simulation at 1-degree resolution takes about three hours when running on 40 processors, and produces roughly 20 GB of output as 50000 separate files. 50-year simulations are common, during which the model is resubmitted as a new job after each year. Running NEMO relies on a set of complicated shell scripts and command utilities for data pre-processing and post-processing prior to job resubmission. Grid Remote Execution (G-Rex) is a pure Java grid middleware system that allows scientific applications to be deployed as Web services on remote computer systems, and then launched and controlled as if they are running on the user's own computer. Although G-Rex is general purpose middleware it has two key features that make it particularly suitable for remote execution of climate models: (1) Output from the model is transferred back to the user while the run is in progress to prevent it from accumulating on the remote system and to allow the user to monitor the model; (2) The client component is a command-line program that can easily be incorporated into existing model work-flow scripts. G-Rex has a REST (Fielding, 2000) architectural style, which allows client programs to be very simple and lightweight and allows users to interact with model runs using only a basic HTTP client (such as a Web browser or the curl utility) if they wish. This design also allows for new client interfaces to be developed in other programming languages with relatively little effort. The G-Rex server is a standard Web application that runs inside a servlet container such as Apache Tomcat and is therefore easy to install and maintain by system administrators. G-Rex is employed as the middleware for the NERC1 Cluster Grid, a small grid of HPC2 clusters belonging to collaborating NERC research institutes. Currently the NEMO (Smith et al. 2008) and POLCOMS (Holt et al, 2008) ocean models are installed, and there are plans to install the Hadley Centre’s HadCM3 model for use in the decadal climate prediction project GCEP (Haines et al., 2008). The science projects involving NEMO on the Grid have a particular focus on data assimilation (Smith et al. 2008), a technique that involves constraining model simulations with observations. The POLCOMS model will play an important part in the GCOMS project (Holt et al, 2008), which aims to simulate the world’s coastal oceans. A typical use of G-Rex by a scientist to run a climate model on the NERC Cluster Grid proceeds as follows :(1) The scientist prepares input files on his or her local machine. (2) Using information provided by the Grid’s Ganglia3 monitoring system, the scientist selects an appropriate compute resource. (3) The scientist runs the relevant workflow script on his or her local machine. This is unmodified except that calls to run the model (e.g. with “mpirun”) are simply replaced with calls to "GRexRun" (4) The G-Rex middleware automatically handles the uploading of input files to the remote resource, and the downloading of output files back to the user, including their deletion from the remote system, during the run. (5) The scientist monitors the output files, using familiar analysis and visualization tools on his or her own local machine. G-Rex is well suited to climate modelling because it addresses many of the middleware usability issues that have led to limited uptake of grid computing by climate scientists. It is a lightweight, low-impact and easy-to-install solution that is currently designed for use in relatively small grids such as the NERC Cluster Grid. A current topic of research is the use of G-Rex as an easy-to-use front-end to larger-scale Grid resources such as the UK National Grid service.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The service-oriented approach to performing distributed scientific research is potentially very powerful but is not yet widely used in many scientific fields. This is partly due to the technical difficulties involved in creating services and workflows and the inefficiency of many workflow systems with regard to handling large datasets. We present the Styx Grid Service, a simple system that wraps command-line programs and allows them to be run over the Internet exactly as if they were local programs. Styx Grid Services are very easy to create and use and can be composed into powerful workflows with simple shell scripts or more sophisticated graphical tools. An important feature of the system is that data can be streamed directly from service to service, significantly increasing the efficiency of workflows that use large data volumes. The status and progress of Styx Grid Services can be monitored asynchronously using a mechanism that places very few demands on firewalls. We show how Styx Grid Services can interoperate with with Web Services and WS-Resources using suitable adapters.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Description based on: 1986.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

AMS Subj. Classification: 00-02, (General)

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The purpose of this study is to determine the potential impact of selected organizational factors on boundary-spanning-role employees’ perceptions of service recovery performance. This study also aims to assess the impact of service recovery performance on the intention to leave the job and extrinsic job satisfaction. This study uses a sample of frontline employees in Belek, Antalya, Turkey. The empirical findings revealed that education, team work and role ambiguity as frontline job perceptions were found to exert positive influences on the service recovery performance, but, empowerment, reward, and organizational commitment were found to have negative effects on the service recovery performance.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Rapport de stage présenté à la Faculté des arts et des sciences en vue de l’obtention du grade de maîtrise en criminologie (option intervention).

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Notions of Class and Gender in the Employment Service Job Descriptions This article examines whether job descriptions emphasize different characteristics and competences depending on the occupations’ social class and gender relations. The study is partly a replication of a similar analysis conducted by Gesser in the 1970s. The purpose is to examine the prevalence of stereotypes in occupational descriptions provided by the Swedish state, and if the descriptions contribute to class and gender labeling of occupations and, by extension, its practitioners. Previous research has shown that career guiding materials are characterized by notions of the appropriate practitioner’s class and gender. In this study we depart from the concept of doxa and argue that stereotypical images of occupations are based on common sense that remains unquestioned. The study draws on a quantitative content analysis of 420 job descriptions analyzed by various statistical methods. The overall results show that there are systematic differences. In general, social class seems to have greater impact than gender on what kind of competences that are emphasized in the descriptions. Social skills are emphasized in female dominated occupations, while physical abilities are highlighted in male-dominated occupations. To some extent, these results are uncontroversial, as it also portraits abilities necessary to do the work in different kind of occupations

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The requirement for Grid middleware to be largely transparent to individual users and at the same time act in accordance with their personal needs is a difficult challenge. In e-science scenarios, users cannot be repeatedly interrogated for each operational decision made when enacting experiments on the Grid. It is thus important to specify and enforce policies that enable the environment to be configured to take user preferences into account automatically. In particular, we need to consider the context in which these policies are applied, because decisions are based not only on the rules of the policy but also on the current state of the system. Consideration of context is explicitly addressed, in the agent perspective, when deciding how to balance the achievement of goals and reaction to the environment. One commonly-applied abstraction that balances reaction to multiple events with context-based reasoning in the way suggested by our requirements is the belief-desire-intention (BDI) architecture, which has proven successful in many applications. In this paper, we argue that BDI is an appropriate model for policy enforcement, and describe the application of BDI to policy enforcement in personalising Grid service discovery. We show how this has been implemented in the myGrid registry to provide bioinformaticians with control over the services returned to them by the service discovery process.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Another dilemma also had to be dealt with; Lloyd Gaines was determined to attend law school, not just anywhere but at the University of Missouri. Shortly after the Supreme Court decision, Lloyd Gaines left his civil service job in Michigan and returned home to St. Louis, arriving on New Year’s Eve, 1938. In the meantime, to pay his bills, he took a job as a filling station attendant. On January 9, 1939, Gaines spoke to the St. Louis chapter of the NAACP. He told them he stood “ready, willing, and able to enroll at MU.” Gaines later quit his gas station job. He explained to his family that the station owner substituted inferior gas and that he could not, in good conscience, continue to work there. In the meantime, the state Supreme Court sent the Gaines case back to Boone County to determine whether the new law school at Lincoln would comply with the US Supreme Court’s requirement of “substantial equality.”