945 resultados para execution traces


Relevância:

30.00% 30.00%

Publicador:

Resumo:

A Execução Condicional Dinâmica (DCE) é uma alternativa para redução dos custos relacionados a desvios previstos incorretamente. A idéia básica é buscar todos os fluxos produzidos por um desvio que obedecem algumas restrições relativas à complexidade e tamanho. Como conseqüência, um número menor de previsões é executado, e assim, um número mais baixo de desvios é incorretamente previsto. Contudo, tal como outras soluções multi-fluxo, o DCE requer uma estrutura de controle mais complexa. Na arquitetura DCE, é observado que várias réplicas da mesma instrução são despachadas para as unidades funcionais, bloqueando recursos que poderiam ser utilizados por outras instruções. Essas réplicas são geradas após o ponto de convergência dos diversos fluxos em execução e são necessárias para garantir a semântica correta entre instruções dependentes de dados. Além disso, o DCE continua produzindo réplicas até que o desvio que gerou os fluxos seja resolvido. Assim, uma seção completa do código pode ser replicado, reduzindo o desempenho. Uma alternativa natural para esse problema é reusar essas seções (ou traços) que são replicadas. O objetivo desse trabalho é analisar e avaliar a efetividade do reuso de valores na arquitetura DCE. Como será apresentado, o princípio do reuso, em diferentes granularidades, pode reduzir efetivamente o problema das réplicas e levar a aumentos de desempenho.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Composite web services comprise several component web services. When a composite web service is executed centrally, a single web service engine is responsible for coordinating the execution of the components, which may create a bottleneck and degrade the overall throughput of the composite service when there are a large number of service requests. Potentially this problem can be handled by decentralizing execution of the composite web service, but this raises the issue of how to partition a composite service into groups of component services such that each group can be orchestrated by its own execution engine while ensuring acceptable overall throughput of the composite service. Here we present a novel penalty-based genetic algorithm to solve the composite web service partitioning problem. Empirical results show that our new algorithm outperforms existing heuristic-based solutions.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An Asset Management (AM) life-cycle constitutes a set of processes that align with the development, operation and maintenance of assets, in order to meet the desired requirements and objectives of the stake holders of the business. The scope of AM is often broad within an organization due to the interactions between its internal elements such as human resources, finance, technology, engineering operation, information technology and management, as well as external elements such as governance and environment. Due to the complexity of the AM processes, it has been proposed that in order to optimize asset management activities, process modelling initiatives should be adopted. Although organisations adopt AM principles and carry out AM initiatives, most do not document or model their AM processes, let alone enacting their processes (semi-) automatically using a computer-supported system. There is currently a lack of knowledge describing how to model AM processes through a methodical and suitable manner so that the processes are streamlines and optimized and are ready for deployment in a computerised way. This research aims to overcome this deficiency by developing an approach that will aid organisations in constructing AM process models quickly and systematically whilst using the most appropriate techniques, such as workflow technology. Currently, there is a wealth of information within the individual domains of AM and workflow. Both fields are gaining significant popularity in many industries thus fuelling the need for research in exploring the possible benefits of their cross-disciplinary applications. This research is thus inspired to investigate these two domains to exploit the application of workflow to modelling and execution of AM processes. Specifically, it will investigate appropriate methodologies in applying workflow techniques to AM frameworks. One of the benefits of applying workflow models to AM processes is to adapt and enable both ad-hoc and evolutionary changes over time. In addition, this can automate an AM process as well as to support the coordination and collaboration of people that are involved in carrying out the process. A workflow management system (WFMS) can be used to support the design and enactment (i.e. execution) of processes and cope with changes that occur to the process during the enactment. So far few literatures can be found in documenting a systematic approach to modelling the characteristics of AM processes. In order to obtain a workflow model for AM processes commonalities and differences between different AM processes need to be identified. This is the fundamental step in developing a conscientious workflow model for AM processes. Therefore, the first stage of this research focuses on identifying the characteristics of AM processes, especially AM decision making processes. The second stage is to review a number of contemporary workflow techniques and choose a suitable technique for application to AM decision making processes. The third stage is to develop an intermediate ameliorated AM decision process definition that improves the current process description and is ready for modelling using the workflow language selected in the previous stage. All these lead to the fourth stage where a workflow model for an AM decision making process is developed. The process model is then deployed (semi-) automatically in a state-of-the-art WFMS demonstrating the benefits of applying workflow technology to the domain of AM. Given that the information in the AM decision making process is captured at an abstract level within the scope of this work, the deployed process model can be used as an executable guideline for carrying out an AM decision process in practice. Moreover, it can be used as a vanilla system that, once being incorporated with rich information from a specific AM decision making process (e.g. in the case of a building construction or a power plant maintenance), is able to support the automation of such a process in a more elaborated way.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we examine the use of a Kalman filter to aid in the mission planning process for autonomous gliders. Given a set of waypoints defining the planned mission and a prediction of the ocean currents from a regional ocean model, we present an approach to determine the best, constant, time interval at which the glider should surface to maintain a prescribed tracking error, and minimizing time on the ocean surface. We assume basic parameters for the execution of a given mission, and provide the results of the Kalman filter mission planning approach. These results are compared with previous executions of the given mission scenario.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Altered expression of the INT6 gene, encoding the e subunit of the translational initiation factor eIF3, occurs in human breast cancers, but how INT6 relates to carcinogenesis remains unestablished. Here, we show that INT6 is involved in the DNA damage response. INT6 was required for cell survival following γ-irradiation and G(2)-M checkpoint control. RNA interference-mediated silencing of INT6 reduced phosphorylation of the checkpoint kinases CHK1 and CHK2 after DNA damage. In addition, INT6 silencing prevented sustained accumulation of ataxia telangiectasia mutated (ATM) at DNA damage sites in cells treated with γ-radiation or the radiomimetic drug neocarzinostatin. Mechanistically, this result could be explained by interaction of INT6 with ATM, which together with INT6 was recruited to the sites of DNA damage. Finally, INT6 silencing also reduced ubiquitylation events that promote retention of repair proteins at DNA lesions. Accordingly, accumulation of the repair factor BRCA1 was defective in the absence of INT6. Our findings reveal unexpected and striking connections of INT6 with ATM and BRCA1 and suggest that the protective action of INT6 in the onset of breast cancers relies on its involvement in the DNA damage response.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes a technique that supports process participants in making risk-informed decisions, with the aim to reduce the process risks. Risk reduction involves decreasing the likelihood and severity of a process fault from occurring. Given a process exposed to risks, e.g. a financial process exposed to a risk of reputation loss, we enact this process and whenever a process participant needs to provide input to the process, e.g. by selecting the next task to execute or by filling out a form, we prompt the participant with the expected risk that a given fault will occur given the particular input. These risks are predicted by traversing decision trees generated from the logs of past process executions and considering process data, involved resources, task durations and contextual information like task frequencies. The approach has been implemented in the YAWL system and its effectiveness evaluated. The results show that the process instances executed in the tests complete with substantially fewer faults and with lower fault severities, when taking into account the recommendations provided by our technique.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In Virgtel Ltd v Zabusky [2009] QCA 92 the Queensland Court of Appeal considered the scope of an order “as to costs only” within the meaning of s 253 of the Supreme Court Act 1995 (Qld) (‘the Act”). The Court also declined to accept submissions from one of the parties after oral hearing, and made some useful comments which serve as a reminder to practitioners of their obligations in that regard.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

How dance history should be conserved, like any other human event is problematical. This article refers to the dancer’s embodied repertoire of movement and questions how it might connect with the archive, so that the recording of dance work is not taken as fully representative of the work itself. I propose opening out the multi-dimensional configuring that is part of the creation of a dance piece in order to flesh out the bones of the archive. This could be done through engaging with the choreographer and the performers (who are often co-creators of the work) to draw a range of insights and perspectives together.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A straightforward procedure for the acid digestion of geological samples with SiO2 concentrations ranging between about 40 to 80%, is described. A powdered sample (200 mesh) of 500 mg was used and fused with 1000 mg spectroflux at about 1000 degreesC in a platinum crucible. The molten was subsequently digested in an aqueous solution of HNO3 at 100 degreesC. Several systematic digestion procedures were followed using various concentrations of HNO3. It was found that a relationship could be established between the dissolution-time and acid concentration. For an acid concentration of 15% an optimum dissolution-time of under 4 min was recorded. To verify that the dissolutions were complete, they were subjected to rigorous quality control tests. The turbidity and viscosity were examined at different intervals and the results were compared with that of deionised water. No significant change in either parameter was observed. The shelf-life of each solution lasted for several months, after which time polymeric silicic acid formed in some solutions, resulting in the presence of a gelatinous solid. The method is cost effective and is clearly well suited for routine applications on a small scale, especially in laboratories in developing countries. ICP-MS was applied to the determination of 13 Rare Earth Elements and Hf in a set of 107 archaeological samples subjected to the above digestion procedure. The distribution of these elements was examined and the possibility of using the REE's for provenance studies is discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Business Process Management describes a holistic management approach for the systematic design, modeling, execution, validation, monitoring and improvement of organizational business processes. Traditionally, most attention within this community has been given to control-flow aspects, i.e., the ordering and sequencing of business activities, oftentimes in isolation with regards to the context in which these activities occur. In this paper, we propose an approach that allows executable process models to be integrated with Geographic Information Systems. This approach enables process models to take geospatial and other geographic aspects into account in an explicit manner both during the modeling phase and the execution phase. We contribute a structured modeling methodology, based on the well-known Business Process Model and Notation standard, which is formalized by means of a mapping to executable Colored Petri nets. We illustrate the feasibility of our approach by means of a sustainability-focused case example of a process with important ecological concerns.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents a technique for the automated removal of noise from process execution logs. Noise is the result of data quality issues such as logging errors and manifests itself in the form of infrequent process behavior. The proposed technique generates an abstract representation of an event log as an automaton capturing the direct follows relations between event labels. This automaton is then pruned from arcs with low relative frequency and used to remove from the log those events not fitting the automaton, which are identified as outliers. The technique has been extensively evaluated on top of various auto- mated process discovery algorithms using both artificial logs with different levels of noise, as well as a variety of real-life logs. The results show that the technique significantly improves the quality of the discovered process model along fitness, appropriateness and simplicity, without negative effects on generalization. Further, the technique scales well to large and complex logs.