916 resultados para workflow scheduling
Resumo:
This chapter aims at developing a taxonomic framework to classify the studies on the flexible job shop scheduling problem (FJSP). The FJSP is a generalization of the classical job shop scheduling problem (JSP), which is one of the oldest NP-hard problems. Although various solution methodologies have been developed to obtain good solutions in reasonable time for FSJPs with different objective functions and constraints, no study which systematically reviews the FJSP literature has been encountered. In the proposed taxonomy, the type of study, type of problem, objective, methodology, data characteristics, and benchmarking are the main categories. In order to verify the proposed taxonomy, a variety of papers from the literature are classified. Using this classification, several inferences are drawn and gaps in the FJSP literature are specified. With the proposed taxonomy, the aim is to develop a framework for a broad view of the FJSP literature and construct a basis for future studies.
Resumo:
The authors propose a mathematical model to minimize the project total cost where there are multiple resources constrained by maximum availability. They assume the resources as renewable and the activities can use any subset of resources requiring any quantity from a limited real interval. The stochastic nature is inferred by means of a stochastic work content defined per resource within an activity and following a known distribution and the total cost is the sum of the resource allocation cost with the tardiness cost or earliness bonus in case the project finishes after or before the due date, respectively. The model was computationally implemented relying upon an interchange of two global optimization metaheuristics – the electromagnetism-like mechanism and the evolutionary strategies. Two experiments were conducted testing the implementation to projects with single and multiple resources, and with or without maximum availability constraints. The set of collected results shows good behavior in general and provide a tool to further assist project manager decision making in the planning phase.
Resumo:
Scheduling, job shop, uncertainty, mixed (disjunctive) graph, stability analysis
Resumo:
Report for the scientific sojourn at the University of California at Berkeley between September 2007 to February 2008. The globalization combined with the success of containerization has brought about tremendous increases in the transportation of containers across the world. This leads to an increasing size of container ships which causes higher demands on seaport container terminals and their equipment. In this situation, the success of container terminals resides in a fast transhipment process with reduced costs. For these reasons it is necessary to optimize the terminal’s processes. There are three main logistic processes in a seaport container terminal: loading and unloading of containerships, storage, and reception/deliver of containers from/to the hinterland. Moreover there is an additional process that ensures the interconnection between previous logistic activities: the internal transport subsystem. The aim of this paper is to optimize the internal transport cycle in a marine container terminal managed by straddle carriers, one of the most used container transfer technologies. Three sub-systems are analyzed in detail: the landside transportation, the storage of containers in the yard, and the quayside transportation. The conflicts and decisions that arise from these three subsystems are analytically investigated, and optimization algorithms are proposed. Moreover, simulation has been applied to TCB (Barcelona Container Terminal) to test these algorithms and compare different straddle carrier’s operation strategies, such as single cycle versus double cycle, and different sizes of the handling equipment fleet. The simulation model is explained in detail and the main decision-making algorithms from the model are presented and formulated.
Resumo:
For the execution of the scientific applications, different methods have been proposed to dynamically provide execution environments for such applications that hide the complexity of underlying distributed and heterogeneous infrastructures. Recently virtualization has emerged as a promising technology to provide such environments. Virtualization is a technology that abstracts away the details of physical hardware and provides virtualized resources for high-level scientific applications. Virtualization offers a cost-effective and flexible way to use and manage computing resources. Such an abstraction is appealing in Grid computing and Cloud computing for better matching jobs (applications) to computational resources. This work applies the virtualization concept to the Condor dynamic resource management system by using Condor Virtual Universe to harvest the existing virtual computing resources to their maximum utility. It allows existing computing resources to be dynamically provisioned at run-time by users based on application requirements instead of statically at design-time thereby lay the basis for efficient use of the available resources, thus providing way for the efficient use of the available resources.
Resumo:
El treball final de carrera està compost de l¿anàlisi, disseny i implementació d'una utilitat de Workflow que dóna servei a una web per gestionar les reclamacions i queixes d'una empresa. La part del Workflow s'ha pensat per ser totalment configurable a les particularitats d'un determinat aplicatiu i per donar servei a tantes aplicacions com faci falta simultàniament.L'anàlisi i el disseny s'han basat en la metodologia UML dins el paradigma de l'Orientació a Objectes juntament amb diversos patrons de disseny com pot ser el MVC que ha permès separar completament la lògica de negoci (Workflow) de la presentació (Web). Per les 2 parts: Workflow i web s¿ha fet ús d'una arquitectura distribuïda.
Resumo:
Estudio de estrategias sobre la utilización de un GRID para realizar la gestión administrativa y la puesta en marcha de localizaciones remotas para la colaboración en este tipo de tareas.
Resumo:
In this paper, we are proposing a methodology to determine the most efficient and least costly way of crew pairing optimization. We are developing a methodology based on algorithm optimization on Eclipse opensource IDE using the Java programming language to solve the crew scheduling problems.
Resumo:
Business processes designers take into account the resources that the processes would need, but, due to the variable cost of certain parameters (like energy) or other circumstances, this scheduling must be done when business process enactment. In this report we formalize the energy aware resource cost, including time and usage dependent rates. We also present a constraint programming approach and an auction-based approach to solve the mentioned problem including a comparison of them and a comparison of the proposed algorithms for solving them
Resumo:
We present a polyhedral framework for establishing general structural properties on optimal solutions of stochastic scheduling problems, where multiple job classes vie for service resources: the existence of an optimal priority policy in a given family, characterized by a greedoid (whose feasible class subsets may receive higher priority), where optimal priorities are determined by class-ranking indices, under restricted linear performance objectives (partial indexability). This framework extends that of Bertsimas and Niño-Mora (1996), which explained the optimality of priority-index policies under all linear objectives (general indexability). We show that, if performance measures satisfy partial conservation laws (with respect to the greedoid), which extend previous generalized conservation laws, then the problem admits a strong LP relaxation over a so-called extended greedoid polytope, which has strong structural and algorithmic properties. We present an adaptive-greedy algorithm (which extends Klimov's) taking as input the linear objective coefficients, which (1) determines whether the optimal LP solution is achievable by a policy in the given family; and (2) if so, computes a set of class-ranking indices that characterize optimal priority policies in the family. In the special case of project scheduling, we show that, under additional conditions, the optimal indices can be computed separately for each project (index decomposition). We further apply the framework to the important restless bandit model (two-action Markov decision chains), obtaining new index policies, that extend Whittle's (1988), and simple sufficient conditions for their validity. These results highlight the power of polyhedral methods (the so-called achievable region approach) in dynamic and stochastic optimization.
Resumo:
A crescente complexidade da gestão portuária, a rápida evolução dos meios tecnológicos aliada à enorme demanda que os sistemas de informação têm no seio das organizações como ferramenta de suporte á execução de diferentes processos, faz com que haja uma enorme pressão na procura de novas soluções para um problema colocado já inúmeras vezes: como conceber sistemas, que desenvolvidas e disponibilizadas no mais curto espaço de tempo, sejam impulsionadores de mudanças efectivas nos processos de negócio. No contexto de uma empresa de gestão portuária em que tipicamente há diferentes entidades implicadas nos processos de negócio conceber tais sistemas é um enorme desafio. Neste trabalho pretendemos seguir uma nova abordagem para um Sistema de Gestão Portuária utilizando Tecnologias que permitem implementar a automatização de processos – Sistemas de Gestão de Workflow.
Resumo:
The standard one-machine scheduling problem consists in schedulinga set of jobs in one machine which can handle only one job at atime, minimizing the maximum lateness. Each job is available forprocessing at its release date, requires a known processing timeand after finishing the processing, it is delivery after a certaintime. There also can exists precedence constraints between pairsof jobs, requiring that the first jobs must be completed beforethe second job can start. An extension of this problem consistsin assigning a time interval between the processing of the jobsassociated with the precedence constrains, known by finish-starttime-lags. In presence of this constraints, the problem is NP-hardeven if preemption is allowed. In this work, we consider a specialcase of the one-machine preemption scheduling problem with time-lags, where the time-lags have a chain form, and propose apolynomial algorithm to solve it. The algorithm consist in apolynomial number of calls of the preemption version of the LongestTail Heuristic. One of the applicability of the method is to obtainlower bounds for NP-hard one-machine and job-shop schedulingproblems. We present some computational results of thisapplication, followed by some conclusions.
Resumo:
Most research on single machine scheduling has assumedthe linearity of job holding costs, which is arguablynot appropriate in some applications. This motivates ourstudy of a model for scheduling $n$ classes of stochasticjobs on a single machine, with the objective of minimizingthe total expected holding cost (discounted or undiscounted). We allow general holding cost rates that are separable,nondecreasing and convex on the number of jobs in eachclass. We formulate the problem as a linear program overa certain greedoid polytope, and establish that it issolved optimally by a dynamic (priority) index rule,whichextends the classical Smith's rule (1956) for the linearcase. Unlike Smith's indices, defined for each class, ournew indices are defined for each extended class, consistingof a class and a number of jobs in that class, and yieldan optimal dynamic index rule: work at each time on a jobwhose current extended class has larger index. We furthershow that the indices possess a decomposition property,as they are computed separately for each class, andinterpret them in economic terms as marginal expected cost rate reductions per unit of expected processing time.We establish the results by deploying a methodology recentlyintroduced by us [J. Niño-Mora (1999). "Restless bandits,partial conservation laws, and indexability. "Forthcomingin Advances in Applied Probability Vol. 33 No. 1, 2001],based on the satisfaction by performance measures of partialconservation laws (PCL) (which extend the generalizedconservation laws of Bertsimas and Niño-Mora (1996)):PCL provide a polyhedral framework for establishing theoptimality of index policies with special structure inscheduling problems under admissible objectives, which weapply to the model of concern.
Resumo:
In todays competitive markets, the importance of goodscheduling strategies in manufacturing companies lead to theneed of developing efficient methods to solve complexscheduling problems.In this paper, we studied two production scheduling problemswith sequence-dependent setups times. The setup times areone of the most common complications in scheduling problems,and are usually associated with cleaning operations andchanging tools and shapes in machines.The first problem considered is a single-machine schedulingwith release dates, sequence-dependent setup times anddelivery times. The performance measure is the maximumlateness.The second problem is a job-shop scheduling problem withsequence-dependent setup times where the objective is tominimize the makespan.We present several priority dispatching rules for bothproblems, followed by a study of their performance. Finally,conclusions and directions of future research are presented.
Resumo:
We address the problem of scheduling a multiclass $M/M/m$ queue with Bernoulli feedback on $m$ parallel servers to minimize time-average linear holding costs. We analyze the performance of a heuristic priority-index rule, which extends Klimov's optimal solution to the single-server case: servers select preemptively customers with larger Klimov indices. We present closed-form suboptimality bounds (approximate optimality) for Klimov's rule, which imply that its suboptimality gap is uniformly bounded above with respect to (i) external arrival rates, as long as they stay within system capacity;and (ii) the number of servers. It follows that its relativesuboptimality gap vanishes in a heavy-traffic limit, as external arrival rates approach system capacity (heavy-traffic optimality). We obtain simpler expressions for the special no-feedback case, where the heuristic reduces to the classical $c \mu$ rule. Our analysis is based on comparing the expected cost of Klimov's ruleto the value of a strong linear programming (LP) relaxation of the system's region of achievable performance of mean queue lengths. In order to obtain this relaxation, we derive and exploit a new set ofwork decomposition laws for the parallel-server system. We further report on the results of a computational study on the quality of the $c \mu$ rule for parallel scheduling.