931 resultados para crew-scheduling
Resumo:
In modem hitec industry Advanced Planning and Scheduling (APS) systems provide the basis for e-business solutions towards the suppliers and the customers. One objective of this thesis was to clarify the modem supply chain management with the APS systems and especially concentrate on the area of Collaborative Planning. In order Advanced Planning and Scheduling systems to be complete and usable, user interfaces are needed. Current Visual Basic user interfaces have faced many complaints and arguments from the users as well as from the development team. This thesis is trying to analyze the reasons and causes for the encountered problems and also provide ways to overcome them. The decision has been made to build the new user interfaces to be Web-enabled. Therefore another objective of this thesis was to research and find suitable technologies for building the Web-based user interfaces for Advanced Planning and Scheduling Systems in Nokia Demand/Supply Planning business area. Comparison between the most suitable technologies is made. Usability issues of Web-enabled user interfaces are also covered. The empirical part of the thesis includes design and implementation of a Web-based user interface with the chosen technology for a particular APS module that enables Collaborative Planning with suppliers.
Researching Manufacturing Planning and Control system and Master Scheduling in a manufacturing firm.
Resumo:
The objective of this thesis is to research Manufacturing Planning and Control (MPC) system and Master Scheduling (MS) in a manufacturing firm. The study is conducted at Ensto Finland Corporation, which operates on a field of electrical systems and supplies. The paper consists of theoretical and empirical parts. The empirical part is based on weekly operating at Ensto and includes inter-firm material analysis, learning and meetings. Master Scheduling is an important module of an MPC system, since it is beneficial on transforming strategic production plans based on demand forecasting into operational schedules. Furthermore, capacity planning tools can remarkably contribute to production planning: by Rough-Cut Capacity Planning (RCCP) tool, a MS plan can be critically analyzed in terms of available key resources in real manufacturing environment. Currently, there are remarkable inefficiencies when it comes to Ensto’s practices: the system is not able to take into consideration seasonal demand and react on market changes on time; This can cause significant lost sales. However, these inefficiencies could be eliminated through the appropriate utilization of MS and RCCP tools. To utilize MS and RCCP tools in Ensto’s production environment, further testing in real production environment is required. Moreover, data accuracy, appropriate commitment to adapting and learning the new tools, and continuous developing of functions closely related to MS, such as sales forecasting, need to be ensured.
Resumo:
The objective of this study was to verify the potential of SNAP III (Scheduling and Network Analysis Program) as a support tool for harvesting and wood transport planning in Brazil harvesting subsystem definition and establishment of a compatible route were assessed. Initially, machine operational and production costs were determined in seven subsystems for the study area, and quality indexes, construction and maintenance costs of forest roads were obtained and used as SNAP III program input data. The results showed, that three categories of forest road occurrence were observed in the study area: main, secondary and tertiary which, based on quality index, allowed a medium vehicle speed of about 41, 30 and 24 km/hours and a construction cost of about US$ 5,084.30, US$ 2,275.28 and US$ 1,650.00/km, respectively. The SNAP III program used as a support tool for the planning, was found to have a high potential tool in the harvesting and wood transport planning. The program was capable of defining efficiently, the harvesting subsystem on technical and economical basis, the best wood transport route and the forest road to be used in each period of the horizon planning.
Resumo:
This work contains a series of studies on the optimization of three real-world scheduling problems, school timetabling, sports scheduling and staff scheduling. These challenging problems are solved to customer satisfaction using the proposed PEAST algorithm. The customer satisfaction refers to the fact that implementations of the algorithm are in industry use. The PEAST algorithm is a product of long-term research and development. The first version of it was introduced in 1998. This thesis is a result of a five-year development of the algorithm. One of the most valuable characteristics of the algorithm has proven to be the ability to solve a wide range of scheduling problems. It is likely that it can be tuned to tackle also a range of other combinatorial problems. The algorithm uses features from numerous different metaheuristics which is the main reason for its success. In addition, the implementation of the algorithm is fast enough for real-world use.
Resumo:
The maintenance of electric distribution network is a topical question for distribution system operators because of increasing significance of failure costs. In this dissertation the maintenance practices of the distribution system operators are analyzed and a theory for scheduling maintenance activities and reinvestment of distribution components is created. The scheduling is based on the deterioration of components and the increasing failure rates due to aging. The dynamic programming algorithm is used as a solving method to maintenance problem which is caused by the increasing failure rates of the network. The other impacts of network maintenance like environmental and regulation reasons are not included to the scope of this thesis. Further the tree trimming of the corridors and the major disturbance of the network are not included to the problem optimized in this thesis. For optimizing, four dynamic programming models are presented and the models are tested. Programming is made in VBA-language to the computer. For testing two different kinds of test networks are used. Because electric distribution system operators want to operate with bigger component groups, optimal timing for component groups is also analyzed. A maintenance software package is created to apply the presented theories in practice. An overview of the program is presented.
Resumo:
Formal methods provide a means of reasoning about computer programs in order to prove correctness criteria. One subtype of formal methods is based on the weakest precondition predicate transformer semantics and uses guarded commands as the basic modelling construct. Examples of such formalisms are Action Systems and Event-B. Guarded commands can intuitively be understood as actions that may be triggered when an associated guard condition holds. Guarded commands whose guards hold are nondeterministically chosen for execution, but no further control flow is present by default. Such a modelling approach is convenient for proving correctness, and the Refinement Calculus allows for a stepwise development method. It also has a parallel interpretation facilitating development of concurrent software, and it is suitable for describing event-driven scenarios. However, for many application areas, the execution paradigm traditionally used comprises more explicit control flow, which constitutes an obstacle for using the above mentioned formal methods. In this thesis, we study how guarded command based modelling approaches can be conveniently and efficiently scheduled in different scenarios. We first focus on the modelling of trust for transactions in a social networking setting. Due to the event-based nature of the scenario, the use of guarded commands turns out to be relatively straightforward. We continue by studying modelling of concurrent software, with particular focus on compute-intensive scenarios. We go from theoretical considerations to the feasibility of implementation by evaluating the performance and scalability of executing a case study model in parallel using automatic scheduling performed by a dedicated scheduler. Finally, we propose a more explicit and non-centralised approach in which the flow of each task is controlled by a schedule of its own. The schedules are expressed in a dedicated scheduling language, and patterns assist the developer in proving correctness of the scheduled model with respect to the original one.
Resumo:
With the shift towards many-core computer architectures, dataflow programming has been proposed as one potential solution for producing software that scales to a varying number of processor cores. Programming for parallel architectures is considered difficult as the current popular programming languages are inherently sequential and introducing parallelism is typically up to the programmer. Dataflow, however, is inherently parallel, describing an application as a directed graph, where nodes represent calculations and edges represent a data dependency in form of a queue. These queues are the only allowed communication between the nodes, making the dependencies between the nodes explicit and thereby also the parallelism. Once a node have the su cient inputs available, the node can, independently of any other node, perform calculations, consume inputs, and produce outputs. Data ow models have existed for several decades and have become popular for describing signal processing applications as the graph representation is a very natural representation within this eld. Digital lters are typically described with boxes and arrows also in textbooks. Data ow is also becoming more interesting in other domains, and in principle, any application working on an information stream ts the dataflow paradigm. Such applications are, among others, network protocols, cryptography, and multimedia applications. As an example, the MPEG group standardized a dataflow language called RVC-CAL to be use within reconfigurable video coding. Describing a video coder as a data ow network instead of with conventional programming languages, makes the coder more readable as it describes how the video dataflows through the different coding tools. While dataflow provides an intuitive representation for many applications, it also introduces some new problems that need to be solved in order for data ow to be more widely used. The explicit parallelism of a dataflow program is descriptive and enables an improved utilization of available processing units, however, the independent nodes also implies that some kind of scheduling is required. The need for efficient scheduling becomes even more evident when the number of nodes is larger than the number of processing units and several nodes are running concurrently on one processor core. There exist several data ow models of computation, with different trade-offs between expressiveness and analyzability. These vary from rather restricted but statically schedulable, with minimal scheduling overhead, to dynamic where each ring requires a ring rule to evaluated. The model used in this work, namely RVC-CAL, is a very expressive language, and in the general case it requires dynamic scheduling, however, the strong encapsulation of dataflow nodes enables analysis and the scheduling overhead can be reduced by using quasi-static, or piecewise static, scheduling techniques. The scheduling problem is concerned with nding the few scheduling decisions that must be run-time, while most decisions are pre-calculated. The result is then an, as small as possible, set of static schedules that are dynamically scheduled. To identify these dynamic decisions and to find the concrete schedules, this thesis shows how quasi-static scheduling can be represented as a model checking problem. This involves identifying the relevant information to generate a minimal but complete model to be used for model checking. The model must describe everything that may affect scheduling of the application while omitting everything else in order to avoid state space explosion. This kind of simplification is necessary to make the state space analysis feasible. For the model checker to nd the actual schedules, a set of scheduling strategies are de ned which are able to produce quasi-static schedulers for a wide range of applications. The results of this work show that actor composition with quasi-static scheduling can be used to transform data ow programs to t many different computer architecture with different type and number of cores. This in turn, enables dataflow to provide a more platform independent representation as one application can be fitted to a specific processor architecture without changing the actual program representation. Instead, the program representation is in the context of design space exploration optimized by the development tools to fit the target platform. This work focuses on representing the dataflow scheduling problem as a model checking problem and is implemented as part of a compiler infrastructure. The thesis also presents experimental results as evidence of the usefulness of the approach.
Resumo:
The objective of this project was to introduce a new software product to pulp industry, a new market for case company. An optimization based scheduling tool has been developed to allow pulp operations to better control their production processes and improve both production efficiency and stability. Both the work here and earlier research indicates that there is a potential for savings around 1-5%. All the supporting data is available today coming from distributed control systems, data historians and other existing sources. The pulp mill model together with the scheduler, allows what-if analyses of the impacts and timely feasibility of various external actions such as planned maintenance of any particular mill operation. The visibility gained from the model proves also to be a real benefit. The aim is to satisfy demand and gain extra profit, while achieving the required customer service level. Research effort has been put both in understanding the minimum features needed to satisfy the scheduling requirements in the industry and the overall existence of the market. A qualitative study was constructed to both identify competitive situation and the requirements vs. gaps on the market. It becomes clear that there is no such system on the marketplace today and also that there is room to improve target market overall process efficiency through such planning tool. This thesis also provides better overall understanding of the different processes in this particular industry for the case company.
Resumo:
Operational excellence of individual tramp shipping companies is important in today’s market, where competition is intense, freight revenues are modest and capital costs high due to global financial crisis, and tighter regulatory framework is generating additional costs and challenges to the industry. This thesis concentrates on tramp shipping, where a tramp operator in a form of an individual case company, specialized in short-sea shipping activities in the Baltic Sea region, is searching ways to map their current fleet operations and better understand potential ways to improve the overall routing and scheduling decisions. The research problem is related to tramp fleet planning where several cargoes are carried on board at the same time, which are here systematically referred to as part cargoes. The purpose is to determine the pivotal dimensions and characteristics of these part cargo operations in tramp shipping, and offer both the individual case company and wider research community better understanding of potential risks and benefits related to utilization of part cargo operations. A mixed method research approach is utilized in this research, as the objectives are related to complex, real-life business practices in the field of supply chain management and more specifically, maritime logistics. A quantitative analysis of different voyage scenarios is executed, including alternative voyage legs with varying cost structure and customer involvement. An on-line-based questionnaire designed and prepared by case company’s decision group again provides desired data of predominant attitudes and views of most important industrial customers regarding the part cargo-related operations and potential future utilization of this business model. The results gained from these quantitative methods are complied with qualitative data collection tools, along with suitable secondary data sources. Based on results and logical analysis of different data sources, a framework for characterizing the different aspects of part cargo operations is developed, utilizing both existing research and empirical investigation of the phenomenon. As conclusions, part cargoes have the ability to be part of viable fleet operations, and even increase flexibility among the fleet to a certain extent. Naturally, several hinderers for this development is recognized as well, such as potential issues with information gathering and sharing, inefficient port activities, and increased transit times.
Resumo:
Pictured here are Tony Biernacki, Mary Cumberland, Gwen Francis, Helen Kukasch, Marcia Gillespie, Bev Pearl, Jeniffer Stephenson, Sandy Allen, Sue Auld, and Lisa Librant (Coxie).
Resumo:
From left to right. Dave Commerford (Coach), Lori Abernathy, Tina Dubanow, Heather White, Judy Brunskill, Nancy Tuffin (Coxie), Debbie Coughlin, Yvonne Hojow, Sue Pinder, and Heather Moyer
Resumo:
From left to right. Top: Sean Dowd, Dave Muirhead, Mike Thompson, Rick Cicchine, Mark Thomas, John Ahlstedt, Mark Reynolds, Tom Kent, and Tony Biernacki (Coach). Bottom: Bob Nguyen (Coxie).
Resumo:
Several irrigation treatments were evaluated on Sovereign Coronation table grapes at two sites over a 3-year period in the cool humid Niagara Peninsula of Ontario. Trials were conducted in the Hippie (Beamsville, ON) and the Lambert Vineyards (Niagara-on-the-Lake, ON) in 2003 to 2005 with the objective of assessing the usefulness of the modified Penman-Monteith equation to accurately schedule vine irrigation needs. Data (relative humidity, windspeed, solar radiation, and temperature) required to precisely calculate evapotranspiration (ETq) were downloaded from the Ontario Weather Network. One of two ETq values (either 100 or 150%) were used in combination with one of two crop coefficients (Kc; either fixed at 0.75 or 0.2 to 0.8 based upon increasing canopy volume) to calculate the amount of irrigation water required. Five irrigation treatments were: un irrigated control; (lOOET) X Kc =0.75; 150ET X Kc =0.75; lOOET X Kc =0.2-0.8; 150ET X Kc =0.2-0.8. Transpiration, water potential (v|/), and soil moisture data were collected each growing seasons. Yield component data was collected and berries from each treatment were analyzed for soluble solids (Brix), pH, titratable acidity (TA), anthocyanins, methyl anthranilate (MA), and total volatile esters (TVE). Irrigation showed a substantial positive effect on transpiration rate and soil moisture; the control treatment showed consistently lower transpiration and soil moisture over the 3 seasons. Transpiration appeared accurately reflect Sovereign Coronation grapevines water status. Soil moisture also accurately reflected level of irrigation. Moreover, irrigation showed impact of leaf \|/, which was more negative throughout the 3 seasons for vines that were not irrigated. Irrigation had a substantial positive effect on yield (kg/vine) and its various components (clusters/vine, cluster weight, and berries/cluster) in 2003 and 2005. Berry weights were higher under the irrigated treatments at both sites. Berry weight consistently appeared to be the main factor leading to these increased yields, as inconsistent responses were noted for some yield variables. Soluble solids was highest under the ET150 and ET100 treatments both with Kc at 0.75. Both pH and TA were highest under control treatments in 2003 and 2004, but highest under irrigated treatments in 2005. Anthocyanins and phenols were highest under the control treatments in 2003 and 2004, but highest under irrigated treatments in 2005. MA and TVE were highest under the ET150 treatments. Vine and soil water status measurements (soil moisture, leaf \|/, and transpiration) confirmed that irrigation was required for the summers of 2003 and 2005 due to dry weather in those years. They also partially supported the hypothesis that the Penman-Monteith equation is useful for calculating vineyard water needs. Both ET treatments gave clear evidence that irrigation could be effective in reducing water stress and for improving vine performance, yield and fruit composition. Use of properly scheduled irrigation was beneficial for Sovereign Coronation table grapes in the Niagara region. Findings herein should give growers some strong guidehnes on when, how and how much to irrigate their vineyards.
Resumo:
This qualitative study explored secondary teachers' perceptions of scheduling in relation to pedagogy, curriculum, and observation of student learning. Its objective was to determine the best way to organize the scheduling for the delivery of Ontario's new 4-year curriculum. Six participants were chosen. Two were teaching in a semestered timetable, 1 in a traditional timetable, and 3 had experience in both schedules. Participants related a pressure cooker "lived experience" with weaker students in the semester system experiencing a particularly harsh environment. The inadequate amount of time for review in content-heavy courses, gap scheduling problems, catch-up difficulties for students missing classes, and the fast pace of semestering are identified as factors negatively impacting on these students. Government testing adds to the pressure by shifting teachers' time and attention in the classroom from deeper learning to a superficial coverage of material, from curriculum as lived to curriculum as text to be covered. Scheduling choice should be available in public education to accommodate the needs of all students. Curriculum guidelines need to be revamped to reflect the content that teachers believe is necessary for a successful course delivery. Applied level courses need to be developed for students who are not academically inferior but learn differently.
Resumo:
Caption title.