39 resultados para Many-to-many-assignment problem

em Aston University Research Archive


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The existing assignment problems for assigning n jobs to n individuals are limited to the considerations of cost or profit measured as crisp. However, in many real applications, costs are not deterministic numbers. This paper develops a procedure based on Data Envelopment Analysis method to solve the assignment problems with fuzzy costs or fuzzy profits for each possible assignment. It aims to obtain the points with maximum membership values for the fuzzy parameters while maximizing the profit or minimizing the assignment cost. In this method, a discrete approach is presented to rank the fuzzy numbers first. Then, corresponding to each fuzzy number, we introduce a crisp number using the efficiency concept. A numerical example is used to illustrate the usefulness of this new method. © 2012 Operational Research Society Ltd. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper re-assesses three independently developed approaches that are aimed at solving the problem of zero-weights or non-zero slacks in Data Envelopment Analysis (DEA). The methods are weights restricted, non-radial and extended facet DEA models. Weights restricted DEA models are dual to envelopment DEA models with restrictions on the dual variables (DEA weights) aimed at avoiding zero values for those weights; non-radial DEA models are envelopment models which avoid non-zero slacks in the input-output constraints. Finally, extended facet DEA models recognize that only projections on facets of full dimension correspond to well defined rates of substitution/transformation between all inputs/outputs which in turn correspond to non-zero weights in the multiplier version of the DEA model. We demonstrate how these methods are equivalent, not only in their aim but also in the solutions they yield. In addition, we show that the aforementioned methods modify the production frontier by extending existing facets or creating unobserved facets. Further we propose a new approach that uses weight restrictions to extend existing facets. This approach has some advantages in computational terms, because extended facet models normally make use of mixed integer programming models, which are computationally demanding.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Purpose: Considering the UK's limited capacity for waste disposal (particularly for hazardous/radiological waste) there is growing focus on waste avoidance and minimisation to lower the volumes of waste being sent to disposal. The hazardous nature of some waste can complicate its management and reduction. To address this problem there was a need for a decision making methodology to support managers in the nuclear industry as they identify ways to reduce the production of avoidable hazardous waste. The methodology we developed is called Waste And Sourcematter Analysis (WASAN). A methodology that begins the thought process at the pre-waste creation stage (i.e. Avoid). Design/methodology/ approach: The methodology analyses the source of waste, the production of waste inside the facility, the knock on effects from up/downstream facilities on waste production, and the down-selection of waste minimisation actions/options. WASAN has been applied to case studies with licencees and this paper reports on one such case study - the management of plastic bags in Enriched Uranium Residues Recovery Plant (EURRP) at Springfields (UK) where it was used to analyse the generation of radioactive plastic bag waste. Findings: Plastic bags are used in EURRP as a strategy to contain hazard. Double bagging of materials led to the proliferation of these bags as a waste. The paper reports on the philosophy behind WASAN, the application of the methodology to this problem, the results, and views from managers in EURRP. Originality/value: This paper presents WASAN as a novel methodology for analyzing the minimization of avoidable hazardous waste. This addresses an issue that is important to many industries e.g. where legislation enforces waste minimization, where waste disposal costs encourage waste avoidance, or where plant design can reduce waste. The paper forms part of the HSE Nuclear Installations Inspectorate's desire to work towards greater openness and transparency in its work and the development in its thinking.© Crown Copyright 2011.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Physical distribution plays an imporant role in contemporary logistics management. Both satisfaction level of of customer and competitiveness of company can be enhanced if the distribution problem is solved optimally. The multi-depot vehicle routing problem (MDVRP) belongs to a practical logistics distribution problem, which consists of three critical issues: customer assignment, customer routing, and vehicle sequencing. According to the literatures, the solution approaches for the MDVRP are not satisfactory because some unrealistic assumptions were made on the first sub-problem of the MDVRP, ot the customer assignment problem. To refine the approaches, the focus of this paper is confined to this problem only. This paper formulates the customer assignment problem as a minimax-type integer linear programming model with the objective of minimizing the cycle time of the depots where setup times are explicitly considered. Since the model is proven to be MP-complete, a genetic algorithm is developed for solving the problem. The efficiency and effectiveness of the genetic algorithm are illustrated by a numerical example.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Data Envelopment Analysis (DEA) is one of the most widely used methods in the measurement of the efficiency and productivity of Decision Making Units (DMUs). DEA for a large dataset with many inputs/outputs would require huge computer resources in terms of memory and CPU time. This paper proposes a neural network back-propagation Data Envelopment Analysis to address this problem for the very large scale datasets now emerging in practice. Neural network requirements for computer memory and CPU time are far less than that needed by conventional DEA methods and can therefore be a useful tool in measuring the efficiency of large datasets. Finally, the back-propagation DEA algorithm is applied to five large datasets and compared with the results obtained by conventional DEA.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Purpose – This paper sets out to study a production-planning problem for printed circuit board (PCB) assembly. A PCB assembly company may have a number of assembly lines for production of several product types in large volume. Design/methodology/approach – Pure integer linear programming models are formulated for assigning the product types to assembly lines, which is the line assignment problem, with the objective of minimizing the total production cost. In this approach, unrealistic assignment, which was suffered by previous researchers, is avoided by incorporating several constraints into the model. In this paper, a genetic algorithm is developed to solve the line assignment problem. Findings – The procedure of the genetic algorithm to the problem and a numerical example for illustrating the models are provided. It is also proved that the algorithm is effective and efficient in dealing with the problem. Originality/value – This paper studies the line assignment problem arising in a PCB manufacturing company in which the production volume is high.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Purpose – Increasing turnover of frontline staff in call centres is detrimental to the delivery of quality service to customers. This paper aims to present the context for the rapid growth of the business process outsourcing (BPO) sector in India, and to address a critical issue faced by call centre organisations in this sector – the high employee turnover. Design/methodology/approach – Following a triangulation approach, two separate empirical investigations are conducted to examine various aspects of high labour turnover rates in the call centre sector in India. Study one examines the research issue via 51 in-depth interviews in as many units. Study two reports results from a questionnaire survey with 204 frontline agents across 11 call centres regarding employee turnover. Findings – This research reveals a range of reasons – from monotonous work, stressful work environment, adverse working conditions, lack of career development opportunities; to better job opportunities elsewhere, which emerge as the key causes of increasing attrition rates in the Indian call centre industry. Research limitations/implications – The research suggests that there are several issues that need to be handled carefully by management of call centres in India to overcome the problem of increasing employee turnover, and that this also demands support from the Indian government. Originality/value – The contributions of this study untangle the issues underlying a key problem in the call centre industry, i.e. employee turnover in the Indian call centre industry context. Adopting an internal marketing approach, it provides useful information for both academics and practitioners and suggests internal marketing interventions, and avenues for future research to combat the problem of employee turnover.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of the investigation was to study the problem of colonization of shipboard fuel systems and to examine the effect of a number of environmental factors on microbial growth and survival in order to find potential preservative treatments. A variety of microbial species were isolated from samples taken from fuel storage tanks. Bacteria were more numerous than yeasts or fungi and most microorganisms were found at the fuel/water interface. 1he salinity, pH and phosphate concentration of some water bottoms were characteristic of sea water. Others were brackish, acidic and varied in phosphate content. Microorganisms were cultured under a number of environmental conditions. After prolonged incubation, the inoculum size had no effect on the final biomass of Cladosporium resinae but the time required to achieve the final mass decreased with increasing spore number. Undecane supported better growth of the fungus than diesel fuel and of four types of diesel fuel, two allowed more profuse growth. With sea water as the aqueous phase, a number of isolates were inhibited but the addition of nutrients allowed the development of many of the organisms. Agitation increased the growth of C. resinae on glucose but inhibited it on hydrocarbons. The optimum temperature fgr growth of C. resinae on surface culture lay between 25º C and 30º C and growth was evident at 5º C but not at 45º C. In aqueous suspension, 90% of spores were inactivated in around 60 hours at 45ºC and the same proportion of spores of C. resinae and Penicillium corylophilum were destroyed after about 30 seconds at 65ºC. The majority of bacteria and all yeasts in a water bottom sample were killed within 10 seconds at this temperature. An increase in the concentration of an organo-boron compound caused more rapid inactivation of C. resinae spores and raising the temperature from 25ºC to 45°C significantly enhanced the potency of the biocide.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Many planning and control tools, especially network analysis, have been developed in the last four decades. The majority of them were created in military organization to solve the problem of planning and controlling research and development projects. The original version of the network model (i.e. C.P.M/PERT) was transplanted to the construction industry without the consideration of the special nature and environment of construction projects. It suited the purpose of setting up targets and defining objectives, but it failed in satisfying the requirement of detailed planning and control at the site level. Several analytical and heuristic rules based methods were designed and combined with the structure of C.P.M. to eliminate its deficiencies. None of them provides a complete solution to the problem of resource, time and cost control. VERT was designed to deal with new ventures. It is suitable for project evaluation at the development stage. CYCLONE, on the other hand, is concerned with the design and micro-analysis of the production process. This work introduces an extensive critical review of the available planning techniques and addresses the problem of planning for site operation and control. Based on the outline of the nature of site control, this research developed a simulation based network model which combines part of the logics of both VERT and CYCLONE. Several new nodes were designed to model the availability and flow of resources, the overhead and operating cost and special nodes for evaluating time and cost. A large software package is written to handle the input, the simulation process and the output of the model. This package is designed to be used on any microcomputer using MS-DOS operating system. Data from real life projects were used to demonstrate the capability of the technique. Finally, a set of conclusions are drawn regarding the features and limitations of the proposed model, and recommendations for future work are outlined at the end of this thesis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In recent years the optical domain has been traditionally reserved for node-to-node transmission with the processing and switching achieved entirely in the electrical domain. However, with the constantly increasing demand for bandwidth and the resultant increase in transmission speeds, there is a very real fear that current electronic technology as used for processing will not be able to cope with future demands. Fuelled by this requirement for faster processing speeds, considerable research is currently being carried out into the potential of All-optical processing. One of the fundamental obstacles in realising All-optical processing is the requirement for All-optical buffering. Without all-optical buffers it is extremely difficult to resolve situations such as contention and congestion. Many devices have been proposed to solve this problem however none of them provide the perfect solution. The subject of this research is to experimentally demonstrate a novel all-optical memory device. Unlike many previously demonstrated optical storage devices the device under consideration utilises only a single loop mirror and a single SOA as its switch, whilst providing full regenerative capabilities required for long-term storage. I will explain some of the principles and characteristics of the device, which will then be experimentally demonstrated. The device configuration will then be studied and investigated as to its suitability for Hybrid Integrated Technology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The activities of many mammalian membrane proteins including G-protein coupled receptors are cholesterol-dependent. Unlike higher eukaryotes, yeast do not make cholesterol. Rather they make a related molecule called ergosterol. As cholesterol and ergosterol are biologically non-equivalent, the potential of yeast as hosts for overproducing mammalian membrane proteins has never been fully realised. To address this problem, we are trying to engineer a novel strain of Saccharomyces cerevisiae in which the cholesterol biosynthetic pathway of mammalian cells has been fully reconstituted. Thus far, we have created a modified strain that makes cholesterol-like sterols which has an increased capacity to make G-protein coupled receptors compared to control yeast.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A nature inspired decentralised multi-agent algorithm is proposed to solve a problem of distributed task selection in which cities produce and store batches of different mail types. Agents must collect and process the mail batches, without a priori knowledge of the available mail at the cities or inter-agent communication. In order to process a different mail type than the previous one, agents must undergo a change-over during which it remains inactive. We propose a threshold based algorithm in order to maximise the overall efficiency (the average amount of mail collected). We show that memory, i.e. the possibility for agents to develop preferences for certain cities, not only leads to emergent cooperation between agents, but also to a significant increase in efficiency (above the theoretical upper limit for any memoryless algorithm), and we systematically investigate the influence of the various model parameters. Finally, we demonstrate the flexibility of the algorithm to changes in circumstances, and its excellent scalability.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In many real applications of Data Envelopment Analysis (DEA), the decision makers have to deteriorate some inputs and some outputs. This could be because of limitation of funds available. This paper proposes a new DEA-based approach to determine highest possible reduction in the concern input variables and lowest possible deterioration in the concern output variables without reducing the efficiency in any DMU. A numerical example is used to illustrate the problem. An application in banking sector with limitation of IT investment shows the usefulness of the proposed method. © 2010 Elsevier Ltd. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Completing projects faster than the normal duration is always a challenge to the management of any project, as it often demands many paradigm shifts. Opportunities of globalization, competition from private sectors and multinationals force the management of public sector organizations in the Indian petroleum sector to take various aggressive strategies to maintain their profitability. Constructing infrastructure for handling petroleum products is one of them. Moreover, these projects are required to be completed in faster duration compared to normal schedules to remain competitive, to get faster return on investment, and to give longer project life. However, using conventional tools and techniques of project management, it is impossible to handle the problem of reducing the project duration from a normal period. This study proposes the use of concurrent engineering in managing projects for radically reducing project duration. The phases of the project are accomplished concurrently/simultaneously instead of in a series. The complexities that arise in managing projects are tackled through restructuring project organization, improving management commitment, strengthening project-planning activities, ensuring project quality, managing project risk objectively and integrating project activities through management information systems. These would not only ensure completion of projects in fast track, but also improve project effectiveness in terms of quality, cost effectiveness, team building, etc. and in turn overall productivity of the project organization would improve.