947 resultados para linear programming applications
Resumo:
Many combinatorial problems coming from the real world may not have a clear and well defined structure, typically being dirtied by side constraints, or being composed of two or more sub-problems, usually not disjoint. Such problems are not suitable to be solved with pure approaches based on a single programming paradigm, because a paradigm that can effectively face a problem characteristic may behave inefficiently when facing other characteristics. In these cases, modelling the problem using different programming techniques, trying to ”take the best” from each technique, can produce solvers that largely dominate pure approaches. We demonstrate the effectiveness of hybridization and we discuss about different hybridization techniques by analyzing two classes of problems with particular structures, exploiting Constraint Programming and Integer Linear Programming solving tools and Algorithm Portfolios and Logic Based Benders Decomposition as integration and hybridization frameworks.
Resumo:
One of the most interesting challenge of the next years will be the Air Space Systems automation. This process will involve different aspects as the Air Traffic Management, the Aircrafts and Airport Operations and the Guidance and Navigation Systems. The use of UAS (Uninhabited Aerial System) for civil mission will be one of the most important steps in this automation process. In civil air space, Air Traffic Controllers (ATC) manage the air traffic ensuring that a minimum separation between the controlled aircrafts is always provided. For this purpose ATCs use several operative avoidance techniques like holding patterns or rerouting. The use of UAS in these context will require the definition of strategies for a common management of piloted and piloted air traffic that allow the UAS to self separate. As a first employment in civil air space we consider a UAS surveillance mission that consists in departing from a ground base, taking pictures over a set of mission targets and coming back to the same ground base. During all mission a set of piloted aircrafts fly in the same airspace and thus the UAS has to self separate using the ATC avoidance as anticipated. We consider two objective, the first consists in the minimization of the air traffic impact over the mission, the second consists in the minimization of the impact of the mission over the air traffic. A particular version of the well known Travelling Salesman Problem (TSP) called Time-Dependant-TSP has been studied to deal with traffic problems in big urban areas. Its basic idea consists in a cost of the route between two clients depending on the period of the day in which it is crossed. Our thesis supports that such idea can be applied to the air traffic too using a convenient time horizon compatible with aircrafts operations. The cost of a UAS sub-route will depend on the air traffic that it will meet starting such route in a specific moment and consequently on the avoidance maneuver that it will use to avoid that conflict. The conflict avoidance is a topic that has been hardly developed in past years using different approaches. In this thesis we purpose a new approach based on the use of ATC operative techniques that makes it possible both to model the UAS problem using a TDTSP framework both to use an Air Traffic Management perspective. Starting from this kind of mission, the problem of the UAS insertion in civil air space is formalized as the UAS Routing Problem (URP). For this reason we introduce a new structure called Conflict Graph that makes it possible to model the avoidance maneuvers and to define the arc cost function of the departing time. Two Integer Linear Programming formulations of the problem are proposed. The first is based on a TDTSP formulation that, unfortunately, is weaker then the TSP formulation. Thus a new formulation based on a TSP variation that uses specific penalty to model the holdings is proposed. Different algorithms are presented: exact algorithms, simple heuristics used as Upper Bounds on the number of time steps used, and metaheuristic algorithms as Genetic Algorithm and Simulated Annealing. Finally an air traffic scenario has been simulated using real air traffic data in order to test our algorithms. Graphic Tools have been used to represent the Milano Linate air space and its air traffic during different days. Such data have been provided by ENAV S.p.A (Italian Agency for Air Navigation Services).
Resumo:
This thesis addresses the formulation of a referee assignment problem for the Italian Volleyball Serie A Championships. The problem has particular constraints such as a referee must be assigned to different teams in a given period of times, and the minimal/maximal level of workload for each referee is obtained by considering cost and profit in the objective function. The problem has been solved through an exact method by using an integer linear programming formulation and a clique based decomposition for improving the computing time. Extensive computational experiments on real-world instances have been performed to determine the effectiveness of the proposed approach.
Resumo:
Il presente lavoro trae origine dagli obiettivi e dalle relative misure applicative della riforma dell’OCM zucchero del 2006 e nello specifico dal Piano nazionale per la razionalizzazione e riconversione della produzione bieticolo-saccarifera approvato dal MIPAF nel 2007. Lo studio riguarda la riconversione dello zuccherificio di Finale Emilia (MO), di appartenenza del Gruppo bieticolo-saccarifero Co.Pro.B, in un impianto di generazione di energia elettrica e termica che utilizza biomassa di origine agricola per la combustione diretta. L'alimentazione avviene principalmente dalla coltivazione dedicata del sorgo da fibra (Sorghum bicolor), integrata con risorse agro-forestali. Lo studio mostra la necessità di coltivazione di 4.400 ettari di sorgo da fibra con una produzione annua di circa 97.000 t di prodotto al 75% di sostanza secca necessari per l’alimentazione della centrale a biomassa. L’obiettivo é quello di valutare l’impatto della nuova coltura energetica sul comprensorio agricolo e sulla economia dell’impresa agricola. La metodologia adottata si basa sulla simulazione di modelli aziendali di programmazione lineare che prevedono l’inserimento del sorgo da fibra come coltura energetica nel piano ottimo delle aziende considerate. I modelli predisposti sono stati calibrati su aziende RICA al fine di riprodurre riparti medi reali su tre tipologie dimensionali rappresentative: azienda piccola entro i 20 ha, media da 20 a 50 ha e grande oltre i 50 ha. La superficie di entrata a livello aziendale, se rapportata alla rappresentatività delle aziende dell’area di studio, risulta insufficiente per soddisfare la richiesta di approvvigionamento dell’impianto a biomassa. Infatti con tale incremento la superficie di coltivazione nel comprensorio si attesta sui 2.500 ettari circa contro i 4.400 necessari alla centrale. Lo studio mostra pertanto che occorre un incentivo superiore, di circa 80-90 €/ha, per soddisfare la richiesta della superficie colturale a livello di territorio. A questi livelli, la disponibilità della coltura energetica sul comprensorio risulta circa 9.500 ettari.
Resumo:
Khutoretsky dealt with the problem of maximising a linear utility function (MUF) over the set of short-term equilibria in a housing market by reducing it to a linear programming problem, and suggested a combinatorial algorithm for this problem. Two approaches to the market adjustment were considered: the funding of housing construction and the granting of housing allowances. In both cases, locally optimal regulatory measures can be developed using the corresponding dual prices. The optimal effects (with the regulation expenditures restricted by an amount K) can be found using specialised models based on MUF: a model M1 for choice of the optimum structure of investment in housing construction, and a model M2 for optimum distribution of housing allowances. The linear integer optimisation problems corresponding to these models are initially difficult but can be solved after slight modifications of the parameters. In particular, the necessary modification of K does not exceed the maximum construction cost of one dwelling (for M1) or the maximum size of one housing allowance (for M2). The result is particularly useful since slight modification of K is not essential in practice.
Resumo:
BACKGROUND: Despite recent algorithmic and conceptual progress, the stoichiometric network analysis of large metabolic models remains a computationally challenging problem. RESULTS: SNA is a interactive, high performance toolbox for analysing the possible steady state behaviour of metabolic networks by computing the generating and elementary vectors of their flux and conversions cones. It also supports analysing the steady states by linear programming. The toolbox is implemented mainly in Mathematica and returns numerically exact results. It is available under an open source license from: http://bioinformatics.org/project/?group_id=546. CONCLUSION: Thanks to its performance and modular design, SNA is demonstrably useful in analysing genome scale metabolic networks. Further, the integration into Mathematica provides a very flexible environment for the subsequent analysis and interpretation of the results.
Resumo:
Small-scale farmers in the Chipata District of Zambia rely on their farm fields to grow maize and groundnuts for food security. Cotton production and surplus food security crops are used to generate income to provide for their families. With increasing population pressure, available land has decreased and farmers struggle to provide the necessary food requirements and income to meet their family’s needs. The purpose of the study was to determine how a farmer can best allocate his land to produce maize, groundnuts and cotton when constrained by labor and capital resources to generate the highest potential for food security and financial gains. Data from the 2008-2009 growing season was compiled and analyzed using a linear programming model. The study determined that farmers make the most profit by allocating all additional land and resources to cotton after meeting their minimum food security requirements. The study suggests growing cotton is a beneficial practice for small-scale subsistence farmers to generate income when restricted by limited resources.
Resumo:
In reverse logistics networks, products (e.g., bottles or containers) have to be transported from a depot to customer locations and, after use, from customer locations back to the depot. In order to operate economically beneficial, companies prefer a simultaneous delivery and pick-up service. The resulting Vehicle Routing Problem with Simultaneous Delivery and Pick-up (VRPSDP) is an operational problem, which has to be solved daily by many companies. We present two mixed-integer linear model formulations for the VRPSDP, namely a vehicle-flow and a commodity-flow model. In order to strengthen the models, domain-reducing preprocessing techniques, and effective cutting planes are outlined. Symmetric benchmark instances known from the literature as well as new asymmetric instances derived from real-world problems are solved to optimality using CPLEX 12.1.
Resumo:
In process industries, make-and-pack production is used to produce food and beverages, chemicals, and metal products, among others. This type of production process allows the fabrication of a wide range of products in relatively small amounts using the same equipment. In this article, we consider a real-world production process (cf. Honkomp et al. 2000. The curse of reality – why process scheduling optimization problems are diffcult in practice. Computers & Chemical Engineering, 24, 323–328.) comprising sequence-dependent changeover times, multipurpose storage units with limited capacities, quarantine times, batch splitting, partial equipment connectivity, and transfer times. The planning problem consists of computing a production schedule such that a given demand of packed products is fulfilled, all technological constraints are satisfied, and the production makespan is minimised. None of the models in the literature covers all of the technological constraints that occur in such make-and-pack production processes. To close this gap, we develop an efficient mixed-integer linear programming model that is based on a continuous time domain and general-precedence variables. We propose novel types of symmetry-breaking constraints and a preprocessing procedure to improve the model performance. In an experimental analysis, we show that small- and moderate-sized instances can be solved to optimality within short CPU times.
Resumo:
A patient classification system was developed integrating a patient acuity instrument with a computerized nursing distribution method based on a linear programming model. The system was designed for real-time measurement of patient acuity (workload) and allocation of nursing personnel to optimize the utilization of resources.^ The acuity instrument was a prototype tool with eight categories of patients defined by patient severity and nursing intensity parameters. From this tool, the demand for nursing care was defined in patient points with one point equal to one hour of RN time. Validity and reliability of the instrument was determined as follows: (1) Content validity by a panel of expert nurses; (2) predictive validity through a paired t-test analysis of preshift and postshift categorization of patients; (3) initial reliability by a one month pilot of the instrument in a practice setting; and (4) interrater reliability by the Kappa statistic.^ The nursing distribution system was a linear programming model using a branch and bound technique for obtaining integer solutions. The objective function was to minimize the total number of nursing personnel used by optimally assigning the staff to meet the acuity needs of the units. A penalty weight was used as a coefficient of the objective function variables to define priorities for allocation of staff.^ The demand constraints were requirements to meet the total acuity points needed for each unit and to have a minimum number of RNs on each unit. Supply constraints were: (1) total availability of each type of staff and the value of that staff member (value was determined relative to that type of staff's ability to perform the job function of an RN (i.e., value for eight hours RN = 8 points, LVN = 6 points); (2) number of personnel available for floating between units.^ The capability of the model to assign staff quantitatively and qualitatively equal to the manual method was established by a thirty day comparison. Sensitivity testing demonstrated appropriate adjustment of the optimal solution to changes in penalty coefficients in the objective function and to acuity totals in the demand constraints.^ Further investigation of the model documented: correct adjustment of assignments in response to staff value changes; and cost minimization by an addition of a dollar coefficient to the objective function. ^
Resumo:
Due to the ongoing trend towards increased product variety, fast-moving consumer goods such as food and beverages, pharmaceuticals, and chemicals are typically manufactured through so-called make-and-pack processes. These processes consist of a make stage, a pack stage, and intermediate storage facilities that decouple these two stages. In operations scheduling, complex technological constraints must be considered, e.g., non-identical parallel processing units, sequence-dependent changeovers, batch splitting, no-wait restrictions, material transfer times, minimum storage times, and finite storage capacity. The short-term scheduling problem is to compute a production schedule such that a given demand for products is fulfilled, all technological constraints are met, and the production makespan is minimised. A production schedule typically comprises 500–1500 operations. Due to the problem size and complexity of the technological constraints, the performance of known mixed-integer linear programming (MILP) formulations and heuristic approaches is often insufficient. We present a hybrid method consisting of three phases. First, the set of operations is divided into several subsets. Second, these subsets are iteratively scheduled using a generic and flexible MILP formulation. Third, a novel critical path-based improvement procedure is applied to the resulting schedule. We develop several strategies for the integration of the MILP model into this heuristic framework. Using these strategies, high-quality feasible solutions to large-scale instances can be obtained within reasonable CPU times using standard optimisation software. We have applied the proposed hybrid method to a set of industrial problem instances and found that the method outperforms state-of-the-art methods.
Resumo:
Data compiled within the IMPENSO project. The Impact of ENSO on Sustainable Water Management and the Decision-Making Community at a Rainforest Margin in Indonesia (IMPENSO), http://www.gwdg.de/~impenso, was a German-Indonesian research project (2003-2007) that has studied the impact of ENSO (El Nino-Southern Oscillation) on the water resources and the agricultural production in the PALU RIVER watershed in Central Sulawesi. ENSO is a climate variability that causes serious droughts in Indonesia and other countries of South-East Asia. The last ENSO event occurred in 1997. As in other regions, many farmers in Central Sulawesi suffered from reduced crop yields and lost their livestock. A better prediction of ENSO and the development of coping strategies would help local communities mitigate the impact of ENSO on rural livelihoods and food security. The IMPENSO project deals with the impact of the climate variability ENSO (El Niño Southern Oscillation) on water resource management and the local communities in the Palu River watershed of Central Sulawesi, Indonesia. The project consists of three interrelated sub-projects, which study the local and regional manifestation of ENSO using the Regional Climate Models REMO and GESIMA (Sub-project A), quantify the impact of ENSO on the availability of water for agriculture and other uses, using the distributed hydrological model WaSiM-ETH (Sub-project B), and analyze the socio-economic impact and the policy implications of ENSO on the basis of a production function analysis, a household vulnerability analysis, and a linear programming model (Sub-project C). The models used in the three sub-projects will be integrated to simulate joint scenarios that are defined in collaboration with local stakeholders and are relevant for the design of coping strategies.
Resumo:
Koopman et al. (2014) developed a method to consistently decompose gross exports in value-added terms that accommodate infinite repercussions of international and inter-sector transactions. This provides a better understanding of trade in value added in global value chains than does the conventional gross exports method, which is affected by double-counting problems. However, the new framework is based on monetary input--output (IO) tables and cannot distinguish prices from quantities; thus, it is unable to consider financial adjustments through the exchange market. In this paper, we propose a framework based on a physical IO system, characterized by its linear programming equivalent that can clarify the various complexities relevant to the existing indicators and is proved to be consistent with Koopman's results when the physical decompositions are evaluated in monetary terms. While international monetary tables are typically described in current U.S. dollars, the physical framework can elucidate the impact of price adjustments through the exchange market. An iterative procedure to calculate the exchange rates is proposed, and we also show that the physical framework is also convenient for considering indicators associated with greenhouse gas (GHG) emissions.
Resumo:
Environmental constraints imposed on hydropoweroperation are usually given in the form of minimum environmental flows and maximum and minimum rates of change of flows, or ramp rates. One solution proposed to mitigate the environmental impact caused by the flows discharged by a hydropower plant while reducing the economic impact of the above-mentioned constraints consists in building a re-regulationreservoir, or afterbay, downstream of the power plant. Adding pumpingcapability between the re-regulationreservoir and the main one could contribute both to reducing the size of the re-regulationreservoir, with the consequent environmental improvement, and to improving the economic feasibility of the project, always fulfilling the environmental constraints imposed to hydropoweroperation. The objective of this paper is studying the contribution of a re-regulationreservoir to fulfilling the environmental constraints while reducing the economic impact of said constraints. For that purpose, a revenue-driven optimization model based on mixed integer linear programming is used. Additionally, the advantages of adding pumpingcapability are analysed. In order to illustrate the applicability of the methodology, a case study based on a real hydropower plant is presented
Resumo:
Este trabajo tiene como objetivos la monitorización en tiempo real de la actividad sísmica, tanto próxima como lejana, a partir de los datos sísmicos registrados por una estación de banda ancha, y el desarrollo de un sistema de difusión interactiva de información actualizada de terremotos, destinado al público general. Ambas fuentes de información se mostrarán a través de una Unidad de Visualización denominada “Monitor Sísmico Interactivo”. El registro de los datos sísmicos se realiza utilizando el sensor de tres componentes de la estación sísmica GUD, perteneciente a la Red Digital de Banda Ancha y transmisión digital del Instituto Geográfico Nacional, instalada en la Basílica del Valle de los Caídos, en lalocalidad de Guadarrama (Madrid). En la E.T.S.I. Topografía, Geodesia y Cartografía se ha instalado un ordenador con conexión a Internet, para la recepción y almacenamiento de los datos, y los programas Scream y Drumplot desarrollados por Guralp, necesarios para la monitorización de la señal sísmica en tiempo real. A partir de estos datos, mediante aplicaciones desarrolladas bajo programación Linux y haciendo uso de las herramientas que ofrece el software SAC (Seismic Analysis Code), se genera además un registro gráfico y una película animada de dicha segmentación para cada evento. Se ha configurado un servidor de correo y una cuenta para la recepción de dos tipos de mensajes de correo, enviados desde la sede central del Instituto Geográfico Nacional, con la información de los eventos registrados por GUD una vez revisados: - Mensajes enviados diariamente, con un listado de eventos ocurridos en los 30 últimos días. - Mensajes con la información en cuasi tiempo real de la última alerta sísmica. Se ha desarrollado el programa “saco” para la gestión del correo recibido que analiza la información sísmica, la almacena en ficheros y ejecuta sobre ellos las aplicaciones de dibujo. Estas aplicaciones han sido previamente desarrolladas bajo programación Linux y software GMT (Generic Mapping Tools), y a partir de ellas se generan automáticamente las distintas imágenes que se visualizan en el Monitor Sísmico: un mapa de sismicidad próxima en la Península Ibérica, un mapa de sismicidad lejana en el mundo, un mapa de detalle para localizar y representar la última alerta generada, los listados con la información de los eventos representados en los mapas, los registros gráficos y las películas animadas de dichos sismogramas. Monitor Sísmico Interactivo ha sido desarrollado para ofrecer además la posibilidad de interactuar con la Unidad de Visualización: se ha creado una base de datos para uso científico donde se almacenan todos los eventos registrados por GUD. Así el usuario puede realizar una petición, a través del envío de un mensaje de correo, que le permite visualizar de forma instantánea las imágenes que muestran la información de cualquier terremoto de su interés. ABSTRACT This study is aimed at real-time monitoring of both near and distant seismic activityfrom the seismic data recorded by a broadband seismic station, and the development of an interactive broadcast system of updated information of earthquakes, for the general public. Bothsources of information are displayed through a display unit called "Interactive Seismic Monitor". The seismic data recording is carried out by using the three-component sensor of the GUD seismic station, which belongs to the Digital Network Broadband and digital broadcast of the National Geographic Institute, housed in the Basilica of The Valley of the Fallen, in the town of Guadarrama (Madrid). A computer with Internet connection has been installed in E.T.S.I. Surveying, Geodesy and Cartography for receiving and storing data, together with Scream and Drumplot programs, developed by Guralp, which are necessary for monitoring the real time seismic signal. Based on the data collected, through programming applications developed under Linux system and using the software tools provided by the SAC (Seismic Analysis Code), a chart recorder and an animated gif image of the segmentation for each event are also generated. A mail server and a mail account have been configured for the receipt of two types of email messages, sent from the National Geographic Institute head office, with the information of the events recorded by GUD after being reviewed: - Messages sent daily, providing a list of events in the past 30 days. - Messages containing information on near real-time seismic of the last seismic alert. A program called "saco" has also been developed for handling mail received that analyzes the seismic data, which stores it in files and runs drawing applications on them. These applications have been previously developed under Linux system and software programming GMT (Generic Mapping Tools), and from them different images that are displayed on the Seismic Monitor are automatically generated: a near seismicity Iberian peninsula map, a distant seismicity world map, a detailed map to locate and represent the last seismic alert generated, the lists with the information of the events depicted in the maps,together with the charts and the animated gif image of such seismograms. Interactive Seismic Monitor has been developed to offer any user the possibility of interacting with the display unit: a database has been created for scientific use which stores all the events recorded by GUD. Thus, any user could make a request, by sending an e-mail that allows them to view instantly all the images showing the information of any earthquake of interest on the display unit.