35 resultados para operating systems
Resumo:
With the advent of distributed computer systems with a largely transparent user interface, new questions have arisen regarding the management of such an environment by an operating system. One fertile area of research is that of load balancing, which attempts to improve system performance by redistributing the workload submitted to the system by the users. Early work in this field concentrated on static placement of computational objects to improve performance, given prior knowledge of process behaviour. More recently this has evolved into studying dynamic load balancing with process migration, thus allowing the system to adapt to varying loads. In this thesis, we describe a simulated system which facilitates experimentation with various load balancing algorithms. The system runs under UNIX and provides functions for user processes to communicate through software ports; processes reside on simulated homogeneous processors, connected by a user-specified topology, and a mechanism is included to allow migration of a process from one processor to another. We present the results of a study of adaptive load balancing algorithms, conducted using the aforementioned simulated system, under varying conditions; these results show the relative merits of different approaches to the load balancing problem, and we analyse the trade-offs between them. Following from this study, we present further novel modifications to suggested algorithms, and show their effects on system performance.
Resumo:
The computer systems of today are characterised by data and program control that are distributed functionally and geographically across a network. A major issue of concern in this environment is the operating system activity of resource management for different processors in the network. To ensure equity in load distribution and improved system performance, load balancing is often undertaken. The research conducted in this field so far, has been primarily concerned with a small set of algorithms operating on tightly-coupled distributed systems. More recent studies have investigated the performance of such algorithms in loosely-coupled architectures but using a small set of processors. This thesis describes a simulation model developed to study the behaviour and general performance characteristics of a range of dynamic load balancing algorithms. Further, the scalability of these algorithms are discussed and a range of regionalised load balancing algorithms developed. In particular, we examine the impact of network diameter and delay on the performance of such algorithms across a range of system workloads. The results produced seem to suggest that the performance of simple dynamic policies are scalable but lack the load stability of more complex global average algorithms.
Resumo:
The recent explosive growth in advanced manufacturing technology (AMT) and continued development of sophisticated information technologies (IT) is expected to have a profound effect on the way we design and operate manufacturing businesses. Furthermore, the escalating capital requirements associated with these developments have significantly increased the level of risk associated with initial design, ongoing development and operation. This dissertation has examined the integration of two key sub-elements of the Computer Integrated Manufacturing (CIM) system, namely the manufacturing facility and the production control system. This research has concentrated on the interactions between production control (MRP) and an AMT based production facility. The disappointing performance of such systems has been discussed in the context of a number of potential technological and performance incompatibilities between these two elements. It was argued that the design and selection of operating policies for both is the key to successful integration. Furthermore, policy decisions are shown to play an important role in matching the performance of the total system to the demands of the marketplace. It is demonstrated that a holistic approach to policy design must be adopted if successful integration is to be achieved. It is shown that the complexity of the issues resulting from such an approach required the formulation of a structured design methodology. Such a methodology was subsequently developed and discussed. This combined a first principles approach to the behaviour of system elements with the specification of a detailed holistic model for use in the policy design environment. The methodology aimed to make full use of the `low inertia' characteristics of AMT, whilst adopting a JIT configuration of MRP and re-coupling the total system to the market demands. This dissertation discussed the application of the methodology to an industrial case study and the subsequent design of operational policies. Consequently a novel approach to production control resulted. A central feature of which was a move toward reduced manual intervention in the MRP processing and scheduling logic with increased human involvement and motivation in the management of work-flow on the shopfloor. Experimental results indicated that significant performance advantages would result from the adoption of the recommended policy set.
Resumo:
It is well established that hydrodynamic journal bearings are responsible for self-excited vibrations and have the effect of lowering the critical speeds of rotor systems. The forces within the oil film wedge, generated by the vibrating journal, may be represented by displacement and velocity coefficient~ thus allowing the dynamical behaviour of the rotor to be analysed both for stability purposes and for anticipating the response to unbalance. However, information describing these coefficients is sparse, misleading, and very often not applicable to industrial type bearings. Results of a combined analytical and experimental investigation into the hydrodynamic oil film coefficients operating in the laminar region are therefore presented, the analysis being applied to a 120 degree partial journal bearing having a 5.0 in diameter journal and a LID ratio of 1.0. The theoretical analysis shows that for this type of popular bearing, the eight linearized coefficients do not accurately describe the behaviour of the vibrating journal based on the theory of small perturbations, due to them being masked by the presence of nonlinearity. A method is developed using the second order terms of Taylor expansion whereby design charts are provided which predict the twentyeight force coefficients for both aligned, and for varying amounts of journal misalignment. The resulting non-linear equations of motion are solved using a modified Newton-Raphson method whereby the whirl trajectories are obtained, thus providing a physical appreciation of the bearing characteristics under dynamically loaded conditions.
Resumo:
This study is concerned with several proposals concerning multiprocessor systems and with the various possible methods of evaluating such proposals. After a discussion of the advantages and disadvantages of several performance evaluation tools, the author decides that simulation is the only tool powerful enough to develop a model which would be of practical use, in the design, comparison and extension of systems. The main aims of the simulation package developed as part of this study are cost effectiveness, ease of use and generality. The methodology on which the simulation package is based is described in detail. The fundamental principles are that model design should reflect actual systems design, that measuring procedures should be carried out alongside design that models should be well documented and easily adaptable and that models should be dynamic. The simulation package itself is modular, and in this way reflects current design trends. This approach also aids documentation and ensures that the model is easily adaptable. It contains a skeleton structure and a library of segments which can be added to or directly swapped with segments of the skeleton structure, to form a model which fits a user's requirements. The study also contains the results of some experimental work carried out using the model, the first part of which tests• the model's capabilities by simulating a large operating system, the ICL George 3 system; the second part deals with general questions and some of the many proposals concerning multiprocessor systems.
Resumo:
Communication and portability are the two main problems facing the user. An operating system, called PORTOS, was developed to solve these problems for users on dedicated microcomputer systems. Firstly, an interface language was defined, according to the anticipated requirements and behaviour of its potential users. Secondly, the PORTOS operating system was developed as a processor for this language. The system is currently running on two minicomputers of highly different architectures. PORTOS achieves its portability through its high-level design, and implementation in CORAL66. The interface language consists of a set of user cotnmands and system responses. Although only a subset has been implemented, owing to time and manpower constraints, promising results were achieved regarding the usability of the language, and its portability.
Resumo:
This thesis deals with the problems associated with the planning and control of production, with particular reference to a small aluminium die casting company. The main problem areas were identified as: (a) A need to be able to forecast the customers demands upon the company's facilities. (b) A need to produce a manufacturing programme in which the output of the foundry (or die casting section) was balanced with the available capacity in the machine shop. (c) The need to ensure that the resultant system enabled the company's operating budget to have a reasonable chance of being achieved. At the commencement of the research work the major customers were members of the automobile industry and had their own system of forecasting, from which they issued manufacturing schedules to their component suppliers, The errors in the forecast were analysed and the distributions noted. Using these distributions the customer's forecast was capable of being modified to enable his final demand to be met with a known degree of confidence. Before a manufacturing programme could be developed the actual manufacturing system had to be reviewed and it was found that as with many small companies there was a remarkable lack of formal control and written data. Relevant data with regards to the component and the manufacturing process had therefore to be collected and analysed. The foundry process was fixed but the secondary machining operations were analysed by a technique similar to Component Flow Analysis and as a result the machines were arranged in a series of flow lines. A system of manual production control was proposed and for comparison, a local computer bureau was approached and a system proposed incorporating the production of additional management information. These systems are compared and the relative merits discussed and a proposal made for implementation.
Resumo:
Drying is an important unit operation in process industry. Results have suggested that the energy used for drying has increased from 12% in 1978 to 18% of the total energy used in 1990. A literature survey of previous studies regarding overall drying energy consumption has demonstrated that there is little continuity of methods and energy trends could not be established. In the ceramics, timber and paper industrial sectors specific energy consumption and energy trends have been investigated by auditing drying equipment. Ceramic products examined have included tableware, tiles, sanitaryware, electrical ceramics, plasterboard, refractories, bricks and abrasives. Data from industry has shown that drying energy has not varied significantly in the ceramics sector over the last decade, representing about 31% of the total energy consumed. Information from the timber industry has established that radical changes have occurred over the last 20 years, both in terms of equipment and energy utilisation. The energy efficiency of hardwood drying has improved by 15% since the 1970s, although no significant savings have been realised for softwood. A survey estimating the energy efficiency and operating characteristics of 192 paper dryer sections has been conducted. Drying energy was found to increase to nearly 60% of the total energy used in the early 1980s, but has fallen over the last decade, representing 23% of the total in 1993. These results have demonstrated that effective energy saving measures, such as improved pressing and heat recovery, have been successfully implemented since the 1970s. Artificial neural networks have successfully been applied to model process characteristics of microwave and convective drying of paper coated gypsum cove. Parameters modelled have included product moisture loss, core gypsum temperature and quality factors relating to paper burning and bubbling defects. Evaluation of thermal and dielectric properties have highlighted gypsum's heat sensitive characteristics in convective and electromagnetic regimes. Modelling experimental data has shown that the networks were capable of simulating drying process characteristics to a high degree of accuracy. Product weight and temperature were predicted to within 0.5% and 5C of the target data respectively. Furthermore, it was demonstrated that the underlying properties of the data could be predicted through a high level of input noise.
Resumo:
Using a hydraulic equipment manufacturing plant as the case study, this work explores the problems of systems integration in manufacturing systems design, stressing the behavioural aspects of motivation and participation, and the constraints involved in the proper consideration of the human sub-system. The need for a simple manageable modular organisation structure is illustrated, where it is shown, by reference to systems theory, how a business can be split into semi-autonomous operating units. The theme is the development of a manufacturing system based on an analysis of the business, its market, product, technology and constraints, coupled with a critical survey of modern management literature to develop an integrated systems design to suit a specific company in the current social environment. Society currently moves through a socio-technical revolution with man seeking higher levels of motivation. The transitory environment from an autocratic/paternalistic to a participative operating mode demands systems parameters only found to a limited extent in manufacturing systems today. It is claimed, that modern manufacturing systems design needs to be based on group working, job enrichment, delegation of decision making and reduced job monotony. The analysis shows how negative aspects of cellular manufacture such as lack of flexibility and poor fixed asset utilisation are relatively irrelevant and misleading in the broader context of the need to come to terms with the social stresses imposed on a company operating in the industrial environment of the present and the immediate future.
Resumo:
Renewable non-edible plant oils such as jatropha and karanj have potential to substitute fossil diesel fuels in CI engines. A multi-cylinder water cooled IDI type CI engine has been tested with jatropha and karanj oils and comparisons made against fossil diesel. The physical and chemical properties of the three fuels were measured to investigate the suitability of jatropha and karanj oils as fuels for CI engines. The engine cooling water circuit and fuel supply systems were modified such that hot jacket water preheated the neat plant oil prior to injection. Between jatropha and karanj there was little difference in the performance, emission and combustion results. Compared to fossil diesel, the brake specific fuel consumption on volume basis was around 3% higher for the plant oils and the brake thermal efficiency was almost similar. Jatropha and karanj operation resulted in higher CO 2 and NO x emissions by 7% and 8% respectively, as compared to diesel. The cylinder gas pressure diagram showed stable engine operation with both plant oils. At full load, the plant oils gave around 3% higher peak cylinder pressure than fossil diesel. With the plant oils, cumulative heat release was smaller at low load and almost similar at full load, compared to diesel. At full load, the plant oils exhibited 5% shorter combustion duration. The study concludes that the IDI type CI engine can be efficiently operated with neat jatropha (or karanj) oil preheated by jacket water, after small modifications of the engine cooling and fuel supply circuits. © 2012 Elsevier Ltd.
Resumo:
In this letter, we numerically demonstrate that the use of inline nonlinear optical loop mirrors in strongly dispersion-managed transmission systems dominated by pulse distortion and amplitude noise can achieve all-optical passive 2R regeneration of a 40-Gb/s return-to-zero data stream. We define the tolerance limits of this result to the parameters of the input pulses.
Resumo:
Desalination of brackish groundwater (BW) is an effective approach to augment water supply, especially for inland regions that are far from seawater resources. Brackish water reverse osmosis (BWRO) desalination is still subject to intensive energy consumption compared to the theoretical minimum energy demand. Here, we review some of the BWRO plants with various system arrangements. We look at how to minimize energy demands, as these contribute considerably to the cost of desalinated water. Different configurations of BWRO system have been compared from the view point of normalized specific energy consumption (SEC). Analysis is made at theoretical limits. The SEC reduction of BWRO can be achieved by (i) increasing number of stages, (ii) using an energy recovery device (ERD), or (iii) operating the BWRO in batch mode or closed circuit mode. Application of more stages not only reduces SEC but also improves water recovery. However, this improvement is less pronounced when the number of stages exceeds four. Alternatively and more favourably, the BWRO system can be operated in Closed Circuit Desalination (CCD) mode and gives a comparative SEC to that of the 3-stage system with a recovery ratio of 80%. A further reduction of about 30% in SEC can be achieved through batch-RO operation. Moreover, the costly ERDs and booster pumps are avoided with both CCD and batch-RO, thus furthering the effectiveness of lowering the costs of these innovative approaches. © 2012 by the authors.
Resumo:
This paper proposes a novel design of optical filters based on a cascade of tailored fiber Bragg gratings (FBGs) operating in the transmission regime. As an example of the application of the proposed general technique, ultranarrow optical vestigial sideband (VSB) filtering based on two FBGs operating in the transmission regime was examined. This design can be easily implemented by writing FBG-based filters for each wavelengthdivision-multiplexing channel before multiplexing. © 2006 IEEE.
Resumo:
This work is part of a bigger project which aims to research the potential development of commercial opportunities for the re-use of batteries after their use in low carbon vehicles on an electricity grid or microgrid system. There are three main revenue streams (peak load lopping on the distribution Network to allow for network re-enforcement deferral, National Grid primary/ secondary/ high frequency response, customer energy management optimization). These incomes streams are dependent on the grid system being present. However, there is additional opportunity to be gained from also using these batteries to provide UPS backup when the grid is no longer present. Most UPS or ESS on the market use new batteries in conjunction with a two level converter interface. This produces a reliable backup solution in the case of loss of mains power, but may be expensive to implement. This paper introduces a modular multilevel cascade converter (MMCC) based ESS using second-life batteries for use on a grid independent industrial plant without any additional onsite generator as a potentially cheaper alternative. The number of modules has been designed for a given reliability target and these modules could be used to minimize/eliminate the output filter. An appropriate strategy to provide voltage and frequency control in a grid independent system is described and simulated under different disturbance conditions such as load switching, fault conditions or a large motor starting. A comparison of the results from the modular topology against a traditional two level converter is provided to prove similar performance criteria. The proposed ESS and control strategy is an acceptable way of providing backup power in the event of loss of grid. Additional financial benefit to the customer may be obtained by using a second life battery in this way.
Resumo:
This thesis describes advances in the characterisation, calibration and data processing of optical coherence tomography (OCT) systems. Femtosecond (fs) laser inscription was used for producing OCT-phantoms. Transparent materials are generally inert to infra-red radiations, but with fs lasers material modification occurs via non-linear processes when the highly focused light source interacts with the materials. This modification is confined to the focal volume and is highly reproducible. In order to select the best inscription parameters, combination of different inscription parameters were tested, using three fs laser systems, with different operating properties, on a variety of materials. This facilitated the understanding of the key characteristics of the produced structures with the aim of producing viable OCT-phantoms. Finally, OCT-phantoms were successfully designed and fabricated in fused silica. The use of these phantoms to characterise many properties (resolution, distortion, sensitivity decay, scan linearity) of an OCT system was demonstrated. Quantitative methods were developed to support the characterisation of an OCT system collecting images from phantoms and also to improve the quality of the OCT images. Characterisation methods include the measurement of the spatially variant resolution (point spread function (PSF) and modulation transfer function (MTF)), sensitivity and distortion. Processing of OCT data is a computer intensive process. Standard central processing unit (CPU) based processing might take several minutes to a few hours to process acquired data, thus data processing is a significant bottleneck. An alternative choice is to use expensive hardware-based processing such as field programmable gate arrays (FPGAs). However, recently graphics processing unit (GPU) based data processing methods have been developed to minimize this data processing and rendering time. These processing techniques include standard-processing methods which includes a set of algorithms to process the raw data (interference) obtained by the detector and generate A-scans. The work presented here describes accelerated data processing and post processing techniques for OCT systems. The GPU based processing developed, during the PhD, was later implemented into a custom built Fourier domain optical coherence tomography (FD-OCT) system. This system currently processes and renders data in real time. Processing throughput of this system is currently limited by the camera capture rate. OCTphantoms have been heavily used for the qualitative characterization and adjustment/ fine tuning of the operating conditions of OCT system. Currently, investigations are under way to characterize OCT systems using our phantoms. The work presented in this thesis demonstrate several novel techniques of fabricating OCT-phantoms and accelerating OCT data processing using GPUs. In the process of developing phantoms and quantitative methods, a thorough understanding and practical knowledge of OCT and fs laser processing systems was developed. This understanding leads to several novel pieces of research that are not only relevant to OCT but have broader importance. For example, extensive understanding of the properties of fs inscribed structures will be useful in other photonic application such as making of phase mask, wave guides and microfluidic channels. Acceleration of data processing with GPUs is also useful in other fields.