904 resultados para Optimal Linear Control


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recent optimizations of NMR spectroscopy have focused their attention on innovations in new hardware, such as novel probes and higher field strengths. Only recently has the potential to enhance the sensitivity of NMR through data acquisition strategies been investigated. This thesis has focused on the practice of enhancing the signal-to-noise ratio (SNR) of NMR using non-uniform sampling (NUS). After first establishing the concept and exact theory of compounding sensitivity enhancements in multiple non-uniformly sampled indirect dimensions, a new result was derived that NUS enhances both SNR and resolution at any given signal evolution time. In contrast, uniform sampling alternately optimizes SNR (t < 1.26T2) or resolution (t~3T2), each at the expense of the other. Experiments were designed and conducted on a plant natural product to explore this behavior of NUS in which the SNR and resolution continue to improve as acquisition time increases. Possible absolute sensitivity improvements of 1.5 and 1.9 are possible in each indirect dimension for matched and 2x biased exponentially decaying sampling densities, respectively, at an acquisition time of ¿T2. Recommendations for breaking into the linear regime of maximum entropy (MaxEnt) are proposed. Furthermore, examination into a novel sinusoidal sampling density resulted in improved line shapes in MaxEnt reconstructions of NUS data and comparable enhancement to a matched exponential sampling density. The Absolute Sample Sensitivity derived and demonstrated here for NUS holds great promise in expanding the adoption of non-uniform sampling.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE: Current pulsatile ventricular assist devices operate asynchronous with the left ventricle in fixed-rate or fill-to-empty modes because electrocardiogram-triggered modes have been abandoned. We hypothesize that varying the ejection delay in the synchronized mode yields more precise control of hemodynamics and left ventricular loading. This allows for a refined management that may be clinically beneficial. METHODS: Eight sheep received a Thoratec paracorporeal ventricular assist device (Thoratec Corp, Pleasanton, Calif) via ventriculo-aortic cannulation. Left ventricular pressure and volume, aortic pressure, pulmonary flow, pump chamber pressure, and pump inflow and outflow were recorded. The pump was driven by a clinical pneumatic drive unit (Medos Medizintechnik AG, Stolberg, Germany) synchronously with the native R-wave. The start of pump ejection was delayed between 0% and 100% of the cardiac period in 10% increments. For each of these delays, hemodynamic variables were compared with baseline data using paired t tests. RESULTS: The location of the minimum of stroke work was observed at a delay of 10% (soon after aortic valve opening), resulting in a median of 43% reduction in stroke work compared with baseline. Maximum stroke work occurred at a median delay of 70% with a median stroke work increase of 11% above baseline. Left ventricular volume unloading expressed by end-diastolic volume was most pronounced for copulsation (delay 0%). CONCLUSIONS: The timing of pump ejection in synchronized mode yields control over left ventricular energetics and can be a method to achieve gradual reloading of a recoverable left ventricle. The traditionally suggested counterpulsation is not optimal in ventriculo-aortic cannulation when maximum unloading is desired.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Linear programs, or LPs, are often used in optimization problems, such as improving manufacturing efficiency of maximizing the yield from limited resources. The most common method for solving LPs is the Simplex Method, which will yield a solution, if one exists, but over the real numbers. From a purely numerical standpoint, it will be an optimal solution, but quite often we desire an optimal integer solution. A linear program in which the variables are also constrained to be integers is called an integer linear program or ILP. It is the focus of this report to present a parallel algorithm for solving ILPs. We discuss a serial algorithm using a breadth-first branch-and-bound search to check the feasible solution space, and then extend it into a parallel algorithm using a client-server model. In the parallel mode, the search may not be truly breadth-first, depending on the solution time for each node in the solution tree. Our search takes advantage of pruning, often resulting in super-linear improvements in solution time. Finally, we present results from sample ILPs, describe a few modifications to enhance the algorithm and improve solution time, and offer suggestions for future work.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this research was to develop a high-fidelity dynamic model of a parafoilpayload system with respect to its application for the Ship Launched Aerial Delivery System (SLADS). SLADS is a concept in which cargo can be transfered from ship to shore using a parafoil-payload system. It is accomplished in two phases: An initial towing phase when the glider follows the towing vessel in a passive lift mode and an autonomous gliding phase when the system is guided to the desired point. While many previous researchers have analyzed the parafoil-payload system when it is released from another airborne vehicle, limited work has been done in the area of towing up the system from ground or sea. One of the main contributions of this research was the development of a nonlinear dynamic model of a towed parafoil-payload system. After performing an extensive literature review of the existing methods of modeling a parafoil-payload system, a five degree-of-freedom model was developed. The inertial and geometric properties of the system were investigated to predict accurate results in the simulation environment. Since extensive research has been done in determining the aerodynamic characteristics of a paraglider, an existing aerodynamic model was chosen to incorporate the effects of air flow around the flexible paraglider wing. During the towing phase, it is essential that the parafoil-payload system follow the line of the towing vessel path to prevent an unstable flight condition called ‘lockout’. A detailed study of the causes of lockout, its mathematical representation and the flight conditions and the parameters related to lockout, constitute another contribution of this work. A linearized model of the parafoil-payload system was developed and used to analyze the stability of the system about equilibrium conditions. The relationship between the control surface inputs and the stability was investigated. In addition to stability of flight, one more important objective of SLADS is to tow up the parafoil-payload system as fast as possible. The tension in the tow cable is directly proportional to the rate of ascent of the parafoil-payload system. Lockout instability is more favorable when tow tensions are large. Thus there is a tradeoff between susceptibility to lockout and rapid deployment. Control strategies were also developed for optimal tow up and to maintain stability in the event of disturbances.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An optimizing compiler internal representation fundamentally affects the clarity, efficiency and feasibility of optimization algorithms employed by the compiler. Static Single Assignment (SSA) as a state-of-the-art program representation has great advantages though still can be improved. This dissertation explores the domain of single assignment beyond SSA, and presents two novel program representations: Future Gated Single Assignment (FGSA) and Recursive Future Predicated Form (RFPF). Both FGSA and RFPF embed control flow and data flow information, enabling efficient traversal program information and thus leading to better and simpler optimizations. We introduce future value concept, the designing base of both FGSA and RFPF, which permits a consumer instruction to be encountered before the producer of its source operand(s) in a control flow setting. We show that FGSA is efficiently computable by using a series T1/T2/TR transformation, yielding an expected linear time algorithm for combining together the construction of the pruned single assignment form and live analysis for both reducible and irreducible graphs. As a result, the approach results in an average reduction of 7.7%, with a maximum of 67% in the number of gating functions compared to the pruned SSA form on the SPEC2000 benchmark suite. We present a solid and near optimal framework to perform inverse transformation from single assignment programs. We demonstrate the importance of unrestricted code motion and present RFPF. We develop algorithms which enable instruction movement in acyclic, as well as cyclic regions, and show the ease to perform optimizations such as Partial Redundancy Elimination on RFPF.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis studies the minimization of the fuel consumption for a Hybrid Electric Vehicle (HEV) using Model Predictive Control (MPC). The presented MPC – based controller calculates an optimal sequence of control inputs to a hybrid vehicle using the measured plant outputs, the current dynamic states, a system model, system constraints, and an optimization cost function. The MPC controller is developed using Matlab MPC control toolbox. To evaluate the performance of the presented controller, a power-split hybrid vehicle, 2004 Toyota Prius, is selected. The vehicle uses a planetary gear set to combine three power components, an engine, a motor, and a generator, and transfer energy from these components to the vehicle wheels. The planetary gear model is developed based on the Willis’s formula. The dynamic models of the engine, the motor, and the generator, are derived based on their dynamics at the planetary gear. The MPC controller for HEV energy management is validated in the MATLAB/Simulink environment. Both the step response performance (a 0 – 60 mph step input) and the driving cycle tracking performance are evaluated. Two standard driving cycles, Urban Dynamometer Driving Schedule (UDDS) and Highway Fuel Economy Driving Schedule (HWFET), are used in the evaluation tests. For the UDDS and HWFET driving cycles, the simulation results, the fuel consumption and the battery state of charge, using the MPC controller are compared with the simulation results using the original vehicle model in Autonomie. The MPC approach shows the feasibility to improve vehicle performance and minimize fuel consumption.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Polycarbonate (PC) is an important engineering thermoplastic that is currently produced in large industrial scale using bisphenol A and monomers such as phosgene. Since phosgene is highly toxic, a non-phosgene approach using diphenyl carbonate (DPC) as an alternative monomer, as developed by Asahi Corporation of Japan, is a significantly more environmentally friendly alternative. Other advantages include the use of CO2 instead of CO as raw material and the elimination of major waste water production. However, for the production of DPC to be economically viable, reactive-distillation units are needed to obtain the necessary yields by shifting the reaction-equilibrium to the desired products and separating the products at the point where the equilibrium reaction occurs. In the field of chemical reaction engineering, there are many reactions that are suffering from the low equilibrium constant. The main goal of this research is to determine the optimal process needed to shift the reactions by using appropriate control strategies of the reactive distillation system. An extensive dynamic mathematical model has been developed to help us investigate different control and processing strategies of the reactive distillation units to increase the production of DPC. The high-fidelity dynamic models include extensive thermodynamic and reaction-kinetics models while incorporating the necessary mass and energy balance of the various stages of the reactive distillation units. The study presented in this document shows the possibility of producing DPC via one reactive distillation instead of the conventional two-column, with a production rate of 16.75 tons/h corresponding to start reactants materials of 74.69 tons/h of Phenol and 35.75 tons/h of Dimethyl Carbonate. This represents a threefold increase over the projected production rate given in the literature based on a two-column configuration. In addition, the purity of the DPC produced could reach levels as high as 99.5% with the effective use of controls. These studies are based on simulation done using high-fidelity dynamic models.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Water resource depletion and sanitation are growing problems around the world. A solution to both of these problems is the use of composting latrines, as it requires no water and has been recommended by the World Health Organization as an improved sanitation technology. However, little analysis has been done on the decomposition process occurring inside the latrine, including what temperatures are reached and what variables most affect the composting process. Having better knowledge of how outside variables affect composting latrines can aid development workers on the choice of implementing such technology, and to better educate the users on the appropriate methods of maintenance. This report presents a full, detailed construction manual and temperature data analysis of a double vault composting latrine. During the author’s two year Peace Corps service in rural Paraguay he was involved with building twenty one composting latrines, and took detailed temperature readings and visual observations of his personal latrine for ten months. The author also took limited temperature readings of fourteen community member’s latrines over a three month period. These data points were analyzed to find correlations between compost temperatures and several variables. The two main variables found to affect the compost temperatures were the seasonal trends of the outside temperatures, and the mixing and addition of moisture to the compost. Outside seasonal temperature changes were compared to those of the compost and a linear regression was performed resulting in a R2-value of 0.89. Mixing the compost and adding water, or a water/urine mixture, resulted in temperature increases of the compost 100% of the time, with seasonal temperatures determining the rate and duration of the temperature increases. The temperature readings were also used to find events when certain temperatures were held for sufficient amounts of time to reach total pathogen destruction in the compost. Four different events were recorded when a temperature of 122°F (50°C) was held for at least 24 hours, ensuring total pathogen destruction in that area of the compost. One event of 114.8°F (46°C) held for one week was also recorded, again ensuring total pathogen destruction. Through the analysis of the temperature data, however, it was found that the compost only reached total pathogen destruction levels during ten percent of the data points. Because of this the storage time recommendation outlined by the World Health Organization should be complied with. The WHO recommends storing compost for 1.5-2 years in climates with ambient temperatures of 2-20°C (35-68°F), and for at least 1 year with ambient temperatures of 20-35°C (68-95°F). If these storage durations are obtainable the use of the double vault composting latrine is an economical and achievable solution to sanitation while conserving water resources.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This research evaluated an Intelligent Compaction (IC) unit on the M-189 highway reconstruction project at Iron River, Michigan. The results from the IC unit were compared to several traditional compaction measurement devices including Nuclear Density Gauge (NDG), Geogauge, Light Weight Deflectometer (LWD), Dynamic Cone Penetrometer (DCP), and Modified Clegg Hammer (MCH). The research collected point measurements data on a test section in which 30 test locations on the final Class II sand base layer and the 22A gravel layer. These point measurements were compared with the IC measurements (ICMVs) on a point-to-point basis through a linear regression analysis. Poor correlations were obtained among different measurements points using simple regression analysis. When comparing the ICMV to the compaction measurements points. Factors attributing to the weak correlation include soil heterogeneity, variation in IC roller operation parameters, in-place moisture content, the narrow range of the compaction devices measurement ranges and support conditions of the support layers. After incorporating some of the affecting factors into a multiple regression analysis, the strength of correlation significantly improved, especially on the stiffer gravel layer. Measurements were also studied from an overall distribution perspective in terms of average, measurement range, standard deviation, and coefficient of variance. Based on data analysis, on-site project observation and literature review, conclusions were made on how IC performed in regards to compaction control on the M-189 reconstruction project.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In power electronic basedmicrogrids, the computational requirements needed to implement an optimized online control strategy can be prohibitive. The work presented in this dissertation proposes a generalized method of derivation of geometric manifolds in a dc microgrid that is based on the a-priori computation of the optimal reactions and trajectories for classes of events in a dc microgrid. The proposed states are the stored energies in all the energy storage elements of the dc microgrid and power flowing into them. It is anticipated that calculating a large enough set of dissimilar transient scenarios will also span many scenarios not specifically used to develop the surface. These geometric manifolds will then be used as reference surfaces in any type of controller, such as a sliding mode hysteretic controller. The presence of switched power converters in microgrids involve different control actions for different system events. The control of the switch states of the converters is essential for steady state and transient operations. A digital memory look-up based controller that uses a hysteretic sliding mode control strategy is an effective technique to generate the proper switch states for the converters. An example dcmicrogrid with three dc-dc boost converters and resistive loads is considered for this work. The geometric manifolds are successfully generated for transient events, such as step changes in the loads and the sources. The surfaces corresponding to a specific case of step change in the loads are then used as reference surfaces in an EEPROM for experimentally validating the control strategy. The required switch states corresponding to this specific transient scenario are programmed in the EEPROM as a memory table. This controls the switching of the dc-dc boost converters and drives the system states to the reference manifold. In this work, it is shown that this strategy effectively controls the system for a transient condition such as step changes in the loads for the example case.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

While revenue management (RM) is traditionally considered a tool of service operations, RM shows considerable potential for application in manufacturing operations. The typical challenges in make-to-order manufacturing are fixed manufacturing capacities and a great variety in offered products, going along with pronounced fluctuations in demand and profitability. Since Harris and Pinder in the mid-90s, numerous papers have furthered the understanding of RM theory in this environment. Nevertheless, results to be expected from applying the developed methods to a practical industry setting have yet to be reported. To this end, this paper investigates a possible application of RM at ThyssenKrupp VDM, leading to considerable improvements in several areas.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Master production schedule (MPS) plays an important role in an integrated production planning system. It converts the strategic planning defined in a production plan into the tactical operation execution. The MPS is also known as a tool for top management to control over manufacture resources and becomes input of the downstream planning levels such as material requirement planning (MRP) and capacity requirement planning (CRP). Hence, inappropriate decision on the MPS development may lead to infeasible execution, which ultimately causes poor delivery performance. One must ensure that the proposed MPS is valid and realistic for implementation before it is released to real manufacturing system. In practice, where production environment is stochastic in nature, the development of MPS is no longer simple task. The varying processing time, random event such as machine failure is just some of the underlying causes of uncertainty that may be hardly addressed at planning stage so that in the end the valid and realistic MPS is tough to be realized. The MPS creation problem becomes even more sophisticated as decision makers try to consider multi-objectives; minimizing inventory, maximizing customer satisfaction, and maximizing resource utilization. This study attempts to propose a methodology for MPS creation which is able to deal with those obstacles. This approach takes into account uncertainty and makes trade off among conflicting multi-objectives at the same time. It incorporates fuzzy multi-objective linear programming (FMOLP) and discrete event simulation (DES) for MPS development.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The execution of a project requires resources that are generally scarce. Classical approaches to resource allocation assume that the usage of these resources by an individual project activity is constant during the execution of that activity; in practice, however, the project manager may vary resource usage over time within prescribed bounds. This variation gives rise to the project scheduling problem which consists in allocating the scarce resources to the project activities over time such that the project duration is minimized, the total number of resource units allocated equals the prescribed work content of each activity, and various work-content-related constraints are met. We formulate this problem for the first time as a mixed-integer linear program. Our computational results for a standard test set from the literature indicate that this model outperforms the state-of-the-art solution methods for this problem.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND Muscle strength greatly influences gait kinematics. The question was whether this association is similar in different diseases. METHODS Data from instrumented gait analysis of 716 patients were retrospectively assessed. The effect of muscle strength on gait deviations, namely the gait profile score (GPS) was evaluated by means of generalised least square models. This was executed for seven different patient groups. The groups were formed according to the type of disease: orthopaedic/neurologic, uni-/bilateral affection, and flaccid/spastic muscles. RESULTS Muscle strength had a negative effect on GPS values, which did not significantly differ amongst the different patient groups. However, an offset of the GPS regression line was found, which was mostly dependent on the basic disease. Surprisingly, spastic patients, who have reduced strength and additionally spasticity in clinical examination, and flaccid neurologic patients showed the same offset. Patients with additional lack of trunk control (Tetraplegia) showed the largest offset. CONCLUSION Gait kinematics grossly depend on muscle strength. This was seen in patients with very different pathologies. Nevertheless, optimal correction of biomechanics and muscle strength may still not lead to a normal gait, especially in that of neurologic patients. The basic disease itself has an additional effect on gait deviations expressed as a GPS-offset of the linear regression line.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the setting of high-dimensional linear models with Gaussian noise, we investigate the possibility of confidence statements connected to model selection. Although there exist numerous procedures for adaptive (point) estimation, the construction of adaptive confidence regions is severely limited (cf. Li in Ann Stat 17:1001–1008, 1989). The present paper sheds new light on this gap. We develop exact and adaptive confidence regions for the best approximating model in terms of risk. One of our constructions is based on a multiscale procedure and a particular coupling argument. Utilizing exponential inequalities for noncentral χ2-distributions, we show that the risk and quadratic loss of all models within our confidence region are uniformly bounded by the minimal risk times a factor close to one.