142 resultados para Optimal Linear Control
Resumo:
This paper illustrates robust fixed order power oscillation damper design for mitigating power systems oscillations. From implementation and tuning point of view, such low and fixed structure is common practice for most practical applications, including power systems. However, conventional techniques of optimal and robust control theory cannot handle the constraint of fixed-order as it is, in general, impossible to ensure a target closed-loop transfer function by a controller of any given order. This paper deals with the problem of synthesizing or designing a feedback controller of dynamic order for a linear time-invariant plant for a fixed plant, as well as for an uncertain family of plants containing parameter uncertainty, so that stability, robust stability and robust performance are attained. The desired closed-loop specifications considered here are given in terms of a target performance vector representing a desired closed-loop design. The performance of the designed controller is validated through non-linear simulations for a range of contingencies.
Resumo:
In this work we present an optimized fuzzy visual servoing system for obstacle avoidance using an unmanned aerial vehicle. The cross-entropy theory is used to optimise the gains of our controllers. The optimization process was made using the ROS-Gazebo 3D simulation with purposeful extensions developed for our experiments. Visual servoing is achieved through an image processing front-end that uses the Camshift algorithm to detect and track objects in the scene. Experimental flight trials using a small quadrotor were performed to validate the parameters estimated from simulation. The integration of cross- entropy methods is a straightforward way to estimate optimal gains achieving excellent results when tested in real flights.
Resumo:
Motorcyclists are the most crash-prone road-user group in many Asian countries including Singapore; however, factors influencing motorcycle crashes are still not well understood. This study examines the effects of various roadway characteristics, traffic control measures and environmental factors on motorcycle crashes at different location types including expressways and intersections. Using techniques of categorical data analysis, this study has developed a set of log-linear models to investigate multi-vehicle motorcycle crashes in Singapore. Motorcycle crash risks in different circumstances have been calculated after controlling for the exposure estimated by the induced exposure technique. Results show that night-time influence increases crash risks of motorcycles particularly during merging and diverging manoeuvres on expressways, and turning manoeuvres at intersections. Riders appear to exercise more care while riding on wet road surfaces particularly during night. Many hazardous interactions at intersections tend to be related to the failure of drivers to notice a motorcycle as well as to judge correctly the speed/distance of an oncoming motorcycle. Road side conflicts due to stopping/waiting vehicles and interactions with opposing traffic on undivided roads have been found to be as detrimental factors on motorcycle safety along arterial, main and local roads away from intersections. Based on the findings of this study, several targeted countermeasures in the form of legislations, rider training, and safety awareness programmes have been recommended.
Resumo:
Hybrid system representations have been exploited in a number of challenging modelling situations, including situations where the original nonlinear dynamics are too complex (or too imprecisely known) to be directly filtered. Unfortunately, the question of how to best design suitable hybrid system models has not yet been fully addressed, particularly in the situations involving model uncertainty. This paper proposes a novel joint state-measurement relative entropy rate based approach for design of hybrid system filters in the presence of (parameterised) model uncertainty. We also present a design approach suitable for suboptimal hybrid system filters. The benefits of our proposed approaches are illustrated through design examples and simulation studies.
Resumo:
The design-build (DB) system is a popular and effective delivery method of construction projects worldwide. After owners decide to procure their projects through the DB system, they may wish to determine the optimal proportion of design to be provided in the DB request for proposals (RFPs), which serve as solicitations for design-builders and describe the scope of work. However, this presents difficulties to DB owners and there is little, if any, systematic research in this area. This paper reports on an empirical study in the USA entailing both an online questionnaire survey and Delphi survey to identify and evaluate the factors influencing owners’ decisions in determining the proportion of design to include in DB RFPs. Eleven factors are identified, i.e. (1) clarity of project scope; (2) applicability of performance specifications; (3) desire for design innovation; (4) site constraints; (5) availability of competent design-builders; (6) project control requirements; (7) user group involvement level; (8) third party requirements; (9) owner experience with DB; (10) project complexity; and (11) schedule constraints. A statistically significant agreement on the eleven factors was also obtained from the (mainly non-owner) Delphi experts. Although some of the experts hold different opinions on how these factors affect the proportion of design, these findings furnish various stakeholders with a better understanding of the delivery process of DB projects and the appropriate provision of project information in DB RFPs. As the result is mainly industry opinion concerning the optimal proportion of design, in addition and for completeness, future studies should be conducted to obtain a big picture of the optimal proportion of design by means of seeking owners’ inputs.
Resumo:
A physiological control system was developed for a rotary left ventricular assist device (LVAD) in which the target pump flow rate (LVADQ) was set as a function of left atrial pressure (LAP), mimicking the Frank-Starling mechanism. The control strategy was implemented using linear PID control and was evaluated in a pulsatile mock circulation loop using a prototyped centrifugal pump by varying pulmonary vascular resistance to alter venous return. The control strategy automatically varied pump speed (2460 to 1740 to 2700 RPM) in response to a decrease and subsequent increase in venous return. In contrast, a fixed-speed pump caused a simulated ventricular suction event during low venous return and higher ventricular volumes during high venous return. The preload sensitivity was increased from 0.011 L/min/mmHg in fixed speed mode to 0.47L/min/mmHg, a value similar to that of the native healthy heart. The sensitivity varied automatically to maintain the LAP and LVADQ within a predefined zone. This control strategy requires the implantation of a pressure sensor in the left atrium and a flow sensor around the outflow cannula of the LVAD. However, appropriate pressure sensor technology is not yet commercially available and so an alternative measure of preload such as pulsatility of pump signals should be investigated.
Resumo:
Background Despite evidence that up to 35% of patients with cancer experience significant distress, access to effective psychosocial care is limited by lack of systematic approaches to assessment, a paucity of psychosocial services, and patient reluctance to accept treatment either because of perceived stigma or difficulties with access to specialist psycho-oncology services due to isolation or disease burden. This paper presents an overview of a randomised study to evaluate the effectiveness of a brief tailored psychosocial Intervention delivered by health professionals in cancer care who undergo focused training and participate in clinical supervision. Methods/design Health professionals from the disciplines of nursing, occupational therapy, speech pathology, dietetics, physiotherapy or radiation therapy will participate in training to deliver the psychosocial Intervention focusing on core concepts of supportive-expressive, cognitive and dignity-conserving care. Health professional training will consist of completion of a self-directed manual and participation in a skills development session. Participating health professionals will be supported through structured clinical supervision whilst delivering the Intervention. In the stepped wedge design each of the 5 participating clinical sites will be allocated in random order from Control condition to Training then delivery of the Intervention. A total of 600 patients will be recruited across all sites. Based on level of distress or risk factors eligible patients will receive up to 4 sessions, each of up to 30 minutes in length, delivered face-to-face or by telephone. Participants will be assessed at baseline and 10-week follow-up. Patient outcome measures include anxiety and depression, quality of life, unmet psychological and supportive care needs. Health professional measures include psychological morbidity, stress and burnout. Process evaluation will be conducted to assess perceptions of participation in the study and the factors that may promote translation of learning into practice. Discussion This study will provide important information about the effectiveness of a brief tailored psychological Intervention for patients with cancer and the potential to prevent development of significant distress in patients considered at risk. It will yield data about the feasibility of this model of care in routine clinical practice and identify enablers and barriers to its systematic implementation in cancer settings.
Resumo:
A major challenge in modern photonics and nano-optics is the diffraction limit of light which does not allow field localisation into regions with dimensions smaller than half the wavelength. Localisation of light into nanoscale regions (beyond its diffraction limit) has applications ranging from the design of optical sensors and measurement techniques with resolutions as high as a few nanometres, to the effective delivery of optical energy into targeted nanoscale regions such as quantum dots, nano-electronic and nano-optical devices. This field has become a major research direction over the last decade. The use of strongly localised surface plasmons in metallic nanostructures is one of the most promising approaches to overcome this problem. Therefore, the aim of this thesis is to investigate the linear and non-linear propagation of surface plasmons in metallic nanostructures. This thesis will focus on two main areas of plasmonic research –– plasmon nanofocusing and plasmon nanoguiding. Plasmon nanofocusing – The main aim of plasmon nanofocusing research is to focus plasmon energy into nanoscale regions using metallic nanostructures and at the same time achieve strong local field enhancement. Various structures for nanofocusing purposes have been proposed and analysed such as sharp metal wedges, tapered metal films on dielectric substrates, tapered metal rods, and dielectric V-grooves in metals. However, a number of important practical issues related to nanofocusing in these structures still remain unclear. Therefore, one of the main aims of this thesis is to address two of the most important of issues which are the coupling efficiency and heating effects of surface plasmons in metallic nanostructures. The method of analysis developed throughout this thesis is a general treatment that can be applied to a diversity of nanofocusing structures, with results shown here for the specific case of sharp metal wedges. Based on the geometrical optics approximation, it is demonstrated that the coupling efficiency from plasmons generated with a metal grating into the nanofocused symmetric or quasi-symmetric modes may vary between ~50% to ~100% depending on the structural parameters. Optimal conditions for nanofocusing with the view to minimise coupling and dissipative losses are also determined and discussed. It is shown that the temperature near the tip of a metal wedge heated by nanosecond plasmonic pulses can increase by several hundred degrees Celsius. This temperature increase is expected to lead to nonlinear effects, self-influence of the focused plasmon, and ultimately self-destruction of the metal tip. This thesis also investigates a different type of nanofocusing structure which consists of a tapered high-index dielectric layer resting on a metal surface. It is shown that the nanofocusing mechanism that occurs in this structure is somewhat different from other structures that have been considered thus far. For example, the surface plasmon experiences significant backreflection and mode transformation at a cut-off thickness. In addition, the reflected plasmon shows negative refraction properties that have not been observed in other nanofocusing structures considered to date. Plasmon nanoguiding – Guiding surface plasmons using metallic nanostructures is important for the development of highly integrated optical components and circuits which are expected to have a superior performance compared to their electronicbased counterparts. A number of different plasmonic waveguides have been considered over the last decade including the recently considered gap and trench plasmon waveguides. The gap and trench plasmon waveguides have proven to be difficult to fabricate. Therefore, this thesis will propose and analyse four different modified gap and trench plasmon waveguides that are expected to be easier to fabricate, and at the same time acquire improved propagation characteristics of the guided mode. In particular, it is demonstrated that the guided modes are significantly screened by the extended metal at the bottom of the structure. This is important for the design of highly integrated optics as it provides the opportunity to place two waveguides close together without significant cross-talk. This thesis also investigates the use of plasmonic nanowires to construct a Fabry-Pérot resonator/interferometer. It is shown that the resonance effect can be achieved with the appropriate resonator length and gap width. Typical quality factors of the Fabry- Pérot cavity are determined and explained in terms of radiative and dissipative losses. The possibility of using a nanowire resonator for the design of plasmonic filters with close to ~100% transmission is also demonstrated. It is expected that the results obtained in this thesis will play a vital role in the development of high resolution near field microscopy and spectroscopy, new measurement techniques and devices for single molecule detection, highly integrated optical devices, and nanobiotechnology devices for diagnostics of living cells.
Resumo:
Quality oriented management systems and methods have become the dominant business and governance paradigm. From this perspective, satisfying customers’ expectations by supplying reliable, good quality products and services is the key factor for an organization and even government. During recent decades, Statistical Quality Control (SQC) methods have been developed as the technical core of quality management and continuous improvement philosophy and now are being applied widely to improve the quality of products and services in industrial and business sectors. Recently SQC tools, in particular quality control charts, have been used in healthcare surveillance. In some cases, these tools have been modified and developed to better suit the health sector characteristics and needs. It seems that some of the work in the healthcare area has evolved independently of the development of industrial statistical process control methods. Therefore analysing and comparing paradigms and the characteristics of quality control charts and techniques across the different sectors presents some opportunities for transferring knowledge and future development in each sectors. Meanwhile considering capabilities of Bayesian approach particularly Bayesian hierarchical models and computational techniques in which all uncertainty are expressed as a structure of probability, facilitates decision making and cost-effectiveness analyses. Therefore, this research investigates the use of quality improvement cycle in a health vii setting using clinical data from a hospital. The need of clinical data for monitoring purposes is investigated in two aspects. A framework and appropriate tools from the industrial context are proposed and applied to evaluate and improve data quality in available datasets and data flow; then a data capturing algorithm using Bayesian decision making methods is developed to determine economical sample size for statistical analyses within the quality improvement cycle. Following ensuring clinical data quality, some characteristics of control charts in the health context including the necessity of monitoring attribute data and correlated quality characteristics are considered. To this end, multivariate control charts from an industrial context are adapted to monitor radiation delivered to patients undergoing diagnostic coronary angiogram and various risk-adjusted control charts are constructed and investigated in monitoring binary outcomes of clinical interventions as well as postintervention survival time. Meanwhile, adoption of a Bayesian approach is proposed as a new framework in estimation of change point following control chart’s signal. This estimate aims to facilitate root causes efforts in quality improvement cycle since it cuts the search for the potential causes of detected changes to a tighter time-frame prior to the signal. This approach enables us to obtain highly informative estimates for change point parameters since probability distribution based results are obtained. Using Bayesian hierarchical models and Markov chain Monte Carlo computational methods, Bayesian estimators of the time and the magnitude of various change scenarios including step change, linear trend and multiple change in a Poisson process are developed and investigated. The benefits of change point investigation is revisited and promoted in monitoring hospital outcomes where the developed Bayesian estimator reports the true time of the shifts, compared to priori known causes, detected by control charts in monitoring rate of excess usage of blood products and major adverse events during and after cardiac surgery in a local hospital. The development of the Bayesian change point estimators are then followed in a healthcare surveillances for processes in which pre-intervention characteristics of patients are viii affecting the outcomes. In this setting, at first, the Bayesian estimator is extended to capture the patient mix, covariates, through risk models underlying risk-adjusted control charts. Variations of the estimator are developed to estimate the true time of step changes and linear trends in odds ratio of intensive care unit outcomes in a local hospital. Secondly, the Bayesian estimator is extended to identify the time of a shift in mean survival time after a clinical intervention which is being monitored by riskadjusted survival time control charts. In this context, the survival time after a clinical intervention is also affected by patient mix and the survival function is constructed using survival prediction model. The simulation study undertaken in each research component and obtained results highly recommend the developed Bayesian estimators as a strong alternative in change point estimation within quality improvement cycle in healthcare surveillances as well as industrial and business contexts. The superiority of the proposed Bayesian framework and estimators are enhanced when probability quantification, flexibility and generalizability of the developed model are also considered. The empirical results and simulations indicate that the Bayesian estimators are a strong alternative in change point estimation within quality improvement cycle in healthcare surveillances. The superiority of the proposed Bayesian framework and estimators are enhanced when probability quantification, flexibility and generalizability of the developed model are also considered. The advantages of the Bayesian approach seen in general context of quality control may also be extended in the industrial and business domains where quality monitoring was initially developed.
Resumo:
Mounting scientific evidence suggests newly imposed disturbance and/or alterations to existing disturbances facilitate invasion. Several empirical studies have explored the role of disturbance in invasion, but little work has been done to fit current understanding into a format useful for practical control efforts. We are working towards addressing this shortcoming by developing a metapopulation model couched in a decision theory framework. This approach has allowed us to investigate how incorporating the negative effects of disturbance on native vegetation into decision-making can change optimal control measures. In this paper, we present some preliminary results.
Resumo:
A priority when designing control strategies for autonomous underwater vehicles is to emphasize their cost of implementation on a real vehicle and at the same time to minimize a prescribed criterion such as time, energy, payload or combination of those. Indeed, the major issue is that due to the vehicles' design and the actuation modes usually under consideration for underwater platforms the number of actuator switchings must be kept to a small value to ensure feasibility and precision. This constraint is typically not verified by optimal trajectories which might not even be piecewise constants. Our goal is to provide a feasible trajectory that minimizes the number of switchings while maintaining some qualities of the desired trajectory, such as optimality with respect to a given criterion. The one-sided Lipschitz constant is used to derive theoretical estimates. The theory is illustrated on two examples, one is a fully actuated underwater vehicle capable of motion in six degrees-of-freedom and one is minimally actuated with control motions constrained to the vertical plane.
Resumo:
Linear adaptive channel equalization using the least mean square (LMS) algorithm and the recursive least-squares(RLS) algorithm for an innovative multi-user (MU) MIMOOFDM wireless broadband communications system is proposed. The proposed equalization method adaptively compensates the channel impairments caused by frequency selectivity in the propagation environment. Simulations for the proposed adaptive equalizer are conducted using a training sequence method to determine optimal performance through a comparative analysis. Results show an improvement of 0.15 in BER (at a SNR of 16 dB) when using Adaptive Equalization and RLS algorithm compared to the case in which no equalization is employed. In general, adaptive equalization using LMS and RLS algorithms showed to be significantly beneficial for MU-MIMO-OFDM systems.
Resumo:
Control of biospecimen quality that is linked to processing is one of the goals of biospecimen science. Consensus is lacking, however, regarding optimal sample quality-control (QC) tools (ie, markers and assays). The aim of this review was to identify QC tools, both for fluid and solid-tissue samples, based on a comprehensive and critical literature review. The most readily applicable tools are those with a known threshold for the preanalytical variation and a known reference range for the QC analyte. Only a few meaningful markers were identified that meet these criteria, such as CD40L for assessing serum exposure at high temperatures and VEGF for assessing serum freeze-thawing. To fully assess biospecimen quality, multiple QC markers are needed. Here we present the most promising biospecimen QC tools that were identified.
Resumo:
The R statistical environment and language has demonstrated particular strengths for interactive development of statistical algorithms, as well as data modelling and visualisation. Its current implementation has an interpreter at its core which may result in a performance penalty in comparison to directly executing user algorithms in the native machine code of the host CPU. In contrast, the C++ language has no built-in visualisation capabilities, handling of linear algebra or even basic statistical algorithms; however, user programs are converted to high-performance machine code, ahead of execution. A new method avoids possible speed penalties in R by using the Rcpp extension package in conjunction with the Armadillo C++ matrix library. In addition to the inherent performance advantages of compiled code, Armadillo provides an easy-to-use template-based meta-programming framework, allowing the automatic pooling of several linear algebra operations into one, which in turn can lead to further speedups. With the aid of Rcpp and Armadillo, conversion of linear algebra centered algorithms from R to C++ becomes straightforward. The algorithms retains the overall structure as well as readability, all while maintaining a bidirectional link with the host R environment. Empirical timing comparisons of R and C++ implementations of a Kalman filtering algorithm indicate a speedup of several orders of magnitude.
Resumo:
Ocean gliders constitute an important advance in the highly demanding ocean monitoring scenario. Their effciency, endurance and increasing robustness make these vehicles an ideal observing platform for many long term oceanographic applications. However, they have proved to be also useful in the opportunis-tic short term characterization of dynamic structures. Among these, mesoscale eddies are of particular interest due to the relevance they have in many oceano-graphic processes.