25 resultados para Monitoring Systems
em Aston University Research Archive
Resumo:
In this article we envision factors and trends that shape the next generation of environmental monitoring systems. One key factor in this respect is the combined effect of end-user needs and the general development of IT services and their availability. Currently, an environmental (monitoring) system is assumed to be reactive. It delivers measurement data and computational results only if the user explicitly asks for it either by query or subscription. There is a temptation to automate this by simply pushing data to end-users. This, however, leads easily to an "advertisement strategy", where data is pushed to end-users regardless of users' needs. Under this strategy, the mere amount of received data obfuscates the individual messages; any "automatic" service, regardless of its fitness, overruns a system that requires the user's initiative. The foreseeable problem is that, unless there is no overall management, each new environmental service is going to compete for end-users' attention and, thus, inadvertently hinder the use of existing services. As the main contribution we investigate the nature of proactive environmental systems, and how they should be designed to avoid the aforementioned problem. We also discuss how semantics, participatory sensing, uncertainty management, and situational awareness link to proactive environmental systems. We illustrate our proposals with some real-life examples.
Resumo:
Self-adaptive systems have the capability to autonomously modify their behavior at run-time in response to changes in their environment. Self-adaptation is particularly necessary for applications that must run continuously, even under adverse conditions and changing requirements; sample domains include automotive systems, telecommunications, and environmental monitoring systems. While a few techniques have been developed to support the monitoring and analysis of requirements for adaptive systems, limited attention has been paid to the actual creation and specification of requirements of self-adaptive systems. As a result, self-adaptivity is often constructed in an ad-hoc manner. In order to support the rigorous specification of adaptive systems requirements, this paper introduces RELAX, a new requirements language for self-adaptive systems that explicitly addresses uncertainty inherent in adaptive systems. We present the formal semantics for RELAX in terms of fuzzy logic, thus enabling a rigorous treatment of requirements that include uncertainty. RELAX enables developers to identify uncertainty in the requirements, thereby facilitating the design of systems that are, by definition, more flexible and amenable to adaptation in a systematic fashion. We illustrate the use of RELAX on smart home applications, including an adaptive assisted living system.
Resumo:
Self-adaptive systems have the capability to autonomously modify their behaviour at run-time in response to changes in their environment. Self-adaptation is particularly necessary for applications that must run continuously, even under adverse conditions and changing requirements; sample domains include automotive systems, telecommunications, and environmental monitoring systems. While a few techniques have been developed to support the monitoring and analysis of requirements for adaptive systems, limited attention has been paid to the actual creation and specification of requirements of self-adaptive systems. As a result, self-adaptivity is often constructed in an ad-hoc manner. In this paper, we argue that a more rigorous treatment of requirements explicitly relating to self-adaptivity is needed and that, in particular, requirements languages for self-adaptive systems should include explicit constructs for specifying and dealing with the uncertainty inherent in self-adaptive systems. We present RELAX, a new requirements language for selfadaptive systems and illustrate it using examples from the smart home domain. © 2009 IEEE.
Resumo:
This thesis presents a novel high-performance approach to time-division-multiplexing (TDM) fibre Bragg grating (FBG) optical sensors, known as the resonant cavity architecture. A background theory of FBG optical sensing includes several techniques for multiplexing sensors. The limitations of current wavelength-division-multiplexing (WDM) schemes are contrasted against the technological and commercial advantage of TDM. The author’s hypothesis that ‘it should be possible to achieve TDM FBG sensor interrogation using an electrically switched semiconductor optical amplifier (SOA)’ is then explained. Research and development of a commercially viable optical sensor interrogator based on the resonant cavity architecture forms the remainder of this thesis. A fully programmable SOA drive system allows interrogation of sensor arrays 10km long with a spatial resolution of 8cm and a variable gain system provides dynamic compensation for fluctuating system losses. Ratiometric filter- and diffractive-element spectrometer-based wavelength measurement systems are developed and analysed for different commercial applications. The ratiometric design provides a low-cost solution that has picometre resolution and low noise using 4% reflective sensors, but is less tolerant to variation in system loss. The spectrometer design is more expensive, but delivers exceptional performance with picometre resolution, low noise and tolerance to 13dB system loss variation. Finally, this thesis details the interrogator’s peripheral components, its compliance for operation in harsh industrial environments and several examples of commercial applications where it has been deployed. Applications include laboratory instruments, temperature monitoring systems for oil production, dynamic control for wind-energy and battery powered, self-contained sub-sea strain monitoring.
Resumo:
The promoters of the large groundwater developments implemented in the 1970's paid little attention to the effects of pumping on soil moisture. A field study, conducted in 1979 in the Tern Area of the Shropshire Groundwater Scheme, revealed that significant quantities of the available moisture could be removed from the root zone of vegetation when drawdown of shallow watertables occurred. Arguments to this effect, supported by the field study evidence, were successfully presented at the Shropshire Groundwater Scheme public inquiry. The aim of this study has been to expand the work which was undertaken in connection with the Shropshire Groundwater Scheme, and to develop a method whereby the effects of groundwater pumping on vegetation can be assessed, and hence the impacts minimised. Two concepts, the critical height and the soil sensitivity depth, formulated during the initial work are at the core of the Environmental Impact Assessment method whose development is described. A programme of laboratory experiments on soil columns is described, as is the derivation of relationships for determining critical heights and field capacity moisture profiles. These relationships are subsequently employed in evaluating the effects of groundwater drawdown. In employing the environmental assessment technique, digitised maps of relevant features of the Tern Area are combined to produce composite maps delineating the extent of the areas which are potentially sensitive to groundwater drawdown. A series of crop yield/moisture loss functions are then employed to estimate the impact of simulated pumping events on the agricultural community of the Tern Area. Finally, guidelines, based on experience gained through evaluation of the Tern Area case study, are presented for use in the design of soil moisture monitoring systems and in the siting of boreholes. In addition recommendations are made for development of the EIA technique, and further research needs are identified.
Resumo:
Self-adaptive systems have the capability to autonomously modify their behaviour at run-time in response to changes in their environment. Such systems are now commonly built in domains as diverse as enterprise computing, automotive control systems, and environmental monitoring systems. To date, however, there has been limited attention paid to how to engineer requirements for such systems. As a result, selfadaptivity is often constructed in an ad-hoc manner. In this paper, we argue that a more rigorous treatment of requirements relating to self-adaptivity is needed and that, in particular, requirements languages for self-adaptive systems should include explicit constructs for specifying and dealing with the uncertainty inherent in self-adaptive systems. We present some initial thoughts on a new requirements language for selfadaptive systems and illustrate it using examples from the services domain. © 2008 IEEE.
Resumo:
DUE TO COPYRIGHT RESTRICTIONS ONLY AVAILABLE FOR CONSULTATION AT ASTON UNIVERSITY LIBRARY AND INFORMATION SERVICES WITH PRIOR ARRANGEMENT
Resumo:
The amplification of demand variation up a supply chain widely termed ‘the Bullwhip Effect’ is disruptive, costly and something that supply chain management generally seeks to minimise. Originally attributed to poor system design; deficiencies in policies, organisation structure and delays in material and information flow all lead to sub-optimal reorder point calculation. It has since been attributed to exogenous random factors such as: uncertainties in demand, supply and distribution lead time but these causes are not exclusive as academic and operational studies since have shown that orders and/or inventories can exhibit significant variability even if customer demand and lead time are deterministic. This increase in the range of possible causes of dynamic behaviour indicates that our understanding of the phenomenon is far from complete. One possible, yet previously unexplored, factor that may influence dynamic behaviour in supply chains is the application and operation of supply chain performance measures. Organisations monitoring and responding to their adopted key performance metrics will make operational changes and this action may influence the level of dynamics within the supply chain, possibly degrading the performance of the very system they were intended to measure. In order to explore this a plausible abstraction of the operational responses to the Supply Chain Council’s SCOR® (Supply Chain Operations Reference) model was incorporated into a classic Beer Game distribution representation, using the dynamic discrete event simulation software Simul8. During the simulation the five SCOR Supply Chain Performance Attributes: Reliability, Responsiveness, Flexibility, Cost and Utilisation were continuously monitored and compared to established targets. Operational adjustments to the; reorder point, transportation modes and production capacity (where appropriate) for three independent supply chain roles were made and the degree of dynamic behaviour in the Supply Chain measured, using the ratio of the standard deviation of upstream demand relative to the standard deviation of the downstream demand. Factors employed to build the detailed model include: variable retail demand, order transmission, transportation delays, production delays, capacity constraints demand multipliers and demand averaging periods. Five dimensions of supply chain performance were monitored independently in three autonomous supply chain roles and operational settings adjusted accordingly. Uniqueness of this research stems from the application of the five SCOR performance attributes with modelled operational responses in a dynamic discrete event simulation model. This project makes its primary contribution to knowledge by measuring the impact, on supply chain dynamics, of applying a representative performance measurement system.
Resumo:
In this paper, we present experimental results for monitoring long distance WDM communication links using a line monitoring system suitable for legacy optically amplified long-haul undersea systems. This monitoring system is based on setting up a simple, passive, low cost high-loss optical loopback circuit at each repeater that provides a connection between the existing anti-directional undersea fibres, and can be used to define fault location. Fault location is achieved by transmitting a short pulse supervisory signal along with the WDM data signals where a portion of the overall signal is attenuated and returned to the transmit terminal by the loopback circuit. A special receiver is used at the terminal to extract the weakly returned supervisory signal where each supervisory signal is received at different times corresponding to different optical repeaters. Therefore, the degradation in any repeater appears on its corresponding supervisory signal level. We use a recirculating loop to simulate a 4600 km fibre link, on which a high-loss loopback supervisory system is implemented. Successful monitoring is accomplished through the production of an appropriate supervisory signal at the terminal that is detected and identified in a satisfactory time period after passing through up to 45 dB attenuation in the loopback circuit. © 2012 Elsevier B.V. All rights reserved.
Resumo:
In May 2006, the Ministers of Health of all the countries on the African continent, at a special session of the African Union, undertook to institutionalise efficiency monitoring within their respective national health information management systems. The specific objectives of this study were: (i) to assess the technical efficiency of National Health Systems (NHSs) of African countries for measuring male and female life expectancies, and (ii) to assess changes in health productivity over time with a view to analysing changes in efficiency and changes in technology. The analysis was based on a five-year panel data (1999-2003) from all the 53 countries of continental Africa. Data Envelopment Analysis (DEA) - a non-parametric linear programming approach - was employed to assess the technical efficiency. Malmquist Total Factor Productivity (MTFP) was used to analyse efficiency and productivity change over time among the 53 countries' national health systems. The data consisted of two outputs (male and female life expectancies) and two inputs (per capital total health expenditure and adult literacy). The DEA revealed that 49 (92.5%) countries' NHSs were run inefficiently in 1999 and 2000; 50 (94.3%), 48 (90.6%) and 47 (88.7%) operated inefficiently in 2001, 2002, and 2003 respectively. All the 53 countries' national health systems registered improvements in total factor productivity attributable mainly to technical progress. Fifty-two countries did not experience any change in scale efficiency, while thirty (56.6%) countries' national health systems had a Pure Efficiency Change (PEFFCH) index of less than one, signifying that those countries' NHSs pure efficiency contributed negatively to productivity change. All the 53 countries' national health systems registered improvements in total factor productivity, attributable mainly to technical progress. Over half of the countries' national health systems had a pure efficiency index of less than one, signifying that those countries' NHSs pure efficiency contributed negatively to productivity change. African countries may need to critically evaluate the utility of institutionalising Malmquist TFP type of analyses to monitor changes in health systems economic efficiency and productivity over time. African national health systems, per capita total health expenditure, technical efficiency, scale efficiency, Malmquist indices of productivity change, DEA
Resumo:
Objective: To assess and explain deviations from recommended practice in National Institute for Clinical Excellence (NICE) guidelines in relation to fetal heart monitoring. Design: Qualitative study. Setting: Large teaching hospital in the UK. Sample: Sixty-six hours of observation of 25 labours and interviews with 20 midwives of varying grades. Methods: Structured observations of labour and semistructured interviews with midwives. Interviews were undertaken using a prompt guide, audiotaped, and transcribed verbatim. Analysis was based on the constant comparative method, assisted by QSR N5 software. Main outcome measures: Deviations from recommended practice in relation to fetal monitoring and insights into why these occur. Results: All babies involved in the study were safely delivered, but 243 deviations from recommended practice in relation to NICE guidelines on fetal monitoring were identified, with the majority (80%) of these occurring in relation to documentation. Other deviations from recommended practice included indications for use of electronic fetal heart monitoring and conduct of fetal heart monitoring. There is evidence of difficulties with availability and maintenance of equipment, and some deficits in staff knowledge and skill. Differing orientations towards fetal monitoring were reported by midwives, which were likely to have impacts on practice. The initiation, management, and interpretation of fetal heart monitoring is complex and distributed across time, space, and professional boundaries, and practices in relation to fetal heart monitoring need to be understood within an organisational and social context. Conclusion: Some deviations from best practice guidelines may be rectified through straightforward interventions including improved systems for managing equipment and training. Other deviations from recommended practice need to be understood as the outcomes of complex processes that are likely to defy easy resolution. © RCOG 2006.
Resumo:
Two in-fiber Bragg grating (FBG) temperature sensor systems for medical applications are demonstrated: (1) an FBG flow-directed thermodilution catheter based on interferometric detection of wavelength shift that is used for cardiac monitoring; and (2) an FBG sensor system with a tunable Fabry-Perot filter for in vivo temperature profiling in nuclear magnetic resonance (NMR) machines. Preliminary results show that the FBG sensor is in good agreement with electrical sensors that are widely used in practice. A field test shows that the FBG sensor system is suitable for in situ temperature profiling in NMR machines for medical applications.
Resumo:
This thesis describes research into business user involvement in the information systems application building process. The main interest of this research is in establishing and testing techniques to quantify the relationships between identified success factors and the outcome effectiveness of 'business user development' (BUD). The availability of a mechanism to measure the levels of the success factors, and quantifiably relate them to outcome effectiveness, is important in that it provides an organisation with the capability to predict and monitor effects on BUD outcome effectiveness. This is particularly important in an era where BUD levels have risen dramatically, user centred information systems development benefits are recognised as significant, and awareness of the risks of uncontrolled BUD activity is becoming more widespread. This research targets the measurement and prediction of BUD success factors and implementation effectiveness for particular business users. A questionnaire instrument and analysis technique has been tested and developed which constitutes a tool for predicting and monitoring BUD outcome effectiveness, and is based on the BUDES (Business User Development Effectiveness and Scope) research model - which is introduced and described in this thesis. The questionnaire instrument is designed for completion by 'business users' - the target community being more explicitly defined as 'people who primarily have a business role within an organisation'. The instrument, named BUD ESP (Business User Development Effectiveness and Scope Predictor), can readily be used with survey participants, and has been shown to give meaningful and representative results.
Resumo:
The future broadband information network will undoubtedly integrate the mobility and flexibility of wireless access systems with the huge bandwidth capacity of photonics solutions to enable a communication system capable of handling the anticipated demand for interactive services. Towards wide coverage and low cost implementations of such broadband wireless photonics communication networks, various aspects of the enabling technologies are continuingly generating intense research interest. Among the core technologies, the optical generation and distribution of radio frequency signals over fibres, and the fibre optic signal processing of optical and radio frequency signals, have been the subjects for study in this thesis. Based on the intrinsic properties of single-mode optical fibres, and in conjunction with the concepts of optical fibre delay line filters and fibre Bragg gratings, a number of novel fibre-based devices, potentially suitable for applications in the future wireless photonics communication systems, have been realised. Special single-mode fibres, namely, the high birefringence (Hi-Bi) fibre and the Er/Yb doped fibre have been employed so as to exploit their merits to achieve practical and cost-effective all-fibre architectures. A number of fibre-based complex signal processors for optical and radio frequencies using novel Hi-Bi fibre delay line filter architectures have been illustrated. In particular, operations such as multichannel flattop bandpass filtering, simultaneous complementary outputs and bidirectional nonreciprocal wavelength interleaving, have been demonstrated. The proposed configurations featured greatly reduced environmental sensitivity typical of coherent fibre delay line filter schemes, reconfigurable transfer functions, negligible chromatic dispersions, and ease of implementation, not easily achievable based on other techniques. A number of unique fibre grating devices for signal filtering and fibre laser applications have been realised. The concept of the superimposed fibre Bragg gratings has been extended to non-uniform grating structures and into Hi-Bi fibres to achieve highly useful grating devices such as overwritten phase-shifted fibre grating structure and widely/narrowly spaced polarization-discriminating filters that are not limited by the intrinsic fibre properties. In terms of the-fibre-based optical millimetre wave transmitters, unique approaches based on fibre laser configurations have been proposed and demonstrated. The ability of the dual-mode distributed feedback (DFB) fibre lasers to generate high spectral purity, narrow linewidth heterodyne signals without complex feedback mechanisms has been illustrated. A novel co-located dual DFB fibre laser configuration, based on the proposed superimposed phase-shifted fibre grating structure, has been further realised with highly desired operation characteristics without the need for costly high frequency synthesizers and complex feedback controls. Lastly, a novel cavity mode condition monitoring and optimisation scheme for short length, linear-cavity fibre lasers has been proposed and achieved. Based on the concept and simplicity of the superimposed fibre laser cavities structure, in conjunction with feedback controls, enhanced output performances from the fibre lasers have been achieved. The importance of such cavity mode assessment and feedback control for optimised fibre laser output performance has been illustrated.
Resumo:
An intelligent agent, operating in an external world which cannot be fully described in its internal world model, must be able to monitor the success of a previously generated plan and to respond to any errors which may have occurred. The process of error analysis requires the ability to reason in an expert fashion about time and about processes occurring in the world. Reasoning about time is needed to deal with causality. Reasoning about processes is needed since the direct effects of a plan action can be completely specified when the plan is generated, but the indirect effects cannot. For example, the action `open tap' leads with certainty to `tap open', whereas whether there will be a fluid flow and how long it might last is more difficult to predict. The majority of existing planning systems cannot handle these kinds of reasoning, thus limiting their usefulness. This thesis argues that both kinds of reasoning require a complex internal representation of the world. The use of Qualitative Process Theory and an interval-based representation of time are proposed as a representation scheme for such a world model. The planning system which was constructed has been tested on a set of realistic planning scenarios. It is shown that even simple planning problems, such as making a cup of coffee, require extensive reasoning if they are to be carried out successfully. The final Chapter concludes that the planning system described does allow the correct solution of planning problems involving complex side effects, which planners up to now have been unable to solve.