923 resultados para Complex Engineering Systems
Resumo:
A product-service system (PSS) is a subtle blend of products and services that are offered as an integrated solution to customers. Much existing work on PSS has originated from Scandinavia and has been motivated by a sustainability agenda. Although valuable, this form has limited appeal to western manufacturers. However, by expanding the concepts of PSS to also embrace leading thinking on large scale complex service systems and informated products and services, it is possible to put forward the basis of a service business model that offers the means to differentiate from competitors who simply offer lower priced products. This paper aims to build this case. It reports the state-of-the-art of PSS, defines the concept, reports on its origin and features, discusses examples of applications, and finally proposes a research strategy for future work on this topic.
Resumo:
Models at runtime can be defined as abstract representations of a system, including its structure and behaviour, which exist in tandem with the given system during the actual execution time of that system. Furthermore, these models should be causally connected to the system being modelled, offering a reflective capability. Significant advances have been made in recent years in applying this concept, most notably in adaptive systems. In this paper we argue that a similar approach can also be used to support the dynamic generation of software artefacts at execution time. An important area where this is relevant is the generation of software mediators to tackle the crucial problem of interoperability in distributed systems. We refer to this approach as emergent middleware, representing a fundamentally new approach to resolving interoperability problems in the complex distributed systems of today. In this context, the runtime models are used to capture meta-information about the underlying networked systems that need to interoperate, including their interfaces and additional knowledge about their associated behaviour. This is supplemented by ontological information to enable semantic reasoning. This paper focuses on this novel use of models at runtime, examining in detail the nature of such runtime models coupled with consideration of the supportive algorithms and tools that extract this knowledge and use it to synthesise the appropriate emergent middleware.
Developing a probabilistic graphical structure from a model of mental-health clinical risk expertise
Resumo:
This paper explores the process of developing a principled approach for translating a model of mental-health risk expertise into a probabilistic graphical structure. The Galatean Risk Screening Tool [1] is a psychological model for mental health risk assessment based on fuzzy sets. This paper details how the knowledge encapsulated in the psychological model was used to develop the structure of the probability graph by exploiting the semantics of the clinical expertise. These semantics are formalised by a detailed specification for an XML structure used to represent the expertise. The component parts were then mapped to equivalent probabilistic graphical structures such as Bayesian Belief Nets and Markov Random Fields to produce a composite chain graph that provides a probabilistic classification of risk expertise to complement the expert clinical judgements. © Springer-Verlag 2010.
Resumo:
We present modulation instability analysis including azimuthal perturbations of steady-state continuous wave (CW) propagation in multicore-fiber configurations with a central core. In systems with a central core, a steady CW evolution regime requires power-controlled phase matching, which offers interesting spatial-division applications. Our results have general applicability and are relevant to a range of physical and engineering systems, including high-power fiber lasers, optical transmission in multicore fiber, and systems of coupled nonlinear waveguides. © 2013 Optical Society of America.
Resumo:
Purpose – This paper aims to provide a critical comment on complex funding systems. Design/methodology/approach – This is a critical comment written in the form of a poem. The poem is in the style of the English light opera composers Gilbert and Sullivan, and is a variation on their song “I Am the Very Model of a Modern Major General”, from The Pirates of Penzance. Findings - The poem spotlights financial failure. Originality/value - The poem spotlights the crazy names and poor transparency of special purpose vehicles.
Resumo:
Many software engineers have found that it is difficult to understand, incorporate and use different formal models consistently in the process of software developments, especially for large and complex software systems. This is mainly due to the complex mathematical nature of the formal methods and the lack of tool support. It is highly desirable to have software models and their related software artefacts systematically connected and used collaboratively, rather than in isolation. The success of the Semantic Web, as the next generation of Web technology, can have profound impact on the environment for formal software development. It allows both the software engineers and machines to understand the content of formal models and supports more effective software design in terms of understanding, sharing and reusing in a distributed manner. To realise the full potential of the Semantic Web in formal software development, effectively creating proper semantic metadata for formal software models and their related software artefacts is crucial. This paper proposed a framework that allows users to interconnect the knowledge about formal software models and other related documents using the semantic technology. We first propose a methodology with tool support is proposed to automatically derive ontological metadata from formal software models and semantically describe them. We then develop a Semantic Web environment for representing and sharing formal Z/OZ models. A method with prototype tool is presented to enhance semantic query to software models and other artefacts. © 2014.
Resumo:
The traditional use of global and centralised control methods, fails for large, complex, noisy and highly connected systems, which typify many real world industrial and commercial systems. This paper provides an efficient bottom up design of distributed control in which many simple components communicate and cooperate to achieve a joint system goal. Each component acts individually so as to maximise personal utility whilst obtaining probabilistic information on the global system merely through local message-passing. This leads to an implied scalable and collective control strategy for complex dynamical systems, without the problems of global centralised control. Robustness is addressed by employing a fully probabilistic design, which can cope with inherent uncertainties, can be implemented adaptively and opens a systematic rich way to information sharing. This paper opens the foreseen direction and inspects the proposed design on a linearised version of coupled map lattice with spatiotemporal chaos. A version close to linear quadratic design gives an initial insight into possible behaviours of such networks.
Resumo:
We investigate the theoretical and numerical computation of rare transitions in simple geophysical turbulent models. We consider the barotropic quasi-geostrophic and two-dimensional Navier–Stokes equations in regimes where bistability between two coexisting large-scale attractors exist. By means of large deviations and instanton theory with the use of an Onsager–Machlup path integral formalism for the transition probability, we show how one can directly compute the most probable transition path between two coexisting attractors analytically in an equilibrium (Langevin) framework and numerically otherWe adapt a class of numerical optimization algorithms known as minimum action methods to simple geophysical turbulent models. We show that by numerically minimizing an appropriate action functional in a large deviation limit, one can predict the most likely transition path for a rare transition between two states. By considering examples where theoretical predictions can be made, we show that the minimum action method successfully predicts the most likely transition path. Finally, we discuss the application and extension of such numerical optimization schemes to the computation of rare transitions observed in direct numerical simulations and experiments and to other, more complex, turbulent systems.
Resumo:
This research aims at a study of the hybrid flow shop problem which has parallel batch-processing machines in one stage and discrete-processing machines in other stages to process jobs of arbitrary sizes. The objective is to minimize the makespan for a set of jobs. The problem is denoted as: FF: batch1,sj:Cmax. The problem is formulated as a mixed-integer linear program. The commercial solver, AMPL/CPLEX, is used to solve problem instances to their optimality. Experimental results show that AMPL/CPLEX requires considerable time to find the optimal solution for even a small size problem, i.e., a 6-job instance requires 2 hours in average. A bottleneck-first-decomposition heuristic (BFD) is proposed in this study to overcome the computational (time) problem encountered while using the commercial solver. The proposed BFD heuristic is inspired by the shifting bottleneck heuristic. It decomposes the entire problem into three sub-problems, and schedules the sub-problems one by one. The proposed BFD heuristic consists of four major steps: formulating sub-problems, prioritizing sub-problems, solving sub-problems and re-scheduling. For solving the sub-problems, two heuristic algorithms are proposed; one for scheduling a hybrid flow shop with discrete processing machines, and the other for scheduling parallel batching machines (single stage). Both consider job arrival and delivery times. An experiment design is conducted to evaluate the effectiveness of the proposed BFD, which is further evaluated against a set of common heuristics including a randomized greedy heuristic and five dispatching rules. The results show that the proposed BFD heuristic outperforms all these algorithms. To evaluate the quality of the heuristic solution, a procedure is developed to calculate a lower bound of makespan for the problem under study. The lower bound obtained is tighter than other bounds developed for related problems in literature. A meta-search approach based on the Genetic Algorithm concept is developed to evaluate the significance of further improving the solution obtained from the proposed BFD heuristic. The experiment indicates that it reduces the makespan by 1.93 % in average within a negligible time when problem size is less than 50 jobs.
Resumo:
Background: Biologists often need to assess whether unfamiliar datasets warrant the time investment required for more detailed exploration. Basing such assessments on brief descriptions provided by data publishers is unwieldy for large datasets that contain insights dependent on specific scientific questions. Alternatively, using complex software systems for a preliminary analysis may be deemed as too time consuming in itself, especially for unfamiliar data types and formats. This may lead to wasted analysis time and discarding of potentially useful data. Results: We present an exploration of design opportunities that the Google Maps interface offers to biomedical data visualization. In particular, we focus on synergies between visualization techniques and Google Maps that facilitate the development of biological visualizations which have both low-overhead and sufficient expressivity to support the exploration of data at multiple scales. The methods we explore rely on displaying pre-rendered visualizations of biological data in browsers, with sparse yet powerful interactions, by using the Google Maps API. We structure our discussion around five visualizations: a gene co-regulation visualization, a heatmap viewer, a genome browser, a protein interaction network, and a planar visualization of white matter in the brain. Feedback from collaborative work with domain experts suggests that our Google Maps visualizations offer multiple, scale-dependent perspectives and can be particularly helpful for unfamiliar datasets due to their accessibility. We also find that users, particularly those less experienced with computer use, are attracted by the familiarity of the Google Maps API. Our five implementations introduce design elements that can benefit visualization developers. Conclusions: We describe a low-overhead approach that lets biologists access readily analyzed views of unfamiliar scientific datasets. We rely on pre-computed visualizations prepared by data experts, accompanied by sparse and intuitive interactions, and distributed via the familiar Google Maps framework. Our contributions are an evaluation demonstrating the validity and opportunities of this approach, a set of design guidelines benefiting those wanting to create such visualizations, and five concrete example visualizations.
Resumo:
The promise of Wireless Sensor Networks (WSNs) is the autonomous collaboration of a collection of sensors to accomplish some specific goals which a single sensor cannot offer. Basically, sensor networking serves a range of applications by providing the raw data as fundamentals for further analyses and actions. The imprecision of the collected data could tremendously mislead the decision-making process of sensor-based applications, resulting in an ineffectiveness or failure of the application objectives. Due to inherent WSN characteristics normally spoiling the raw sensor readings, many research efforts attempt to improve the accuracy of the corrupted or "dirty" sensor data. The dirty data need to be cleaned or corrected. However, the developed data cleaning solutions restrict themselves to the scope of static WSNs where deployed sensors would rarely move during the operation. Nowadays, many emerging applications relying on WSNs need the sensor mobility to enhance the application efficiency and usage flexibility. The location of deployed sensors needs to be dynamic. Also, each sensor would independently function and contribute its resources. Sensors equipped with vehicles for monitoring the traffic condition could be depicted as one of the prospective examples. The sensor mobility causes a transient in network topology and correlation among sensor streams. Based on static relationships among sensors, the existing methods for cleaning sensor data in static WSNs are invalid in such mobile scenarios. Therefore, a solution of data cleaning that considers the sensor movements is actively needed. This dissertation aims to improve the quality of sensor data by considering the consequences of various trajectory relationships of autonomous mobile sensors in the system. First of all, we address the dynamic network topology due to sensor mobility. The concept of virtual sensor is presented and used for spatio-temporal selection of neighboring sensors to help in cleaning sensor data streams. This method is one of the first methods to clean data in mobile sensor environments. We also study the mobility pattern of moving sensors relative to boundaries of sub-areas of interest. We developed a belief-based analysis to determine the reliable sets of neighboring sensors to improve the cleaning performance, especially when node density is relatively low. Finally, we design a novel sketch-based technique to clean data from internal sensors where spatio-temporal relationships among sensors cannot lead to the data correlations among sensor streams.
Resumo:
Rodents are often involved at several stages of trophic dynamics. Consequently they often play crucial roles in the structure and function of many complex ecological systems. This study sought to address the lack of baseline data concerning rodents in tropical areas, and south Florida in particular. Live trapping took place in the four major habitat types of the Long Pine Key area of Everglades National Park over the course of one year. I compared population structures and abundance of murid rodents in the four habitat types, and tested multiple weather variables for their effectiveness as predictors of rodent abundance. I found the Long Pine Key area to be depauperate in terms of species diversity. Each of the four species of rodent encountered favored a particular habitat type. The density of the understory vegetation and the avoidance of avian predators in particular appear to be the most important factors in the distribution and abundance of rodents in the Long Pine Key area of Everglades National Park.
Resumo:
Coral reef maps at various spatial scales and extents are needed for mapping, monitoring, modelling, and management of these environments. High spatial resolution satellite imagery, pixel <10 m, integrated with field survey data and processed with various mapping approaches, can provide these maps. These approaches have been accurately applied to single reefs (10-100 km**2), covering one high spatial resolution scene from which a single thematic layer (e.g. benthic community) is mapped. This article demonstrates how a hierarchical mapping approach can be applied to coral reefs from individual reef to reef-system scales (10-1000 km**2) using object-based image classification of high spatial resolution images guided by ecological and geomorphological principles. The approach is demonstrated for three individual reefs (10-35 km**2) in Australia, Fiji, and Palau; and for three complex reef systems (300-600 km**2) one in the Solomon Islands and two in Fiji. Archived high spatial resolution images were pre-processed and mosaics were created for the reef systems. Georeferenced benthic photo transect surveys were used to acquire cover information. Field and image data were integrated using an object-based image analysis approach that resulted in a hierarchically structured classification. Objects were assigned class labels based on the dominant benthic cover type, or location-relevant ecological and geomorphological principles, or a combination thereof. This generated a hierarchical sequence of reef maps with an increasing complexity in benthic thematic information that included: 'reef', 'reef type', 'geomorphic zone', and 'benthic community'. The overall accuracy of the 'geomorphic zone' classification for each of the six study sites was 76-82% using 6-10 mapping categories. For 'benthic community' classification, the overall accuracy was 52-75% with individual reefs having 14-17 categories and reef systems 20-30 categories. We show that an object-based classification of high spatial resolution imagery, guided by field data and ecological and geomorphological principles, can produce consistent, accurate benthic maps at four hierarchical spatial scales for coral reefs of various sizes and complexities.
Resumo:
Emergence of coherent structures and patterns at the nonlinear stage of modulation instability of a uniform state is an inherent feature of many biological, physical and engineering systems. There are several well-studied classical modulation instabilities, such as Benjamin-Feir, Turing and Faraday instability, which play a critical role in the self-organization of energy and matter in non-equilibrium physical, chemical and biological systems. Here we experimentally demonstrate the dissipative Faraday instability induced by spatially periodic zig-zag modulation of a dissipative parameter of the system - spectrally dependent losses - achieving generation of temporal patterns and high-harmonic mode-locking in a fibre laser. We demonstrate features of this instability that distinguish it from both the Benjamin-Feir and the purely dispersive Faraday instability. Our results open the possibilities for new designs of mode-locked lasers and can be extended to other fields of physics and engineering.
Resumo:
This thesis argues that complex adaptive social–ecological systems (SES) theory has important implications for the design of integrated ocean and coastal governance in the EU. Traditional systems of governance have struggled to deal with the global changes, complexity and uncertainties that challenge a transition towards sustainability in Europe’s maritime macro-regions. There is an apparent disconnect between governance strategies for sustainability in Europe’s maritime macro-regions and a sound theoretical basis for them. My premise is that the design of governance architecture for maritime regional sustainability should be informed by SES theory. Therefore, the aim of this research was to gain insight into a multilevel adaptive governance architecture that combines notions of sustainability and development in the context of the Atlantic Europe maritime macro-region. The central research question asked whether it is possible to achieve this insight by using a SES as a framework and analytical tool. This research adopted social ecology and sustainability science as a foundation for understanding society–nature relations. Concepts from complex adaptive systems, SES and resilience theories were integrated into a conceptual framework that guided the investigation and analysis. A study was conducted to conceptualise the European Atlantic social–ecological system (EASES). This was used to represent and understand the Atlantic Europe macro-region as a SES. The study examined the proposition that governance can be focused on building SES resilience to help achieve maritime regional sustainability. A workbook method was developed and used to elicit expert opinion regarding EASES. The study identified sources of resilience and resilience dynamics that require management in the context of multilevel adaptive governance. This research found that the Atlantic Europe macro-region is a key focal level for multilevel adaptive governance architecture. The majority of the findings are specific to Atlantic Europe and not generalisable to other maritime macro-regions in Europe.