923 resultados para Complex Engineering Systems
Resumo:
The primary objective is to identify the critical factors that have a natural impact on the performance measurement system. It is important to make correct decisions related to measurement systems, which are based on the complex business environment. The performance measurement system is combined with a very complex non-linear factor. The Six Sigma methodology is seen as one potential approach at every organisational level. It will be linked to the performance and financial measurement as well as to the analytical thinking on which the viewpoint of management depends. The complex systems are connected to the customer relationship study. As the primary throughput can be seen in a new well-defined performance measurement structure that will also be facilitated as will an analytical multifactor system. These critical factors should also be seen as a business innovation opportunity at the same time. This master's thesis has been divided into two different theoretical parts. The empirical part consists of both action-oriented and constructive research approaches with an empirical case study. The secondary objective is to seek a competitive advantage factor with a new analytical tool and the Six Sigma thinking. Process and product capabilities will be linked to the contribution of complex system. These critical barriers will be identified by the performance measuring system. The secondary throughput can be recognised as the product and the process cost efficiencies which throughputs are achieved with an advantage of management. The performance measurement potential is related to the different productivity analysis. Productivity can be seen as one essential part of the competitive advantage factor.
Resumo:
Systems suppliers are focal actors in mechanical engineering supply chains, in between general contractors and component suppliers. This research concentrates on the systems suppliers’ competitive flexibility, as a competitive advantage that the systems supplier gains from independence from the competitive forces of the market. The aim is to study the roles that power, dependence relations, social capital, and interorganizational learning have on the competitive flexibility. Research on this particular theme is scarce thus far. The research method applied here is the inductive multiple case study. Interviews from four case companies were used as main source of the qualitative data. The literature review presents previous literature on subcontracting, supply chain flexibility, supply chain relationships, social capital and interorganizational learning. The result of this study are seven propositions and consequently a model on the effects that the dominance of sales of few customers, power of competitors, significance of the manufactured system in the end product, professionalism in procurement and the significance of brand products in the business have on the competitive flexibility. These relationships are moderated by either social capital or interorganizational learning. The main results obtained from this study revolve around social capital and interorganizational learning, which have beneficial effects on systems suppliers’ competitive flexibility, by moderating the effects of other constructs of the model. Further research on this topic should include quantitative research to provide the extent to which the results can be reliably generalized. Also each construct of the model gives possible focus for more thorough research.
Resumo:
Complex System is any system that presents involved behavior, and is hard to be modeled by using the reductionist approach of successive subdivision, searching for ''elementary'' constituents. Nature provides us with plenty of examples of these systems, in fields as diverse as biology, chemistry, geology, physics, and fluid mechanics, and engineering. What happens, in general, is that for these systems we have a situation where a large number of both attracting and unstable chaotic sets coexist. As a result, we can have a rich and varied dynamical behavior, where many competing behaviors coexist. In this work, we present and discuss simple mechanical systems that are nice paradigms of Complex System, when they are subjected to random external noise. We argue that systems with few degrees of freedom can present the same complex behavior under quite general conditions.
Resumo:
The significance of services as business and human activities has increased dramatically throughout the world in the last three decades. Becoming a more and more competitive and efficient service provider while still being able to provide unique value opportunities for customers requires new knowledge and ideas. Part of this knowledge is created and utilized in daily activities in every service organization, but not all of it, and therefore an emerging phenomenon in the service context is information awareness. Terms like big data and Internet of things are not only modern buzz-words but they are also describing urgent requirements for a new type of competences and solutions. When the amount of information increases and the systems processing information become more efficient and intelligent, it is the human understanding and objectives that may get separated from the automated processes and technological innovations. This is an important challenge and the core driver for this dissertation: What kind of information is created, possessed and utilized in the service context, and even more importantly, what information exists but is not acknowledged or used? In this dissertation the focus is on the relationship between service design and service operations. Reframing this relationship refers to viewing the service system from the architectural perspective. The selected perspective allows analysing the relationship between design activities and operational activities as an information system while maintaining the tight connection to existing service research contributions and approaches. This type of an innovative approach is supported by research methodology that relies on design science theory. The methodological process supports the construction of a new design artifact based on existing theoretical knowledge, creation of new innovations and testing the design artifact components in real service contexts. The relationship between design and operations is analysed in the health care and social care service systems. The existing contributions in service research tend to abstract services and service systems as value creation, working or interactive systems. This dissertation adds an important information processing system perspective to the research. The main contribution focuses on the following argument: Only part of the service information system is automated and computerized, whereas a significant part of information processing is embedded in human activities, communication and ad-hoc reactions. The results indicate that the relationship between service design and service operations is more complex and dynamic than the existing scientific and managerial models tend to view it. Both activities create, utilize, mix and share information, making service information management a necessary but relatively unknown managerial task. On the architectural level, service system -specific elements seem to disappear, but access to more general information elements and processes can be found. While this dissertation focuses on conceptual-level design artifact construction, the results provide also very practical implications for service providers. Personal, visual and hidden activities of service, and more importantly all changes that take place in any service system have also an information dimension. Making this information dimension visual and prioritizing the processed information based on service dimensions is likely to provide new opportunities to increase activities and provide a new type of service potential for customers.
Resumo:
Resilience is the property of a system to remain trustworthy despite changes. Changes of a different nature, whether due to failures of system components or varying operational conditions, significantly increase the complexity of system development. Therefore, advanced development technologies are required to build robust and flexible system architectures capable of adapting to such changes. Moreover, powerful quantitative techniques are needed to assess the impact of these changes on various system characteristics. Architectural flexibility is achieved by embedding into the system design the mechanisms for identifying changes and reacting on them. Hence a resilient system should have both advanced monitoring and error detection capabilities to recognise changes as well as sophisticated reconfiguration mechanisms to adapt to them. The aim of such reconfiguration is to ensure that the system stays operational, i.e., remains capable of achieving its goals. Design, verification and assessment of the system reconfiguration mechanisms is a challenging and error prone engineering task. In this thesis, we propose and validate a formal framework for development and assessment of resilient systems. Such a framework provides us with the means to specify and verify complex component interactions, model their cooperative behaviour in achieving system goals, and analyse the chosen reconfiguration strategies. Due to the variety of properties to be analysed, such a framework should have an integrated nature. To ensure the system functional correctness, it should rely on formal modelling and verification, while, to assess the impact of changes on such properties as performance and reliability, it should be combined with quantitative analysis. To ensure scalability of the proposed framework, we choose Event-B as the basis for reasoning about functional correctness. Event-B is a statebased formal approach that promotes the correct-by-construction development paradigm and formal verification by theorem proving. Event-B has a mature industrial-strength tool support { the Rodin platform. Proof-based verification as well as the reliance on abstraction and decomposition adopted in Event-B provides the designers with a powerful support for the development of complex systems. Moreover, the top-down system development by refinement allows the developers to explicitly express and verify critical system-level properties. Besides ensuring functional correctness, to achieve resilience we also need to analyse a number of non-functional characteristics, such as reliability and performance. Therefore, in this thesis we also demonstrate how formal development in Event-B can be combined with quantitative analysis. Namely, we experiment with integration of such techniques as probabilistic model checking in PRISM and discrete-event simulation in SimPy with formal development in Event-B. Such an integration allows us to assess how changes and di erent recon guration strategies a ect the overall system resilience. The approach proposed in this thesis is validated by a number of case studies from such areas as robotics, space, healthcare and cloud domain.
Resumo:
The goal of this thesis is to define and validate a software engineering approach for the development of a distributed system for the modeling of composite materials, based on the analysis of various existing software development methods. We reviewed the main features of: (1) software engineering methodologies; (2) distributed system characteristics and their effect on software development; (3) composite materials modeling activities and the requirements for the software development. Using the design science as a research methodology, the distributed system for creating models of composite materials is created and evaluated. Empirical experiments which we conducted showed good convergence of modeled and real processes. During the study, we paid attention to the matter of complexity and importance of distributed system and a deep understanding of modern software engineering methods and tools.
Resumo:
Ohjelmiston suorituskyky on kokonaisvaltainen asia, johon kaikki ohjelmiston elinkaaren vaiheet vaikuttavat. Suorituskykyongelmat johtavat usein projektien viivästymisiin, kustannusten ylittymisiin sekä joissain tapauksissa projektin täydelliseen epäonnistumiseen. Software performance engineering (SPE) on ohjelmistolähtöinen lähestysmistapa, joka tarjoaa tekniikoita suorituskykyisen ohjelmiston kehittämiseen. Tämä diplomityö tutkii näitä tekniikoita ja valitsee niiden joukosta ne, jotka soveltuvat suorituskykyongelmien ratkaisemiseen kahden IT-laitehallintatuotteen kehityksessä. Työn lopputuloksena on päivitetty versio nykyisestä tuotekehitysprosessista, mikä huomioi sovellusten suorituskykyyn liittyvät haasteet tuotteiden elinkaaren eri vaiheissa.
Towards reverse engineering of Photosystem II: Synergistic Computational and Experimental Approaches
Resumo:
ABSTRACT Photosystem II (PSII) of oxygenic photosynthesis has the unique ability to photochemically oxidize water, extracting electrons from water to result in the evolution of oxygen gas while depositing these electrons to the rest of the photosynthetic machinery which in turn reduces CO2 to carbohydrate molecules acting as fuel for the cell. Unfortunately, native PSII is unstable and not suitable to be used in industrial applications. Consequently, there is a need to reverse-engineer the water oxidation photochemical reactions of PSII using solution-stable proteins. But what does it take to reverse-engineer PSII’s reactions? PSII has the pigment with the highest oxidation potential in nature known as P680. The high oxidation of P680 is in fact the driving force for water oxidation. P680 is made up of a chlorophyll a dimer embedded inside the relatively hydrophobic transmembrane environment of PSII. In this thesis, the electrostatic factors contributing to the high oxidation potential of P680 are described. PSII oxidizes water in a specialized metal cluster known as the Oxygen Evolving Complex (OEC). The pathways that water can take to enter the relatively hydrophobic region of PSII are described as well. A previous attempt to reverse engineer PSII’s reactions using the protein scaffold of E. coli’s Bacterioferritin (BFR) existed. The oxidation potential of the pigment used for the BFR ‘reaction centre’ was measured and the protein effects calculated in a similar fashion to how P680 potentials were calculated in PSII. The BFR-RC’s pigment oxidation potential was found to be 0.57 V, too low to oxidize water or tyrosine like PSII. We suggest that the observed tyrosine oxidation in BRF-RC could be driven by the ZnCe6 di-cation. In order to increase the efficiency of iii tyrosine oxidation, and ultimately oxidize water, the first potential of ZnCe6 would have to attain a value in excess of 0.8 V. The results were used to develop a second generation of BFR-RC using a high oxidation pigment. The hypervalent phosphorous porphyrin forms a radical pair that can be observed using Transient Electron Paramagnetic Resonance (TR-EPR). Finally, the results from this thesis are discussed in light of the development of solar fuel producing systems.
Polysaccharide-based Polyion Complex Micelles as New Delivery Systems for Hydrophilic Cationic Drugs
Resumo:
Les micelles polyioniques ont émergé comme des systèmes prometteurs de relargage de médicaments hydrophiles ioniques. Le but de cette étude était le développement des micelles polyioniques à base de dextrane pour la relargage de médicaments hydrophiles cationiques utilisant une nouvelle famille de copolymères bloc carboxymethyldextran-poly(éthylène glycol) (CMD-PEG). Quatre copolymères CMD-PEG ont été préparés dont deux copolymères identiques en termes de longueurs des blocs de CMD et de PEG mais différent en termes de densité de charges du bloc CMD; et deux autres copolymères dans lesquels les blocs chargés sont les mêmes mais dont les blocs de PEG sont différents. Les propriétés d’encapsulation des micelles CMD-PEG ont été évaluées avec différentes molécules cationiques: le diminazène (DIM), un médicament cationique modèle, le chlorhydrate de minocycline (MH), un analogue semi-synthétique de la tétracycline avec des propriétés neuro-protectives prometteuses et différents antibiotiques aminoglycosidiques. La cytotoxicité des copolymères CMD-PEG a été évaluée sur différentes lignées cellulaires en utilisant le test MTT et le test du Bleu Alamar. La formation de micelles des copolymères de CMD-PEG a été caractérisée par différentes techniques telles que la spectroscopie RMN 1H, la diffusion de la lumière dynamique (DLS) et la titration calorimétrique isotherme (ITC). Le taux de relargage des médicaments et l’activité pharmacologique des micelles contenant des médicaments ont aussi été évalués. Les copolymères CMD-PEG n'ont induit aucune cytotoxicité dans les hépatocytes humains et dans les cellules microgliales murines (N9) après 24 h incubation pour des concentrations allant jusqu’à 15 mg/mL. Les interactions électrostatiques entre les copolymères de CMD-PEG et les différentes drogues cationiques ont amorcé la formation de micelles polyioniques avec un coeur composé du complexe CMD-médicaments cationiques et une couronne composée de PEG. Les propriétés des micelles DIM/CMDPEG ont été fortement dépendantes du degré de carboxyméthylation du bloc CMD. Les micelles de CMD-PEG de degré de carboxyméthylation du bloc CMD ≥ 60 %, ont incorporé jusqu'à 64 % en poids de DIM et ont résisté à la désintégration induite par les sels et ceci jusqu'à 400 mM NaCl. Par contre, les micelles de CMD-PEG de degré de carboxyméthylation ~ 30% avaient une plus faible teneur en médicament (~ 40 % en poids de DIM) et se désagrégeaient à des concentrations en sel inférieures (∼ 100 mM NaCl). Le copolymère de CMD-PEG qui a montré les propriétés micellaires les plus satisfaisantes a été sélectionné comme système de livraison potentiel de chlorhydrate de minocycline (MH) et d’antibiotiques aminoglycosidiques. Les micelles CMD-PEG encapsulantes de MH ou d’aminoglycosides ont une petite taille (< 200 nm de diamètre), une forte capacité de chargement (≥ 50% en poids de médicaments) et une plus longue période de relargage de médicament. Ces micelles furent stables en solution aqueuse pendant un mois; après lyophilisation et en présence d'albumine sérique bovine. De plus, les micelles ont protégé MH contre sa dégradation en solutions aqueuses. Les micelles encapsulant les drogues ont maintenu les activités pharmacologiques de ces dernières. En outre, les micelles MH réduisent l’inflammation induite par les lipopolysaccharides dans les cellules microgliales murines (N9). Les micelles aminoglycosides ont été quant à elles capable de tuer une culture bactérienne test. Toutefois les micelles aminoglycosides/CMDPEG furent instables dans les conditions physiologiques. Les propriétés des micelles ont été considérablement améliorées par des modifications hydrophobiques de CMD-PEG. Ainsi, les micelles aminoglycosides/dodecyl-CMD-PEG ont montré une taille plus petite et une meilleure stabilité aux conditions physiologiques. Les résultats obtenus dans le cadre de cette étude montrent que CMD-PEG copolymères sont des systèmes prometteurs de relargage de médicaments cationiques.
Resumo:
Les logiciels sont en constante évolution, nécessitant une maintenance et un développement continus. Ils subissent des changements tout au long de leur vie, que ce soit pendant l'ajout de nouvelles fonctionnalités ou la correction de bogues. Lorsque les logiciels évoluent, leurs architectures ont tendance à se dégrader et deviennent moins adaptables aux nouvelles spécifications des utilisateurs. En effet, les architectures de ces logiciels deviennent plus complexes et plus difficiles à maintenir à cause des nombreuses dépendances entre les artefacts. Par conséquent, les développeurs doivent comprendre les dépendances entre les artefacts des logiciels pour prendre des mesures proactives qui facilitent les futurs changements et ralentissent la dégradation des architectures des logiciels. D'une part, le maintien d'un logiciel sans la compréhension des les dépendances entre ses artefacts peut conduire à l'introduction de défauts. D'autre part, lorsque les développeurs manquent de connaissances sur l'impact de leurs activités de maintenance, ils peuvent introduire des défauts de conception, qui ont un impact négatif sur l'évolution du logiciel. Ainsi, les développeurs ont besoin de mécanismes pour comprendre comment le changement d'un artefact impacte le reste du logiciel. Dans cette thèse, nous proposons trois contributions principales : La spécification de deux nouveaux patrons de changement et leurs utilisations pour fournir aux développeurs des informations utiles concernant les dépendances de co-changement. La spécification de la relation entre les patrons d'évolutions des artefacts et les fautes. La découverte de la relation entre les dépendances des anti-patrons et la prédisposition des différentes composantes d'un logiciel aux fautes.
Resumo:
Cette thèse présente une revue des réflexions récentes et plus traditionnelles provenant de la théorie des systèmes, de la créativité en emploi, des théories d’organisation du travail et de la motivation afin de proposer une perspective psychologique de la régulation des actions des individus au sein d’environnements de travail complexes et incertains. Des composantes de la Théorie de la Régulation de l’Action (Frese & Zapf, 1994) ainsi que de la Théorie de l’Auto-Détermination (Deci & Ryan, 2000) sont mises en relation afin d’évaluer un modèle définissant certains schémas cognitifs clés associés aux tâches individuelles et collectives en emploi. Nous proposons que ces schémas cognitifs, organisés de manière hiérarchique, jouent un rôle central dans la régulation d’une action efficace au sein d’un système social adaptatif. Nos mesures de ces schémas cognitifs sont basées sur des échelles de mesure proposées dans le cadre des recherches sur l’ambiguïté de rôle (eg. Sawyer, 1992; Breaugh & Colihan, 1994) et sont mis en relation avec des mesures de satisfaction des besoins psychologiques (Van den Broeck, Vansteenkiste, De Witte, Soenens & Lens, 2009) et du bien-être psychologique (Goldberg, 1972). Des données provenant de 153 employés à temps plein d’une compagnie de jeu vidéo ont été récoltées à travers deux temps de mesure. Les résultats révèlent que différents types de schémas cognitifs associés aux tâches individuelles et collectives sont liés à la satisfaction de différents types de besoin psychologiques et que ces derniers sont eux-mêmes liés au bien-être psychologique. Les résultats supportent également l’hypothèse d’une organisation hiérarchique des schémas cognitifs sur la base de leur niveau d’abstraction et de leur proximité avec l’exécution concrète de l’action. Ces résultats permettent de fournir une explication initiale au processus par lequel les différents types de schémas cognitifs développés en emplois et influencé par l’environnement de travail sont associés à l’attitude des employés et à leur bien-être psychologique. Les implications pratiques et théoriques pour la motivation, l’apprentissage, l’habilitation, le bien-être psychologique et l’organisation du travail dans les environnements de travail complexes et incertains sont discutés.
Resumo:
Timely detection of sudden change in dynamics that adversely affect the performance of systems and quality of products has great scientific relevance. This work focuses on effective detection of dynamical changes of real time signals from mechanical as well as biological systems using a fast and robust technique of permutation entropy (PE). The results are used in detecting chatter onset in machine turning and identifying vocal disorders from speech signal.Permutation Entropy is a nonlinear complexity measure which can efficiently distinguish regular and complex nature of any signal and extract information about the change in dynamics of the process by indicating sudden change in its value. Here we propose the use of permutation entropy (PE), to detect the dynamical changes in two non linear processes, turning under mechanical system and speech under biological system.Effectiveness of PE in detecting the change in dynamics in turning process from the time series generated with samples of audio and current signals is studied. Experiments are carried out on a lathe machine for sudden increase in depth of cut and continuous increase in depth of cut on mild steel work pieces keeping the speed and feed rate constant. The results are applied to detect chatter onset in machining. These results are verified using frequency spectra of the signals and the non linear measure, normalized coarse-grained information rate (NCIR).PE analysis is carried out to investigate the variation in surface texture caused by chatter on the machined work piece. Statistical parameter from the optical grey level intensity histogram of laser speckle pattern recorded using a charge coupled device (CCD) camera is used to generate the time series required for PE analysis. Standard optical roughness parameter is used to confirm the results.Application of PE in identifying the vocal disorders is studied from speech signal recorded using microphone. Here analysis is carried out using speech signals of subjects with different pathological conditions and normal subjects, and the results are used for identifying vocal disorders. Standard linear technique of FFT is used to substantiate thc results.The results of PE analysis in all three cases clearly indicate that this complexity measure is sensitive to change in regularity of a signal and hence can suitably be used for detection of dynamical changes in real world systems. This work establishes the application of the simple, inexpensive and fast algorithm of PE for the benefit of advanced manufacturing process as well as clinical diagnosis in vocal disorders.
Resumo:
Identification and Control of Non‐linear dynamical systems are challenging problems to the control engineers.The topic is equally relevant in communication,weather prediction ,bio medical systems and even in social systems,where nonlinearity is an integral part of the system behavior.Most of the real world systems are nonlinear in nature and wide applications are there for nonlinear system identification/modeling.The basic approach in analyzing the nonlinear systems is to build a model from known behavior manifest in the form of system output.The problem of modeling boils down to computing a suitably parameterized model,representing the process.The parameters of the model are adjusted to optimize a performanace function,based on error between the given process output and identified process/model output.While the linear system identification is well established with many classical approaches,most of those methods cannot be directly applied for nonlinear system identification.The problem becomes more complex if the system is completely unknown but only the output time series is available.Blind recognition problem is the direct consequence of such a situation.The thesis concentrates on such problems.Capability of Artificial Neural Networks to approximate many nonlinear input-output maps makes it predominantly suitable for building a function for the identification of nonlinear systems,where only the time series is available.The literature is rich with a variety of algorithms to train the Neural Network model.A comprehensive study of the computation of the model parameters,using the different algorithms and the comparison among them to choose the best technique is still a demanding requirement from practical system designers,which is not available in a concise form in the literature.The thesis is thus an attempt to develop and evaluate some of the well known algorithms and propose some new techniques,in the context of Blind recognition of nonlinear systems.It also attempts to establish the relative merits and demerits of the different approaches.comprehensiveness is achieved in utilizing the benefits of well known evaluation techniques from statistics. The study concludes by providing the results of implementation of the currently available and modified versions and newly introduced techniques for nonlinear blind system modeling followed by a comparison of their performance.It is expected that,such comprehensive study and the comparison process can be of great relevance in many fields including chemical,electrical,biological,financial and weather data analysis.Further the results reported would be of immense help for practical system designers and analysts in selecting the most appropriate method based on the goodness of the model for the particular context.
Resumo:
Genetic Programming can be effectively used to create emergent behavior for a group of autonomous agents. In the process we call Offline Emergence Engineering, the behavior is at first bred in a Genetic Programming environment and then deployed to the agents in the real environment. In this article we shortly describe our approach, introduce an extended behavioral rule syntax, and discuss the impact of the expressiveness of the behavioral description to the generation success, using two scenarios in comparison: the election problem and the distributed critical section problem. We evaluate the results, formulating criteria for the applicability of our approach.