899 resultados para Classificació AMS::93 Systems Theory
Resumo:
In the traditional way, value is created by manufacturer or producer of a product without engaging the customers. So, traditionally value creation is a monopoly in the part of a manufacturer. After gathering all the raw materials the manufacturers are inserting value to a product. And the inserted value is recognized in the time of consuming the product. In the modern time though there is traditional way of value creation but with the increase of more educated, smart, and technically sound customers the idea of value creation has changed. Now, customers are also contributing in value creation as value co-creator even before the product is consumed. This scenario has been encountered in the thesis with the main purpose of how value is cocreated in smart phone operating systems. The purpose is further divided into the following supobjectives: o What is value co-creation in smart phone operating systems? o Who participates in value co-creation in smart phone operating systems? o What are the procedures that are involved in value co-creation in smart phone operating systems? The research was conducted as a qualitative desk study by observing two of the leading smart phone operating system providers. Data has been collected from the official discussion forum of both the operating system providers. Other general concepts relating to the purpose of the study has been encountered through literature review. The research findings reveal that customers and companies both together co-create value of anticipated level when they communicate and interact with each other. However, most of the time customer to customer interactions, dialogues and discussions that come out in the core conversation help the value co-creation. The value co-creation framework sets up the customer at the main focus of value creation theory. By nullifying the inherited notion that companies only create value within its boundary and provide it to their customers in exchange of currencies. Rationally, it has been commenced that the firms are merely compromising value propositions to its customers. But the value has been co-created in a point where offerings are combined and interacted with customers’ capabilities, knowledge, resources and perceptions. This new perspective has radically altered the prospect of firms towards its customers. Typically customers are now taking part in value cocreation as a crucial member.
Resumo:
The challenge the community college faces in helping meet the needs of the living open system of society is examined in this study. It is postulated that internalization student outcomes are required by society to reduce entropy and remain self-renewing. Such behavior is characterized as having an intrinsically motivated energy source and displays the seeking and conquering of challenge, the development of reflective knowledge and skill, full use of all capabilities, internal control, growth orientation, high self-esteem, relativistic thinking and competence. The development of a conceptual systems model that suggests how transactions among students, faculty and administration might occur to best meet the needs of internalization outcomes in students, and intrinsic motivation in faculty is a major purpose of this study. It is a speculative model that is based on a synthesis of a wide variety of variables. Empirical evidence, theoretical considerations, and speculative ideas are gathered together from researchers and theoretici.ans who are working on separate answers to questions of intrinsic motivation, internal control and environments that encourage their development. The model considers the effect administrators·have on faculty anq the corresponding effect faculty may have on students. The major concentration is on the administrator--teacher interface.For administrators the model may serve as a guide in planning effective transactions, and establishing system goals. The teacher is offered a means to coordinate actions toward a specific overall objective, and the administrator, teacher and researcher are invited to use the model to experiment, innovate, verify the assumptions on which the model is based, and raise additional hypotheses. Goals and history of the community colleges in Ontario are examined against current problems, previous progress and open system thinking. The nature of the person as a five part system is explored with emphasis on intrinsic motivation. The nature, operation, conceptualization, and value of this internal energy source is reviewed in detail. The current state of society, education and management theory are considered and the value of intrinsically motivating teaching tasks together with "system four" leadership style are featured. Evidence is reviewed that suggests intrinsically motivated faculty are needed, and "system four" leadership style is the kind of interaction-influence system needed to nurture intrinsic motivation in faculty.
Resumo:
Second-rank tensor interactions, such as quadrupolar interactions between the spin- 1 deuterium nuclei and the electric field gradients created by chemical bonds, are affected by rapid random molecular motions that modulate the orientation of the molecule with respect to the external magnetic field. In biological and model membrane systems, where a distribution of dynamically averaged anisotropies (quadrupolar splittings, chemical shift anisotropies, etc.) is present and where, in addition, various parts of the sample may undergo a partial magnetic alignment, the numerical analysis of the resulting Nuclear Magnetic Resonance (NMR) spectra is a mathematically ill-posed problem. However, numerical methods (de-Pakeing, Tikhonov regularization) exist that allow for a simultaneous determination of both the anisotropy and orientational distributions. An additional complication arises when relaxation is taken into account. This work presents a method of obtaining the orientation dependence of the relaxation rates that can be used for the analysis of the molecular motions on a broad range of time scales. An arbitrary set of exponential decay rates is described by a three-term truncated Legendre polynomial expansion in the orientation dependence, as appropriate for a second-rank tensor interaction, and a linear approximation to the individual decay rates is made. Thus a severe numerical instability caused by the presence of noise in the experimental data is avoided. At the same time, enough flexibility in the inversion algorithm is retained to achieve a meaningful mapping from raw experimental data to a set of intermediate, model-free
Resumo:
In this thesis, I examined the relevance of dual-process theory to understanding forgiveness. Specifically, I argued that the internal conflict experienced by laypersons when forgiving (or finding themselves unable to forgive) and the discrepancies between existing definitions of forgiveness can currently be best understood through the lens of dual-process theory. Dual-process theory holds that individuals engage in two broad forms of mental processing corresponding to two systems, here referred to as System 1 and System 2. System 1 processing is automatic, unconscious, and operates through learned associations and heuristics. System 2 processing is effortful, conscious, and operates through rule-based and hypothetical thinking. Different definitions of forgiveness amongst both lay persons and scholars may reflect different processes within each system. Further, lay experiences with internal conflict concerning forgiveness may frequently result from processes within each system leading to different cognitive, affective, and behavioural responses. The study conducted for this thesis tested the hypotheses that processing within System 1 can directly affect one's likelihood to forgive, and that this effect is moderated by System 2 processing. I used subliminal conditioning to manipulate System 1 processing by creating positive or negative conditioned attitudes towards a hypothetical transgressor. I used working memory load (WML) to inhibit System 2 processing amongst half of the participants. The conditioning phase of the study failed and so no conclusions could be drawn regarding the roles of System 1 and System 2 in forgiveness. The implications of dual-process theory for forgiveness research and clinical practice, and directions for future research are discussed.
Resumo:
In This Paper Several Additional Gmm Specification Tests Are Studied. a First Test Is a Chow-Type Test for Structural Parameter Stability of Gmm Estimates. the Test Is Inspired by the Fact That \"Taste and Technology\" Parameters Are Uncovered. the Second Set of Specification Tests Are Var Encompassing Tests. It Is Assumed That the Dgp Has a Finite Var Representation. the Moment Restrictions Which Are Suggested by Economic Theory and Exploited in the Gmm Procedure Represent One Possible Characterization of the Dgp. the Var Is a Different But Compatible Characterization of the Same Dgp. the Idea of the Var Encompassing Tests Is to Compare Parameter Estimates of the Euler Conditions and Var Representations of the Dgp Obtained Separately with Parameter Estimates of the Euler Conditions and Var Representations Obtained Jointly. There Are Several Ways to Construct Joint Systems Which Are Discussed in the Paper. Several Applications Are Also Discussed.
Resumo:
Un objectif principal du génie logiciel est de pouvoir produire des logiciels complexes, de grande taille et fiables en un temps raisonnable. La technologie orientée objet (OO) a fourni de bons concepts et des techniques de modélisation et de programmation qui ont permis de développer des applications complexes tant dans le monde académique que dans le monde industriel. Cette expérience a cependant permis de découvrir les faiblesses du paradigme objet (par exemples, la dispersion de code et le problème de traçabilité). La programmation orientée aspect (OA) apporte une solution simple aux limitations de la programmation OO, telle que le problème des préoccupations transversales. Ces préoccupations transversales se traduisent par la dispersion du même code dans plusieurs modules du système ou l’emmêlement de plusieurs morceaux de code dans un même module. Cette nouvelle méthode de programmer permet d’implémenter chaque problématique indépendamment des autres, puis de les assembler selon des règles bien définies. La programmation OA promet donc une meilleure productivité, une meilleure réutilisation du code et une meilleure adaptation du code aux changements. Très vite, cette nouvelle façon de faire s’est vue s’étendre sur tout le processus de développement de logiciel en ayant pour but de préserver la modularité et la traçabilité, qui sont deux propriétés importantes des logiciels de bonne qualité. Cependant, la technologie OA présente de nombreux défis. Le raisonnement, la spécification, et la vérification des programmes OA présentent des difficultés d’autant plus que ces programmes évoluent dans le temps. Par conséquent, le raisonnement modulaire de ces programmes est requis sinon ils nécessiteraient d’être réexaminés au complet chaque fois qu’un composant est changé ou ajouté. Il est cependant bien connu dans la littérature que le raisonnement modulaire sur les programmes OA est difficile vu que les aspects appliqués changent souvent le comportement de leurs composantes de base [47]. Ces mêmes difficultés sont présentes au niveau des phases de spécification et de vérification du processus de développement des logiciels. Au meilleur de nos connaissances, la spécification modulaire et la vérification modulaire sont faiblement couvertes et constituent un champ de recherche très intéressant. De même, les interactions entre aspects est un sérieux problème dans la communauté des aspects. Pour faire face à ces problèmes, nous avons choisi d’utiliser la théorie des catégories et les techniques des spécifications algébriques. Pour apporter une solution aux problèmes ci-dessus cités, nous avons utilisé les travaux de Wiels [110] et d’autres contributions telles que celles décrites dans le livre [25]. Nous supposons que le système en développement est déjà décomposé en aspects et classes. La première contribution de notre thèse est l’extension des techniques des spécifications algébriques à la notion d’aspect. Deuxièmement, nous avons défini une logique, LA , qui est utilisée dans le corps des spécifications pour décrire le comportement de ces composantes. La troisième contribution consiste en la définition de l’opérateur de tissage qui correspond à la relation d’interconnexion entre les modules d’aspect et les modules de classe. La quatrième contribution concerne le développement d’un mécanisme de prévention qui permet de prévenir les interactions indésirables dans les systèmes orientés aspect.
Resumo:
Dans cette thèse l’ancienne question philosophique “tout événement a-t-il une cause ?” sera examinée à la lumière de la mécanique quantique et de la théorie des probabilités. Aussi bien en physique qu’en philosophie des sciences la position orthodoxe maintient que le monde physique est indéterministe. Au niveau fondamental de la réalité physique – au niveau quantique – les événements se passeraient sans causes, mais par chance, par hasard ‘irréductible’. Le théorème physique le plus précis qui mène à cette conclusion est le théorème de Bell. Ici les prémisses de ce théorème seront réexaminées. Il sera rappelé que d’autres solutions au théorème que l’indéterminisme sont envisageables, dont certaines sont connues mais négligées, comme le ‘superdéterminisme’. Mais il sera argué que d’autres solutions compatibles avec le déterminisme existent, notamment en étudiant des systèmes physiques modèles. Une des conclusions générales de cette thèse est que l’interprétation du théorème de Bell et de la mécanique quantique dépend crucialement des prémisses philosophiques desquelles on part. Par exemple, au sein de la vision d’un Spinoza, le monde quantique peut bien être compris comme étant déterministe. Mais il est argué qu’aussi un déterminisme nettement moins radical que celui de Spinoza n’est pas éliminé par les expériences physiques. Si cela est vrai, le débat ‘déterminisme – indéterminisme’ n’est pas décidé au laboratoire : il reste philosophique et ouvert – contrairement à ce que l’on pense souvent. Dans la deuxième partie de cette thèse un modèle pour l’interprétation de la probabilité sera proposé. Une étude conceptuelle de la notion de probabilité indique que l’hypothèse du déterminisme aide à mieux comprendre ce que c’est qu’un ‘système probabiliste’. Il semble que le déterminisme peut répondre à certaines questions pour lesquelles l’indéterminisme n’a pas de réponses. Pour cette raison nous conclurons que la conjecture de Laplace – à savoir que la théorie des probabilités présuppose une réalité déterministe sous-jacente – garde toute sa légitimité. Dans cette thèse aussi bien les méthodes de la philosophie que de la physique seront utilisées. Il apparaît que les deux domaines sont ici solidement reliés, et qu’ils offrent un vaste potentiel de fertilisation croisée – donc bidirectionnelle.
Resumo:
Cette thèse présente une revue des réflexions récentes et plus traditionnelles provenant de la théorie des systèmes, de la créativité en emploi, des théories d’organisation du travail et de la motivation afin de proposer une perspective psychologique de la régulation des actions des individus au sein d’environnements de travail complexes et incertains. Des composantes de la Théorie de la Régulation de l’Action (Frese & Zapf, 1994) ainsi que de la Théorie de l’Auto-Détermination (Deci & Ryan, 2000) sont mises en relation afin d’évaluer un modèle définissant certains schémas cognitifs clés associés aux tâches individuelles et collectives en emploi. Nous proposons que ces schémas cognitifs, organisés de manière hiérarchique, jouent un rôle central dans la régulation d’une action efficace au sein d’un système social adaptatif. Nos mesures de ces schémas cognitifs sont basées sur des échelles de mesure proposées dans le cadre des recherches sur l’ambiguïté de rôle (eg. Sawyer, 1992; Breaugh & Colihan, 1994) et sont mis en relation avec des mesures de satisfaction des besoins psychologiques (Van den Broeck, Vansteenkiste, De Witte, Soenens & Lens, 2009) et du bien-être psychologique (Goldberg, 1972). Des données provenant de 153 employés à temps plein d’une compagnie de jeu vidéo ont été récoltées à travers deux temps de mesure. Les résultats révèlent que différents types de schémas cognitifs associés aux tâches individuelles et collectives sont liés à la satisfaction de différents types de besoin psychologiques et que ces derniers sont eux-mêmes liés au bien-être psychologique. Les résultats supportent également l’hypothèse d’une organisation hiérarchique des schémas cognitifs sur la base de leur niveau d’abstraction et de leur proximité avec l’exécution concrète de l’action. Ces résultats permettent de fournir une explication initiale au processus par lequel les différents types de schémas cognitifs développés en emplois et influencé par l’environnement de travail sont associés à l’attitude des employés et à leur bien-être psychologique. Les implications pratiques et théoriques pour la motivation, l’apprentissage, l’habilitation, le bien-être psychologique et l’organisation du travail dans les environnements de travail complexes et incertains sont discutés.
Resumo:
Dans ce travail, nous étendons le nombre de conditions physiques actuellement con- nues du trou d’échange exact avec la dérivation de l’expansion de quatrième ordre du trou d’échange sphérique moyenne exacte. Nous comparons les expansions de deux- ième et de quatrième ordre avec le trou d’échange exact pour des systèmes atomiques et moléculaires. Nous avons constaté que, en général, l’expansion du quatrième ordre reproduit plus fidèlement le trou d’échange exact pour les petites valeurs de la distance interélectronique. Nous démontrons que les ensembles de base de type gaussiennes ont une influence significative sur les termes de cette nouvelle condition, en étudiant com- ment les oscillations causées par ces ensembles de bases affectent son premier terme. Aussi, nous proposons quatre modèles de trous d’échange analytiques auxquels nous imposons toutes les conditions actuellement connues du trou d’échange exact et la nou- velle présentée dans ce travail. Nous évaluons la performance des modèles en calculant des énergies d’échange et ses contributions à des énergies d’atomisation. On constate que les oscillations causeés par les bases de type gaussiennes peuvent compromettre la précision et la solution des modèles.
Resumo:
Nonlinear dynamics has emerged into a prominent area of research in the past few Decades.Turbulence, Pattern formation,Multistability etc are some of the important areas of research in nonlinear dynamics apart from the study of chaos.Chaos refers to the complex evolution of a deterministic system, which is highly sensitive to initial conditions. The study of chaos theory started in the modern sense with the investigations of Edward Lorentz in mid 60's. Later developments in this subject provided systematic development of chaos theory as a science of deterministic but complex and unpredictable dynamical systems. This thesis deals with the effect of random fluctuations with its associated characteristic timescales on chaos and synchronization. Here we introduce the concept of noise, and two familiar types of noise are discussed. The classifications and representation of white and colored noise are introduced. Based on this we introduce the concept of randomness that we deal with as a variant of the familiar concept of noise. The dynamical systems introduced are the Rossler system, directly modulated semiconductor lasers and the Harmonic oscillator. The directly modulated semiconductor laser being not a much familiar dynamical system, we have included a detailed introduction to its relevance in Chaotic encryption based cryptography in communication. We show that the effect of a fluctuating parameter mismatch on synchronization is to destroy the synchronization. Further we show that the relation between synchronization error and timescales can be found empirically but there are also cases where this is not possible. Studies show that under the variation of the parameters, the system becomes chaotic, which appears to be the period doubling route to chaos.
Resumo:
FPS is a more general form of synchronization. Hyperchaotic systems possessing more than one positive Lypaunov exponent exhibit highly complex behaviour and are more suitable for some applications like secure communications. In this thesis we report studies of FPS and MFPS of a few chaotic and hyperchaotic systems. When all the parameters of the system are known we show that active nonlinear control method can be efectively used to obtain FPS. Adaptive nonlinear control and OPCL control method are employed for obtaining FPS and MFPS when some or all parameters of the system are uncertain. A secure communication scheme based on MFPS is also proposed in theory. All our theoretical calculations are verified by numerical simulations.
Resumo:
Present thesis has discussed the design and synthesis of polymers suitable for nonlinear optics. Most of the molecules that were studied have shown good nonlinear optical activity. The second order nonlinear optical activity of the polymers was measured experimentally by Kurtz and Perry powder technique. The thesis comprises of eight chapters.The theory of NLO phenomenon and a review about the various nonlinear optical polymers has been discussed in chapter 1. The review has provided a survey of NLO active polymeric materials with a general introduction, which included the principles and the origin of nonlinear optics, and has given emphasis to polymeric materials for nonlinear optics, including guest-host systems, side chain polymers, main chain polymers, crosslinked polymers, chiral polymers etc.Chapter 2 has discussed the stability of the metal incorporated tetrapyrrole molecules, porphyrin, chlorin and bacteriochlorin.Chapter 3 has provided the NLO properties of certain organic molecules by computational tools. The chapter is divided into four parts. The first part has described the nonlinear optical properties of chromophore (D-n-A) and bichromophore (D-n-A-A-n-D) systems, which were separated by methylene spacer, by making use of DPT and semiempirical calculations.Chapter 4: A series of polyurethanes was prepared from cardanol, a renewable resource and a waste of the cashew industry by previously designed bifunctional and multifunctional polymers using quantum theoretical approach.Chapter 5: A series of chiral polyurethanes with main chain bis azo diol groups in the polymer backbone was designed and NLO activity was predicted by ZlNDO/ CV methods.In Chapter 7, polyurethanes were first designed by computational methods and the NLO properties were predicted by correction vector method. The designed bifunctional and multifunctional polyurethanes were synthesized by varying the chiral-achiral diol compositions
Resumo:
We show that, at high densities, fully variational solutions of solidlike types can be obtained from a density functional formalism originally designed for liquid 4He . Motivated by this finding, we propose an extension of the method that accurately describes the solid phase and the freezing transition of liquid 4He at zero temperature. The density profile of the interface between liquid and the (0001) surface of the 4He crystal is also investigated, and its surface energy evaluated. The interfacial tension is found to be in semiquantitative agreement with experiments and with other microscopic calculations. This opens the possibility to use unbiased density functional (DF) methods to study highly nonhomogeneous systems, like 4He interacting with strongly attractive impurities and/or substrates, or the nucleation of the solid phase in the metastable liquid.
Resumo:
We investigate the influence of the driving mechanism on the hysteretic response of systems with athermal dynamics. In the framework of local mean-field theory at finite temperature (but neglecting thermally activated processes), we compare the rate-independent hysteresis loops obtained in the random field Ising model when controlling either the external magnetic field H or the extensive magnetization M. Two distinct behaviors are observed, depending on disorder strength. At large disorder, the H-driven and M-driven protocols yield identical hysteresis loops in the thermodynamic limit. At low disorder, when the H-driven magnetization curve is discontinuous (due to the presence of a macroscopic avalanche), the M-driven loop is reentrant while the induced field exhibits strong intermittent fluctuations and is only weakly self-averaging. The relevance of these results to the experimental observations in ferromagnetic materials, shape memory alloys, and other disordered systems is discussed.