905 resultados para Dual-Systems Theory
Resumo:
In this thesis we propose a new approach to deduction methods for temporal logic. Our proposal is based on an inductive definition of eventualities that is different from the usual one. On the basis of this non-customary inductive definition for eventualities, we first provide dual systems of tableaux and sequents for Propositional Linear-time Temporal Logic (PLTL). Then, we adapt the deductive approach introduced by means of these dual tableau and sequent systems to the resolution framework and we present a clausal temporal resolution method for PLTL. Finally, we make use of this new clausal temporal resolution method for establishing logical foundations for declarative temporal logic programming languages. The key element in the deduction systems for temporal logic is to deal with eventualities and hidden invariants that may prevent the fulfillment of eventualities. Different ways of addressing this issue can be found in the works on deduction systems for temporal logic. Traditional tableau systems for temporal logic generate an auxiliary graph in a first pass.Then, in a second pass, unsatisfiable nodes are pruned. In particular, the second pass must check whether the eventualities are fulfilled. The one-pass tableau calculus introduced by S. Schwendimann requires an additional handling of information in order to detect cyclic branches that contain unfulfilled eventualities. Regarding traditional sequent calculi for temporal logic, the issue of eventualities and hidden invariants is tackled by making use of a kind of inference rules (mainly, invariant-based rules or infinitary rules) that complicates their automation. A remarkable consequence of using either a two-pass approach based on auxiliary graphs or aone-pass approach that requires an additional handling of information in the tableau framework, and either invariant-based rules or infinitary rules in the sequent framework, is that temporal logic fails to carry out the classical correspondence between tableaux and sequents. In this thesis, we first provide a one-pass tableau method TTM that instead of a graph obtains a cyclic tree to decide whether a set of PLTL-formulas is satisfiable. In TTM tableaux are classical-like. For unsatisfiable sets of formulas, TTM produces tableaux whose leaves contain a formula and its negation. In the case of satisfiable sets of formulas, TTM builds tableaux where each fully expanded open branch characterizes a collection of models for the set of formulas in the root. The tableau method TTM is complete and yields a decision procedure for PLTL. This tableau method is directly associated to a one-sided sequent calculus called TTC. Since TTM is free from all the structural rules that hinder the mechanization of deduction, e.g. weakening and contraction, then the resulting sequent calculus TTC is also free from this kind of structural rules. In particular, TTC is free of any kind of cut, including invariant-based cut. From the deduction system TTC, we obtain a two-sided sequent calculus GTC that preserves all these good freeness properties and is finitary, sound and complete for PLTL. Therefore, we show that the classical correspondence between tableaux and sequent calculi can be extended to temporal logic. The most fruitful approach in the literature on resolution methods for temporal logic, which was started with the seminal paper of M. Fisher, deals with PLTL and requires to generate invariants for performing resolution on eventualities. In this thesis, we present a new approach to resolution for PLTL. The main novelty of our approach is that we do not generate invariants for performing resolution on eventualities. Our method is based on the dual methods of tableaux and sequents for PLTL mentioned above. Our resolution method involves translation into a clausal normal form that is a direct extension of classical CNF. We first show that any PLTL-formula can be transformed into this clausal normal form. Then, we present our temporal resolution method, called TRS-resolution, that extends classical propositional resolution. Finally, we prove that TRS-resolution is sound and complete. In fact, it finishes for any input formula deciding its satisfiability, hence it gives rise to a new decision procedure for PLTL. In the field of temporal logic programming, the declarative proposals that provide a completeness result do not allow eventualities, whereas the proposals that follow the imperative future approach either restrict the use of eventualities or deal with them by calculating an upper bound based on the small model property for PLTL. In the latter, when the length of a derivation reaches the upper bound, the derivation is given up and backtracking is used to try another possible derivation. In this thesis we present a declarative propositional temporal logic programming language, called TeDiLog, that is a combination of the temporal and disjunctive paradigms in Logic Programming. We establish the logical foundations of our proposal by formally defining operational and logical semantics for TeDiLog and by proving their equivalence. Since TeDiLog is, syntactically, a sublanguage of PLTL, the logical semantics of TeDiLog is supported by PLTL logical consequence. The operational semantics of TeDiLog is based on TRS-resolution. TeDiLog allows both eventualities and always-formulas to occur in clause heads and also in clause bodies. To the best of our knowledge, TeDiLog is the first declarative temporal logic programming language that achieves this high degree of expressiveness. Since the tableau method presented in this thesis is able to detect that the fulfillment of an eventuality is prevented by a hidden invariant without checking for it by means of an extra process, since our finitary sequent calculi do not include invariant-based rules and since our resolution method dispenses with invariant generation, we say that our deduction methods are invariant-free.
Resumo:
While some of the deepest results in nature are those that give explicit bounds between important physical quantities, some of the most intriguing and celebrated of such bounds come from fields where there is still a great deal of disagreement and confusion regarding even the most fundamental aspects of the theories. For example, in quantum mechanics, there is still no complete consensus as to whether the limitations associated with Heisenberg's Uncertainty Principle derive from an inherent randomness in physics, or rather from limitations in the measurement process itself, resulting from phenomena like back action. Likewise, the second law of thermodynamics makes a statement regarding the increase in entropy of closed systems, yet the theory itself has neither a universally-accepted definition of equilibrium, nor an adequate explanation of how a system with underlying microscopically Hamiltonian dynamics (reversible) settles into a fixed distribution.
Motivated by these physical theories, and perhaps their inconsistencies, in this thesis we use dynamical systems theory to investigate how the very simplest of systems, even with no physical constraints, are characterized by bounds that give limits to the ability to make measurements on them. Using an existing interpretation, we start by examining how dissipative systems can be viewed as high-dimensional lossless systems, and how taking this view necessarily implies the existence of a noise process that results from the uncertainty in the initial system state. This fluctuation-dissipation result plays a central role in a measurement model that we examine, in particular describing how noise is inevitably injected into a system during a measurement, noise that can be viewed as originating either from the randomness of the many degrees of freedom of the measurement device, or of the environment. This noise constitutes one component of measurement back action, and ultimately imposes limits on measurement uncertainty. Depending on the assumptions we make about active devices, and their limitations, this back action can be offset to varying degrees via control. It turns out that using active devices to reduce measurement back action leads to estimation problems that have non-zero uncertainty lower bounds, the most interesting of which arise when the observed system is lossless. One such lower bound, a main contribution of this work, can be viewed as a classical version of a Heisenberg uncertainty relation between the system's position and momentum. We finally also revisit the murky question of how macroscopic dissipation appears from lossless dynamics, and propose alternative approaches for framing the question using existing systematic methods of model reduction.
Resumo:
Dissipativity is an essential concept of systems theory. The paper provides an extension of dissipativity, named differential dissipativity, by lifting storage functions and supply rates to the tangent bundle. Differential dissipativity is connected to incremental stability in the same way as dissipativity is connected to stability. It leads to a natural formulation of differential passivity when restricting to quadratic supply rates. The paper also shows that the interconnection of differentially passive systems is differentially passive, and provides preliminary examples of differentially passive electrical systems. © IFAC.
Resumo:
The Bifurcation Interpreter is a computer program that autonomously explores the steady-state orbits of one-parameter families of periodically- driven oscillators. To report its findings, the Interpreter generates schematic diagrams and English text descriptions similar to those appearing in the science and engineering research literature. Given a system of equations as input, the Interpreter uses symbolic algebra to automatically generate numerical procedures that simulate the system. The Interpreter incorporates knowledge about dynamical systems theory, which it uses to guide the simulations, to interpret the results, and to minimize the effects of numerical error.
Resumo:
Dynamical systems theory is used as a theoretical language and tool to design a distributed control architecture for teams of mobile robots, that must transport a large object and simultaneously avoid collisions with (either static or dynamic) obstacles. Here we demonstrate in simulations and implementations in real robots that it is possible to simplify the architectures presented in previous work and to extend the approach to teams of n robots. The robots have no prior knowledge of the environment. The motion of each robot is controlled by a time series of asymptotical stable states. The attractor dynamics permits the integration of information from various sources in a graded manner. As a result, the robots show a strikingly smooth an stable team behaviour.
Resumo:
A conceptual analysis of educational leadership explored the influence of managed and living systems on 21st century leadership discourse. Drawing on a detailed understanding of managed and living systems theory compiled from the work of Capra (2002), Morgan (1997), Mitchell and Sackney (2009), and Wheatley (2007), this study draws attention to the managed systems systemic concepts of efficiency, control, and standardization, and the living systems concepts of collaboration, shared meaning, change, and interconnection as markers of systems theory that find resonance within leadership literature. Using these systemic concepts as a framework, this study provides important insights into the espousal of managed and living systems concepts within the leadership discourse.
Resumo:
The technique of linear responsibility analysis is used for a retrospective case study of a private industrial development consisting of an extension to existing buildings to provide a warehouse, services block and packing line. The organizational structure adopted on the project is analysed using concepts from systems theory which are included in Walker's theoretical model of the structure of building project organizations (Walker, 1981). This model proposes that the process of building provision can be viewed as systems and subsystems which are differentiated from each other at decision points. Further to this, the subsystems can be viewed as the interaction of managing system and operating system. Using Walker's model, a systematic analysis of the relationships between the contributors gives a quantitative assessment of the efficacy of the organizational structure used. The causes of the client's dissatisfaction with the outcome of the project were lack of integration and complexity of the managing system. However, there was a high level of satisfaction with the completed project and this is reflected by the way in which the organization structure corresponded to the model's propositions.
Resumo:
The technique of linear responsibility analysis is used for a retrospective case study of a private development consisting of an extension to an existing building to provide a wholesale butchery facility. The project used a conventionally organized management process. The organization structure adopted on the project is analysed using concepts from the systems theory, which are included in Walkers theoretical model of the structure of building project organizations. This model proposes that the process of building provision can be viewed as systems and sub-systems that are differentiated from each other at decision points. Further to this, the sub-systems can be viewed as the interaction of managing system and operating system. Using Walkers model, a systematic analysis of the relationships between the contributors gives a quantitative assessment of the efficiency of the organizational structure used. The project's organization structure diverged from the models propositions resulting in delay to the project's completion and cost overrun but the client was satisfied with the project functionally.
Resumo:
The technique of linear responsibility analysis is used for a retrospective case study of a private industrial development consisting of an engineering factory and offices. A multi-disciplinary professional practice was used to manage and design the project. The organizational structure adopted on the project is analysed using concepts from systems theory which are included in Walker's theoretical model of the structure of building project organizations (Walker, 1981). This model proposes that the process of buildings provision can be viewed as systems and sub-systems which are differentiated form each other at decision points. Further to this, the sub-systematic analysis of the relationship between the contributors gives a quantitative assessment of the efficiency of the organizational structure used. There was a high level of satisfaction with the completed project and this is reflected by the way in which the organization structure corresponded to the model's proposition. However, the project was subject to string environmental forces which the project organization was not capable of entirely overcoming.
Resumo:
The management of a public sector project is analysed using a model developed from systems theory. Linear responsibility analysis is used to identify the primary and key decision structure of the project and to generate quantitative data regarding differentiation and integration of the operating system, the managing system and the client/project team. The environmental context of the project is identified. Conclusions are drawn regarding the project organization structure's ability to cope with the prevailing environmental conditions. It is found that the complexity of the managing system imposed on the project was unable to achieve this and created serious deficiencies in the outcome of the project.
Resumo:
This article begins by identifying a close relationship between the image of children generated by several sociologists working within the new sociology of childhood perspective and the claims and ambitions of the proponents of children's autonomy rights. The image of the child as a competent, self-controlled human agent are then subjected to observation from the perspective of Niklas Luhmann's social systems theory. The new sociology of childhood's constructivist approach is compared and contrasted with Niklas Luhmann's theory of 'operational constructivism'. The article applies tenets of Luhmann's theory, to the emergence of the new childhood sociologist's image of the child as a competent, self-controlled social agent, to the epistemological status of this image and, in particular, to claims that it derives from scientific endeavour. The article proceeds to identify two theoretical developments within sociology - sociology of identity and social agency - which have brought about fundamental changes in what may be considered 'sociological' and so 'scientific' and paved the way for sociological communications about what children,really are'. In conclusion, it argues that the merging of sociology with polemics, ideology, opinion and personal beliefs and, at the level of social systems, between science and politics represents in Luhmann's terms 'dedifferentiation'- a tendency he claims may have serious adverse consequences for modern society. This warning is applied to the scientific status of sociology - its claim to be able to produce 'facts' for society, upon which social systems, such as politics and law, may rely. Like the mass media, sociology may now be capable of producing only information, and not facts, about children.
Resumo:
Studies of construction labour productivity have revealed that limited predictability and multi-agent social complexity make long-range planning of construction projects extremely inaccurate. Fire-fighting, a cultural feature of construction project management, social and structural diversity of involved permanent organizations, and structural temporality all contribute towards relational failures and frequent changes. The main purpose of this paper is therefore to demonstrate that appropriate construction planning may have a profound synergistic effect on structural integration of a project organization. Using the general systems theory perspective it is further a specific objective to investigate and evaluate organizational effects of changes in planning and potentials for achieving continuous project-organizational synergy. The newly developed methodology recognises that planning should also represent a continuous, improvement-leading driving force throughout a project. The synergistic effect of the process planning membership duality fostered project-wide integration, eliminated internal boundaries, and created a pool of constantly upgrading knowledge. It maintained a creative environment that resulted in a number of process-related improvements from all parts of the organization. As a result labour productivity has seen increases of more than 30%, profits have risen from an average of 12% to more than 18%, and project durations have been reduced by several days.
Resumo:
Control systems theory can be a discipline difficult to learn without some laboratory help. With the help of focused laboratories this discipline turns to be very interesting to the students involved. The main problem is that laboratories aren't always available to students, and sometimes, when they are available, aren't big enough to a growing student population. Thus, with computer networks growing so fast, why don't create remote control labs that can be used by a large number of students? Why don't create remote control labs using Internetⓒ Copyright ?2001 IFAC Keywords: Remote Control, Computer Networks, Database, Educational Aids, Laboratory Education, Communication Control Applications.
Resumo:
This text contains papers presented at the Institute of Mathematics and its Applications Conference on Control Theory, held at the University of Strathclyde in Glasgow. The contributions cover a wide range of topics of current interest to theoreticians and practitioners including algebraic systems theory, nonlinear control systems, adaptive control, robustness issues, infinite dimensional systems, applications studies and connections to mathematical aspects of information theory and data-fusion.
Resumo:
This article examines selected methodological insights that complexity theory might provide for planning. In particular, it focuses on the concept of fractals and, through this concept, how ways of organising policy domains across scales might have particular causal impacts. The aim of this article is therefore twofold: (a) to position complexity theory within social science through a ‘generalised discourse’, thereby orienting it to particular ontological and epistemological biases and (b) to reintroduce a comparatively new concept – fractals – from complexity theory in a way that is consistent with the ontological and epistemological biases argued for, and expand on the contribution that this might make to planning. Complexity theory is theoretically positioned as a neo-systems theory with reasons elaborated. Fractal systems from complexity theory are systems that exhibit self-similarity across scales. This concept (as previously introduced by the author in ‘Fractal spaces in planning and governance’) is further developed in this article to (a) illustrate the ontological and epistemological claims for complexity theory, and to (b) draw attention to ways of organising policy systems across scales to emphasise certain characteristics of the systems – certain distinctions. These distinctions when repeated across scales reinforce associated processes/values/end goals resulting in particular policy outcomes. Finally, empirical insights from two case studies in two different policy domains are presented and compared to illustrate the workings of fractals in planning practice.