783 resultados para Fundamentals of computing theory
Resumo:
The application of Concurrency Theory to Systems Biology is in its earliest stage of progress. The metaphor of cells as computing systems by Regev and Shapiro opened the employment of concurrent languages for the modelling of biological systems. Their peculiar characteristics led to the design of many bio-inspired formalisms which achieve higher faithfulness and specificity. In this thesis we present pi@, an extremely simple and conservative extension of the pi-calculus representing a keystone in this respect, thanks to its expressiveness capabilities. The pi@ calculus is obtained by the addition of polyadic synchronisation and priority to the pi-calculus, in order to achieve compartment semantics and atomicity of complex operations respectively. In its direct application to biological modelling, the stochastic variant of the calculus, Spi@, is shown able to model consistently several phenomena such as formation of molecular complexes, hierarchical subdivision of the system into compartments, inter-compartment reactions, dynamic reorganisation of compartment structure consistent with volume variation. The pivotal role of pi@ is evidenced by its capability of encoding in a compositional way several bio-inspired formalisms, so that it represents the optimal core of a framework for the analysis and implementation of bio-inspired languages. In this respect, the encodings of BioAmbients, Brane Calculi and a variant of P Systems in pi@ are formalised. The conciseness of their translation in pi@ allows their indirect comparison by means of their encodings. Furthermore it provides a ready-to-run implementation of minimal effort whose correctness is granted by the correctness of the respective encoding functions. Further important results of general validity are stated on the expressive power of priority. Several impossibility results are described, which clearly state the superior expressiveness of prioritised languages and the problems arising in the attempt of providing their parallel implementation. To this aim, a new setting in distributed computing (the last man standing problem) is singled out and exploited to prove the impossibility of providing a purely parallel implementation of priority by means of point-to-point or broadcast communication.
Resumo:
This dissertation mimics the Turkish college admission procedure. It started with the purpose to reduce the inefficiencies in Turkish market. For this purpose, we propose a mechanism under a new market structure; as we prefer to call, semi-centralization. In chapter 1, we give a brief summary of Matching Theory. We present the first examples in Matching history with the most general papers and mechanisms. In chapter 2, we propose our mechanism. In real life application, that is in Turkish university placements, the mechanism reduces the inefficiencies of the current system. The success of the mechanism depends on the preference profile. It is easy to show that under complete information the mechanism implements the full set of stable matchings for a given profile. In chapter 3, we refine our basic mechanism. The modification on the mechanism has a crucial effect on the results. The new mechanism is, as we call, a middle mechanism. In one of the subdomain, this mechanism coincides with the original basic mechanism. But, in the other partition, it gives the same results with Gale and Shapley's algorithm. In chapter 4, we apply our basic mechanism to well known Roommate Problem. Since the roommate problem is in one-sided game patern, firstly we propose an auxiliary function to convert the game semi centralized two-sided game, because our basic mechanism is designed for this framework. We show that this process is succesful in finding a stable matching in the existence of stability. We also show that our mechanism easily and simply tells us if a profile lacks of stability by using purified orderings. Finally, we show a method to find all the stable matching in the existence of multi stability. The method is simply to run the mechanism for all of the top agents in the social preference.
Resumo:
The theory on the intensities of 4f-4f transitions introduced by B.R. Judd and G.S. Ofelt in 1962 has become a center piece in rare-earth optical spectroscopy over the past five decades. Many fundamental studies have since explored the physical origins of the Judd–Ofelt theory and have proposed numerous extensions to the original model. A great number of studies have applied the Judd–Ofelt theory to a wide range of rare-earth doped materials, many of them with important applications in solid-state lasers, optical amplifiers, phosphors for displays and solid state lighting, upconversion and quantum-cutting materials, and fluorescent markers. This paper takes the view of the experimentalist who is interested in appreciating the basic concepts, implications, assumptions, and limitations of the Judd–Ofelt theory in order to properly apply it to practical problems. We first present the formalism for calculating the wavefunctions of 4f electronic states in a concise form and then show their application to the calculation and fitting of 4f-4f transition intensities. The potential, limitations and pitfalls of the theory are discussed, and a detailed case study of LaCl3:Er3+ is presented.
Resumo:
Family change theory (Kagitcibasi, 1996, 2007) is an approach which can be used to explain how modernisation and globalisation processes affect the family. The most important assumption of the theory is that when traditional interdependent cultures modernise, they need not necessarily develop in the direction of the independent family model typical of Western individualistic societies. Instead, they may develop towards a family model of emotional interdependence that combines continuing emotional interdependencies in the family with declining material interdependencies and with rising personal autonomy. In this chapter a preliminary evaluation of the empirical status of family change theory is given based on a review of recent cross-cultural studies. It will be shown to what extent the few studies that have been systematically conducted in this respect have found results either supporting or not supporting aspects ofthe theory, and where the strengths and problems of this research lie.
Resumo:
Beginning in the early 1980s, the health care system experienced momentous realignments. Fundamental changes in structures of traditional health care organizations, shifts in authority and relationships of professionals and institutions, and the increasing influence of managed care contributed to a relatively stable industry entering into a state of turbulence. The dynamics of these changes are recurring themes in the health services literature. The purpose of this dissertation was to examine the content of this literature over a defined time period and within the perspective of a theory of organizational change. ^ Using a theoretical framework based upon the organizational theory known as Organizational Ecology, secondary data from the period between 1983 and 1994 was reviewed. Analysis of the literature identified through a defined search methodology was focused upon determining the manner in which the literature characterized changes that were described. Using a model constructed from fundamentals of Organizational Ecology with which to structure an assessment of content, literature was summarized for the manner and extent of change in specific organizational forms and for the changes in emphasis by the environmental dynamics directing changes in the population of organizations. Although it was not the intent of the analysis to substantiate causal relationships between environmental resources selected as the determinants of organizational change and the observed changes in organizational forms, the structured review of content of the literature established a strong basis for inferring such a relationship. ^ The results of the integrative review of the literature and the power of the appraisal achieved through the theoretical framework constructed for the analysis indicate that there is considerable value in such an approach. An historical perspective on changes which have transformed the health care system developed within a defined organizational theory provide a unique insight into these changes and indicate the need for further development of such an analytical model. ^
Resumo:
In recent years, the econometrics literature has shown a growing interest in the study of partially identified models, in which the object of economic and statistical interest is a set rather than a point. The characterization of this set and the development of consistent estimators and inference procedures for it with desirable properties are the main goals of partial identification analysis. This review introduces the fundamental tools of the theory of random sets, which brings together elements of topology, convex geometry, and probability theory to develop a coherent mathematical framework to analyze random elements whose realizations are sets. It then elucidates how these tools have been fruitfully applied in econometrics to reach the goals of partial identification analysis.
Resumo:
We regularize compact and non-compact Abelian Chern–Simons–Maxwell theories on a spatial lattice using the Hamiltonian formulation. We consider a doubled theory with gauge fields living on a lattice and its dual lattice. The Hilbert space of the theory is a product of local Hilbert spaces, each associated with a link and the corresponding dual link. The two electric field operators associated with the link-pair do not commute. In the non-compact case with gauge group R, each local Hilbert space is analogous to the one of a charged “particle” moving in the link-pair group space R2 in a constant “magnetic” background field. In the compact case, the link-pair group space is a torus U(1)2 threaded by k units of quantized “magnetic” flux, with k being the level of the Chern–Simons theory. The holonomies of the torus U(1)2 give rise to two self-adjoint extension parameters, which form two non-dynamical background lattice gauge fields that explicitly break the manifest gauge symmetry from U(1) to Z(k). The local Hilbert space of a link-pair then decomposes into representations of a magnetic translation group. In the pure Chern–Simons limit of a large “photon” mass, this results in a Z(k)-symmetric variant of Kitaev’s toric code, self-adjointly extended by the two non-dynamical background lattice gauge fields. Electric charges on the original lattice and on the dual lattice obey mutually anyonic statistics with the statistics angle . Non-Abelian U(k) Berry gauge fields that arise from the self-adjoint extension parameters may be interesting in the context of quantum information processing.
Resumo:
BACKGROUND Compliance with surgical checklist use remains an obstacle in the context of checklist implementation programs. The theory of planned behaviour was applied to analyse attitudes, perceived behaviour control, and norms as psychological antecedents of individuals' intentions to use the checklist. METHODS A cross-sectional survey study with staff (N = 866) of 10 Swiss hospitals was conducted in German and French. Group mean differences between individuals with and without managerial function were computed. Structural equation modelling and confirmatory factor analysis was applied to investigate the structural relation between attitudes, perceived behaviour control, norms, and intentions. RESULTS Significant mean differences in favour of individuals with managerial function emerged for norms, perceived behavioural control, and intentions, but not for attitudes. Attitudes and perceived behavioural control had a significant direct effect on intentions whereas norms had not. CONCLUSIONS Individuals with managerial function exhibit stronger perceived behavioural control, stronger norms, and stronger intentions. This could be applied in facilitating checklist implementation. The structural model of the theory of planned behaviour remains stable across groups, indicating a valid model to describe antecedents of intentions in the context of surgical checklist implementation.
Resumo:
Advancements in cloud computing have enabled the proliferation of distributed applications, which require management and control of multiple services. However, without an efficient mechanism for scaling services in response to changing workload conditions, such as number of connected users, application performance might suffer, leading to violations of Service Level Agreements (SLA) and possible inefficient use of hardware resources. Combining dynamic application requirements with the increased use of virtualised computing resources creates a challenging resource Management context for application and cloud-infrastructure owners. In such complex environments, business entities use SLAs as a means for specifying quantitative and qualitative requirements of services. There are several challenges in running distributed enterprise applications in cloud environments, ranging from the instantiation of service VMs in the correct order using an adequate quantity of computing resources, to adapting the number of running services in response to varying external loads, such as number of users. The application owner is interested in finding the optimum amount of computing and network resources to use for ensuring that the performance requirements of all her/his applications are met. She/he is also interested in appropriately scaling the distributed services so that application performance guarantees are maintained even under dynamic workload conditions. Similarly, the infrastructure Providers are interested in optimally provisioning the virtual resources onto the available physical infrastructure so that her/his operational costs are minimized, while maximizing the performance of tenants’ applications. Motivated by the complexities associated with the management and scaling of distributed applications, while satisfying multiple objectives (related to both consumers and providers of cloud resources), this thesis proposes a cloud resource management platform able to dynamically provision and coordinate the various lifecycle actions on both virtual and physical cloud resources using semantically enriched SLAs. The system focuses on dynamic sizing (scaling) of virtual infrastructures composed of virtual machines (VM) bounded application services. We describe several algorithms for adapting the number of VMs allocated to the distributed application in response to changing workload conditions, based on SLA-defined performance guarantees. We also present a framework for dynamic composition of scaling rules for distributed service, which used benchmark-generated application Monitoring traces. We show how these scaling rules can be combined and included into semantic SLAs for controlling allocation of services. We also provide a detailed description of the multi-objective infrastructure resource allocation problem and various approaches to satisfying this problem. We present a resource management system based on a genetic algorithm, which performs allocation of virtual resources, while considering the optimization of multiple criteria. We prove that our approach significantly outperforms reactive VM-scaling algorithms as well as heuristic-based VM-allocation approaches.
Resumo:
This paper shows that optimal policy and consistent policy outcomes require the use of control-theory and game-theory solution techniques. While optimal policy and consistent policy often produce different outcomes even in a one-period model, we analyze consistent policy and its outcome in a simple model, finding that the cause of the inconsistency with optimal policy traces to inconsistent targets in the social loss function. As a result, the social loss function cannot serve as a direct loss function for the central bank. Accordingly, we employ implementation theory to design a central bank loss function (mechanism design) with consistent targets, while the social loss function serves as a social welfare criterion. That is, with the correct mechanism design for the central bank loss function, optimal policy and consistent policy become identical. In other words, optimal policy proves implementable (consistent).
Resumo:
In this paper we generalize the Continuous Adversarial Queuing Theory (CAQT) model (Blesa et al. in MFCS, Lecture Notes in Computer Science, vol. 3618, pp. 144–155, 2005) by considering the possibility that the router clocks in the network are not synchronized. We name the new model Non Synchronized CAQT (NSCAQT). Clearly, this new extension to the model only affects those scheduling policies that use some form of timing. In a first approach we consider the case in which although not synchronized, all clocks run at the same speed, maintaining constant differences. In this case we show that all universally stable policies in CAQT that use the injection time and the remaining path to schedule packets remain universally stable. These policies include, for instance, Shortest in System (SIS) and Longest in System (LIS). Then, we study the case in which clock differences can vary over time, but the maximum difference is bounded. In this model we show the universal stability of two families of policies related to SIS and LIS respectively (the priority of a packet in these policies depends on the arrival time and a function of the path traversed). The bounds we obtain in this case depend on the maximum difference between clocks. This is a necessary requirement, since we also show that LIS is not universally stable in systems without bounded clock difference. We then present a new policy that we call Longest in Queues (LIQ), which gives priority to the packet that has been waiting the longest in edge queues. This policy is universally stable and, if clocks maintain constant differences, the bounds we prove do not depend on them. To finish, we provide with simulation results that compare the behavior of some of these policies in a network with stochastic injection of packets.
Resumo:
We introduce in this paper a method to calculate the Hessenberg matrix of a sum of measures from the Hessenberg matrices of the component measures. Our method extends the spectral techniques used by G. Mantica to calculate the Jacobi matrix associated with a sum of measures from the Jacobi matrices of each of the measures. We apply this method to approximate the Hessenberg matrix associated with a self-similar measure and compare it with the result obtained by a former method for self-similar measures which uses a fixed point theorem for moment matrices. Results are given for a series of classical examples of self-similar measures. Finally, we also apply the method introduced in this paper to some examples of sums of (not self-similar) measures obtaining the exact value of the sections of the Hessenberg matrix.
Resumo:
The European Credit Transfer and Accumulation System (ECTS) is the credit system for higher education used in the European Higher Education Area (EHEA), which involves all the countries engaged in the Bologna Process. This paper describes a study which is part of the project of the Bologna Experts Team-Spain and was carried out with the following aims: 1) designing some procedures for the assessment of transferable competences; and 2) testing some basic psychometric features that an assessment device with some consequences for the subjects being evaluated needs to prove. We will focus on the degrees of Computing. The sample of students (20) includes first year students from the Technical University of Madrid. In this paper, we will report some results of data analyses carried out to this moment on reliability and validity of the task designed to measure problem solving.
Resumo:
In this paper we generalize the Continuous Adversarial Queuing Theory (CAQT) model (Blesa et al. in MFCS, Lecture Notes in Computer Science, vol. 3618, pp. 144–155, 2005) by considering the possibility that the router clocks in the network are not synchronized. We name the new model Non Synchronized CAQT (NSCAQT). Clearly, this new extension to the model only affects those scheduling policies that use some form of timing. In a first approach we consider the case in which although not synchronized, all clocks run at the same speed, maintaining constant differences. In this case we show that all universally stable policies in CAQT that use the injection time and the remaining path to schedule packets remain universally stable. These policies include, for instance, Shortest in System (SIS) and Longest in System (LIS). Then, we study the case in which clock differences can vary over time, but the maximum difference is bounded. In this model we show the universal stability of two families of policies related to SIS and LIS respectively (the priority of a packet in these policies depends on the arrival time and a function of the path traversed). The bounds we obtain in this case depend on the maximum difference between clocks. This is a necessary requirement, since we also show that LIS is not universally stable in systems without bounded clock difference. We then present a new policy that we call Longest in Queues (LIQ), which gives priority to the packet that has been waiting the longest in edge queues. This policy is universally stable and, if clocks maintain constant differences, the bounds we prove do not depend on them. To finish, we provide with simulation results that compare the behavior of some of these policies in a network with stochastic injection of packets.