816 resultados para Embedded System, Domain Specific Language (DSL), Agenti BDI, Arduino, Agentino
Resumo:
Currently, Portugal assumes itself as a democratic rule of substantive law State, sustained by a legal system seeking the right balance between the guarantee of fundamental rights and freedoms constitutional foreseen in Portugal’s Fundamental Law and criminal persecution. The architecture of the penal code lies with, roughly speaking, a accusatory basic structure, “deliberately attached to one of the most remarkable achievements of the civilizational democratic progress, and by obedience to the constitutional commandment”, in balance with the official investigation principle, valid both for the purpose of prosecution and trial. Regarding the principle of non self-incrimination - nemo tenetur se ipsum accusare, briefly defined as the defendant’s right of not being obliged to contribute to the self-incrimination, it should be stressed that there isn’t an explicit consecration in the Portuguese Constitution, being commonly accepted in an implicit constitutional prediction and deriving from other constitutional rights and principles, first and foremost, the meaning and scope of the concept of democratic rule of Law State, embedded in the Fundamental Law, and in the guidelines of the constitutional principles of human person dignity, freedom of action and the presumption of innocence. In any case, about the (in) applicability of the principle of the prohibition of self-incrimination to the Criminal Police Bodies in the trial hearing in Court, and sharing an idea of Guedes Valente, the truth is that the exercise of criminal action must tread a transparent path and non-compliant with methods to obtain evidence that violate the law, the public order or in violation of democratic principles and loyalty (Guedes Valente, 2013, p. 484). Within the framework of the penal process relating to the trial, which is assumed as the true phase of the process, the witness represents a relevant figure for the administration of criminal justice, for the testimonial proof is, in the idea of Othmar Jauernig, the worst proof of evidence, but also being the most frequent (Jauernig, 1998, p. 289). As coadjutant of the Public Prosecutor and, in specific cases, the investigating judge, the Criminal Police Bodies are invested with high responsibility, being "the arms and eyes of Judicial Authorities in pursuing the criminal investigation..." which has as ultimate goal the fulfillment of the Law pursuing the defense of society" (Guedes Valente, 2013, p. 485). It is in this context and as a witness that, throughout operational career, the Criminal Police Bodies are required to be at the trial hearing and clarify the Court with its view about the facts relating to occurrences of criminal context, thus contributing very significantly and, in some cases, decisively for the proper administration of the portuguese criminal justice. With regards to the intervention of Criminal Police Bodies in the trial hearing in Court, it’s important that they pay attention to a set of standards concerning the preparation of the testimony, the very provision of the testimony and, also, to its conclusion. Be emphasized that these guidelines may become crucial for the quality of the police testimony at the trial hearing, thus leading to an improvement of the enforcement of justice system. In this vein, while preparing the testimony, the Criminal Police Bodies must present itself in court with proper clothing, to read before and carefully the case files, to debate the facts being judged with other Criminal Police Bodies and prepare potential questions. Later, while giving his testimony during the trial, the Criminal Police Bodies must, summing up, to take the oath in a convincing manner, to feel comfortable, to start well by convincingly answering the first question, keep an attitude of serenity, to adopt an attitude of collaboration, to avoid the reading of documents, to demonstrate deference and seriousness before the judicial operators, to use simple and objective language, to adopt a fluent speech, to use nonverbal language correctly, to avoid spontaneity responding only to what is asked, to report only the truth, to avoid hesitations and contradictions, to be impartial and to maintain eye contact with the judge. Finally, at the conclusion of the testimony, the Criminal Police Bodies should rise in a smooth manner, avoiding to show relief, resentment or satisfaction, leaving a credible and professional image and, without much formality, requesting the judge permission to leave the courtroom. As final note, it’s important to stress that "The intervention of the Police Criminal Bodies in the trial hearing in Court” encloses itself on a theme of crucial importance not only for members of the Police and Security Forces, who must welcome this subject with the utmost seriousness and professionalism, but also for the proper administration of the criminal justice system in Portugal.
Resumo:
If you want to know whether a property is true or not in a specific algebraic structure,you need to test that property on the given structure. This can be done by hand, which can be cumbersome and erroneous. In addition, the time consumed in testing depends on the size of the structure where the property is applied. We present an implementation of a system for finding counterexamples and testing properties of models of first-order theories. This system is supposed to provide a convenient and paperless environment for researchers and students investigating or studying such models and algebraic structures in particular. To implement a first-order theory in the system, a suitable first-order language.( and some axioms are required. The components of a language are given by a collection of variables, a set of predicate symbols, and a set of operation symbols. Variables and operation symbols are used to build terms. Terms, predicate symbols, and the usual logical connectives are used to build formulas. A first-order theory now consists of a language together with a set of closed formulas, i.e. formulas without free occurrences of variables. The set of formulas is also called the axioms of the theory. The system uses several different formats to allow the user to specify languages, to define axioms and theories and to create models. Besides the obvious operations and tests on these structures, we have introduced the notion of a functor between classes of models in order to generate more co~plex models from given ones automatically. As an example, we will use the system to create several lattices structures starting from a model of the theory of pre-orders.
Resumo:
Affiliation: Sophie Broussau, Amelie Pilotte & Bernard Massie : Départment de microbiologie et immunologie, Faculté de médecine, Université de Montréal
Resumo:
L'utilisation des méthodes formelles est de plus en plus courante dans le développement logiciel, et les systèmes de types sont la méthode formelle qui a le plus de succès. L'avancement des méthodes formelles présente de nouveaux défis, ainsi que de nouvelles opportunités. L'un des défis est d'assurer qu'un compilateur préserve la sémantique des programmes, de sorte que les propriétés que l'on garantit à propos de son code source s'appliquent également au code exécutable. Cette thèse présente un compilateur qui traduit un langage fonctionnel d'ordre supérieur avec polymorphisme vers un langage assembleur typé, dont la propriété principale est que la préservation des types est vérifiée de manière automatisée, à l'aide d'annotations de types sur le code du compilateur. Notre compilateur implante les transformations de code essentielles pour un langage fonctionnel d'ordre supérieur, nommément une conversion CPS, une conversion des fermetures et une génération de code. Nous présentons les détails des représentation fortement typées des langages intermédiaires, et les contraintes qu'elles imposent sur l'implantation des transformations de code. Notre objectif est de garantir la préservation des types avec un minimum d'annotations, et sans compromettre les qualités générales de modularité et de lisibilité du code du compilateur. Cet objectif est atteint en grande partie dans le traitement des fonctionnalités de base du langage (les «types simples»), contrairement au traitement du polymorphisme qui demande encore un travail substantiel pour satisfaire la vérification de type.
Resumo:
Le Virus Herpès Simplex de type 1 (HSV-1) est un agent infectieux qui cause l’herpès chez une grande proportion de la population mondiale. L’herpès est généralement considéré comme une maladie bénigne dont la forme la plus commune est l'herpès labial (communément appelé « bouton de fièvre »), mais elle peut se révéler très sérieuse et causer la cécité et l’encéphalite, voir létale dans certain cas. Le virus persiste toute la vie dans le corps de son hôte. Jusqu'à présent, aucun traitement ne peut éliminer le virus et aucun vaccin n’a été prouvé efficace pour contrôler l’infection herpétique. HSV-1 est un virus avec un génome d’ADN bicaténaire contenu dans une capside icosaèdrale entourée d’une enveloppe lipidique. Treize glycoprotéines virales se trouvent dans cette enveloppe et sont connues ou supposées jouer des rôles distincts dans différentes étapes du cycle de réplication viral, incluant l'attachement, l'entrée, l’assemblage, et la propagation des virus. La glycoprotéine M (gM) qui figure parmi ces glycoprotéines d’enveloppe, est la seule glycoprotéine non essentielle mais est conservée dans toute la famille herpesviridae. Récemment, l’homologue de gM dans le Pseudorabies virus (PRV), un autre herpesvirus, a été impliqué dans la phase finale de l’assemblage (i.e. l’enveloppement cytoplasmique) au niveau du réseau trans-Golgi (TGN) en reconnaissant spécifiquement des protéines tégumentaires et d’autres glycoprotéines d’enveloppe ([1]). Toutefois, il a été proposé que cette hypothèse ne s’applique pas pour le HSV-1 ([2]). De plus, contrairement à la localisation au TGN dans les cellules transfectées, HSV-1 gM se localise dans la membrane nucléaire et sur les virions périnucléaires durant une infection. L’objectif du projet présenté ici était d’éclaircir la relation de la localisation et la fonction de HSV-1 gM dans le contexte d’une infection. Dans les résultats rapportés ici, nous décrivons tout abord un mécanisme spécifique de ciblage nucléaire de HSV-1 gM. En phase précoce d’une infection, gM est ciblée à la membrane nucléaire d'une manière virus ii dépendante. Cela se produit avant la réorganisation du TGN normalement induite par l’infection et avant que gM n’entre dans la voie de sécrétion. Ce ciblage nucléaire actif et spécifique de gM ne semble pas dépendre des plusieurs des partenaires d’interaction proposés dans la littérature. Ces données suggèrent que la forme nucléaire de gM pourrait avoir un nouveau rôle indépendant de l’enveloppement final dans le cytoplasme. Dans la deuxième partie du travail présenté ici, nous avons concentré nos efforts sur le rôle de gM dans l’assemblage du virus en phase tardive de l’infection et en identifiant un domaine critique de gM. Nos résultats mettent en valeur l’importance du domaine carboxyl-terminal cytoplasmique de gM dans le transport de gM du réticulum endoplasmique (RE) à l’appareil de Golgi, dans l’enveloppement cytoplasmique et la propagation intercellulaire du virus. Ainsi, l’export du RE de gM a été complètement compromis dans les cellules transfectées exprimant un mutant de gM dépourvu de sa région C-terminale. La délétion la queue cytoplasmique de gM cause une réduction légère du titre viral et de la taille des plaques. L'analyse de ces mutants par microscopie électronique a démontré une accumulation des nucléocapsides sans enveloppe dans le cytoplasme par rapport aux virus de type sauvage. Étrangement, ce phénotype était apparent dans les cellules BHK mais absent dans les cellules 143B, suggérant que la fonction de gM dépende du type cellulaire. Finalement, le criblage de partenaires d’interaction du domaine C-terminal de gM identifiés par le système de double-hybride nous a permis de proposer plusieurs candidats susceptibles de réguler la fonction de gM dans la morphogénèse et la propagation de virus.
Resumo:
Le but de cette thèse est d'étudier les corrélats comportementaux et neuronaux du transfert inter-linguistique (TIL) dans l'apprentissage d’une langue seconde (L2). Compte tenu de nos connaissances sur l'influence de la distance linguistique sur le TIL (Paradis, 1987, 2004; Odlin, 1989, 2004, 2005; Gollan, 2005; Ringbom, 2007), nous avons examiné l'effet de facilitation de la similarité phonologique à l’aide de la résonance magnétique fonctionnelle entre des langues linguistiquement proches (espagnol-français) et des langues linguistiquement éloignées (persan-français). L'étude I rapporte les résultats obtenus pour des langues linguistiquement proches (espagnol-français), alors que l'étude II porte sur des langues linguistiquement éloignées (persan-français). Puis, les changements de connectivité fonctionnelle dans le réseau langagier (Price, 2010) et dans le réseau de contrôle supplémentaire impliqué dans le traitement d’une langue seconde (Abutalebi & Green, 2007) lors de l’apprentissage d’une langue linguistiquement éloignée (persan-français) sont rapportés dans l’étude III. Les résultats des analyses d’IRMF suivant le modèle linéaire général chez les bilingues de langues linguistiquement proches (français-espagnol) montrent que le traitement des mots phonologiquement similaires dans les deux langues (cognates et clangs) compte sur un réseau neuronal partagé par la langue maternelle (L1) et la L2, tandis que le traitement des mots phonologiquement éloignés (non-clang-non-cognates) active des structures impliquées dans le traitement de la mémoire de travail et d'attention. Toutefois, chez les personnes bilingues de L1-L2 linguistiquement éloignées (français-persan), même les mots phonologiquement similaires à travers les langues (cognates et clangs) activent des régions connues pour être impliquées dans l'attention et le contrôle cognitif. Par ailleurs, les mots phonologiquement éloignés (non-clang-non-cognates) activent des régions usuellement associées à la mémoire de travail et aux fonctions exécutives. Ainsi, le facteur de distance inter-linguistique entre L1 et L2 module la charge cognitive sur la base du degré de similarité phonologiques entres les items en L1 et L2. Des structures soutenant les processus impliqués dans le traitement exécutif sont recrutées afin de compenser pour des demandes cognitives. Lorsque la compétence linguistique en L2 augmente et que les tâches linguistiques exigent ainsi moins d’effort, la demande pour les ressources cognitives diminue. Tel que déjà rapporté (Majerus, et al, 2008; Prat, et al, 2007; Veroude, et al, 2010; Dodel, et al, 2005; Coynel, et al ., 2009), les résultats des analyses de connectivité fonctionnelle montrent qu’après l’entraînement la valeur d'intégration (connectivité fonctionnelle) diminue puisqu’il y a moins de circulation du flux d'information. Les résultats de cette recherche contribuent à une meilleure compréhension des aspects neurocognitifs et de plasticité cérébrale du TIL ainsi que l'impact de la distance linguistique dans l'apprentissage des langues. Ces résultats ont des implications dans les stratégies d'apprentissage d’une L2, les méthodes d’enseignement d’une L2 ainsi que le développement d'approches thérapeutiques chez des patients bilingues qui souffrent de troubles langagiers.
Resumo:
This paper presents the design and development of a frame based approach for speech to sign language machine translation system in the domain of railways and banking. This work aims to utilize the capability of Artificial intelligence for the improvement of physically challenged, deaf-mute people. Our work concentrates on the sign language used by the deaf community of Indian subcontinent which is called Indian Sign Language (ISL). Input to the system is the clerk’s speech and the output of this system is a 3D virtual human character playing the signs for the uttered phrases. The system builds up 3D animation from pre-recorded motion capture data. Our work proposes to build a Malayalam to ISL
Resumo:
This paper presents a novel approach to recognize Grantha, an ancient script in South India and converting it to Malayalam, a prevalent language in South India using online character recognition mechanism. The motivation behind this work owes its credit to (i) developing a mechanism to recognize Grantha script in this modern world and (ii) affirming the strong connection among Grantha and Malayalam. A framework for the recognition of Grantha script using online character recognition is designed and implemented. The features extracted from the Grantha script comprises mainly of time-domain features based on writing direction and curvature. The recognized characters are mapped to corresponding Malayalam characters. The framework was tested on a bed of medium length manuscripts containing 9-12 sample lines and printed pages of a book titled Soundarya Lahari writtenin Grantha by Sri Adi Shankara to recognize the words and sentences. The manuscript recognition rates with the system are for Grantha as 92.11%, Old Malayalam 90.82% and for new Malayalam script 89.56%. The recognition rates of pages of the printed book are for Grantha as 96.16%, Old Malayalam script 95.22% and new Malayalam script as 92.32% respectively. These results show the efficiency of the developed system
Resumo:
Geochemical composition is a set of data for predicting the climatic condition existing in an ecosystem. Both the surficial and core sediment geochemistry are helpful in monitoring, assessing and evaluating the marine environment. The aim of the research work is to assess the relationship between the biogeochemical constituents in the Cochin Estuarine System (CES), their modifications after a long period of anoxia and also to identify the various processes which control the sediment composition in this region, through a multivariate statistical approach. Therefore the study of present core sediment geochemistry has a critical role in unraveling the benchmark of their characterization. Sediment cores from four prominent zones of CES were examined for various biogeochemical aspects. The results have served as rejuvenating records for the prediction of core sediment status prevailing in the CES
Resumo:
Land use is a crucial link between human activities and the natural environment and one of the main driving forces of global environmental change. Large parts of the terrestrial land surface are used for agriculture, forestry, settlements and infrastructure. Given the importance of land use, it is essential to understand the multitude of influential factors and resulting land use patterns. An essential methodology to study and quantify such interactions is provided by the adoption of land-use models. By the application of land-use models, it is possible to analyze the complex structure of linkages and feedbacks and to also determine the relevance of driving forces. Modeling land use and land use changes has a long-term tradition. In particular on the regional scale, a variety of models for different regions and research questions has been created. Modeling capabilities grow with steady advances in computer technology, which on the one hand are driven by increasing computing power on the other hand by new methods in software development, e.g. object- and component-oriented architectures. In this thesis, SITE (Simulation of Terrestrial Environments), a novel framework for integrated regional sland-use modeling, will be introduced and discussed. Particular features of SITE are the notably extended capability to integrate models and the strict separation of application and implementation. These features enable efficient development, test and usage of integrated land-use models. On its system side, SITE provides generic data structures (grid, grid cells, attributes etc.) and takes over the responsibility for their administration. By means of a scripting language (Python) that has been extended by language features specific for land-use modeling, these data structures can be utilized and manipulated by modeling applications. The scripting language interpreter is embedded in SITE. The integration of sub models can be achieved via the scripting language or by usage of a generic interface provided by SITE. Furthermore, functionalities important for land-use modeling like model calibration, model tests and analysis support of simulation results have been integrated into the generic framework. During the implementation of SITE, specific emphasis was laid on expandability, maintainability and usability. Along with the modeling framework a land use model for the analysis of the stability of tropical rainforest margins was developed in the context of the collaborative research project STORMA (SFB 552). In a research area in Central Sulawesi, Indonesia, socio-environmental impacts of land-use changes were examined. SITE was used to simulate land-use dynamics in the historical period of 1981 to 2002. Analogous to that, a scenario that did not consider migration in the population dynamics, was analyzed. For the calculation of crop yields and trace gas emissions, the DAYCENT agro-ecosystem model was integrated. In this case study, it could be shown that land-use changes in the Indonesian research area could mainly be characterized by the expansion of agricultural areas at the expense of natural forest. For this reason, the situation had to be interpreted as unsustainable even though increased agricultural use implied economic improvements and higher farmers' incomes. Due to the importance of model calibration, it was explicitly addressed in the SITE architecture through the introduction of a specific component. The calibration functionality can be used by all SITE applications and enables largely automated model calibration. Calibration in SITE is understood as a process that finds an optimal or at least adequate solution for a set of arbitrarily selectable model parameters with respect to an objective function. In SITE, an objective function typically is a map comparison algorithm capable of comparing a simulation result to a reference map. Several map optimization and map comparison methodologies are available and can be combined. The STORMA land-use model was calibrated using a genetic algorithm for optimization and the figure of merit map comparison measure as objective function. The time period for the calibration ranged from 1981 to 2002. For this period, respective reference land-use maps were compiled. It could be shown, that an efficient automated model calibration with SITE is possible. Nevertheless, the selection of the calibration parameters required detailed knowledge about the underlying land-use model and cannot be automated. In another case study decreases in crop yields and resulting losses in income from coffee cultivation were analyzed and quantified under the assumption of four different deforestation scenarios. For this task, an empirical model, describing the dependence of bee pollination and resulting coffee fruit set from the distance to the closest natural forest, was integrated. Land-use simulations showed, that depending on the magnitude and location of ongoing forest conversion, pollination services are expected to decline continuously. This results in a reduction of coffee yields of up to 18% and a loss of net revenues per hectare of up to 14%. However, the study also showed that ecological and economic values can be preserved if patches of natural vegetation are conservated in the agricultural landscape. -----------------------------------------------------------------------
Resumo:
Wednesday 12th March 2014 Speaker(s): Dr Tim Chown Organiser: Time: 12/03/2014 11:00-11:50 Location: B32/3077 File size: 642 Mb Abstract The WAIS seminar series is designed to be a blend of classic seminars, research discussions, debates and tutorials. The Domain Name System (DNS) is a critical part of the Internet infrastructure. In this talk we begin by explaining the basic model of operation of the DNS, including how domain names are delegated and how a DNS resolver performs a DNS lookup. We then take a tour of DNS-related topics, including caching, poisoning, governance, the increasing misuse of the DNS in DDoS attacks, and the expansion of the DNS namespace to new top level domains and internationalised domain names. We also present the latest work in the IETF on DNS privacy. The talk will be pitched such that no detailed technical knowledge is required. We hope that attendees will gain some familiarity with how the DNS works, some key issues surrounding DNS operation, and how the DNS might touch on various areas of research within WAIS.
Resumo:
La gestió de l'aigua residual és una tasca complexa. Hi ha moltes substàncies contaminants conegudes però encara moltes per conèixer, i el seu efecte individual o col·lgectiu és difícil de predir. La identificació i avaluació dels impactes ambientals resultants de la interacció entre els sistemes naturals i socials és un assumpte multicriteri. Els gestors ambientals necessiten eines de suport pels seus diagnòstics per tal de solucionar problemes ambientals. Les contribucions d'aquest treball de recerca són dobles: primer, proposar l'ús d'un enfoc basat en la modelització amb agents per tal de conceptualitzar i integrar tots els elements que estan directament o indirectament involucrats en la gestió de l'aigua residual. Segon, proposar un marc basat en l'argumentació amb l'objectiu de permetre als agents raonar efectivament. La tesi conté alguns exemples reals per tal de mostrar com un marc basat amb agents que argumenten pot suportar diferents interessos i diferents perspectives. Conseqüentment, pot ajudar a construir un diàleg més informat i efectiu i per tant descriure millor les interaccions entre els agents. En aquest document es descriu primer el context estudiat, escalant el problema global de la gestió de la conca fluvial a la gestiódel sistema urbà d'aigües residuals, concretament l'escenari dels abocaments industrials. A continuació, s'analitza el sistema mitjançant la descripció d'agents que interaccionen. Finalment, es descriuen alguns prototips capaços de raonar i deliberar, basats en la lògica no monòtona i en un llenguatge declaratiu (answer set programming). És important remarcar que aquesta tesi enllaça dues disciplines: l'enginyeria ambiental (concretament l'àrea de la gestió de les aigües residuals) i les ciències de la computació (concretament l'àrea de la intel·ligència artificial), contribuint així a la multidisciplinarietat requerida per fer front al problema estudiat. L'enginyeria ambiental ens proporciona el coneixement del domini mentre que les ciències de la computació ens permeten estructurar i especificar aquest coneixement.
Resumo:
The proposal presented in this thesis is to provide designers of knowledge based supervisory systems of dynamic systems with a framework to facilitate their tasks avoiding interface problems among tools, data flow and management. The approach is thought to be useful to both control and process engineers in assisting their tasks. The use of AI technologies to diagnose and perform control loops and, of course, assist process supervisory tasks such as fault detection and diagnose, are in the scope of this work. Special effort has been put in integration of tools for assisting expert supervisory systems design. With this aim the experience of Computer Aided Control Systems Design (CACSD) frameworks have been analysed and used to design a Computer Aided Supervisory Systems (CASSD) framework. In this sense, some basic facilities are required to be available in this proposed framework: ·
Resumo:
The aim of this thesis is to narrow the gap between two different control techniques: the continuous control and the discrete event control techniques DES. This gap can be reduced by the study of Hybrid systems, and by interpreting as Hybrid systems the majority of large-scale systems. In particular, when looking deeply into a process, it is often possible to identify interaction between discrete and continuous signals. Hybrid systems are systems that have both continuous, and discrete signals. Continuous signals are generally supposed continuous and differentiable in time, since discrete signals are neither continuous nor differentiable in time due to their abrupt changes in time. Continuous signals often represent the measure of natural physical magnitudes such as temperature, pressure etc. The discrete signals are normally artificial signals, operated by human artefacts as current, voltage, light etc. Typical processes modelled as Hybrid systems are production systems, chemical process, or continuos production when time and continuous measures interacts with the transport, and stock inventory system. Complex systems as manufacturing lines are hybrid in a global sense. They can be decomposed into several subsystems, and their links. Another motivation for the study of Hybrid systems is the tools developed by other research domains. These tools benefit from the use of temporal logic for the analysis of several properties of Hybrid systems model, and use it to design systems and controllers, which satisfies physical or imposed restrictions. This thesis is focused in particular types of systems with discrete and continuous signals in interaction. That can be modelled hard non-linealities, such as hysteresis, jumps in the state, limit cycles, etc. and their possible non-deterministic future behaviour expressed by an interpretable model description. The Hybrid systems treated in this work are systems with several discrete states, always less than thirty states (it can arrive to NP hard problem), and continuous dynamics evolving with expression: with Ki ¡ Rn constant vectors or matrices for X components vector. In several states the continuous evolution can be several of them Ki = 0. In this formulation, the mathematics can express Time invariant linear system. By the use of this expression for a local part, the combination of several local linear models is possible to represent non-linear systems. And with the interaction with discrete events of the system the model can compose non-linear Hybrid systems. Especially multistage processes with high continuous dynamics are well represented by the proposed methodology. Sate vectors with more than two components, as third order models or higher is well approximated by the proposed approximation. Flexible belt transmission, chemical reactions with initial start-up and mobile robots with important friction are several physical systems, which profits from the benefits of proposed methodology (accuracy). The motivation of this thesis is to obtain a solution that can control and drive the Hybrid systems from the origin or starting point to the goal. How to obtain this solution, and which is the best solution in terms of one cost function subject to the physical restrictions and control actions is analysed. Hybrid systems that have several possible states, different ways to drive the system to the goal and different continuous control signals are problems that motivate this research. The requirements of the system on which we work is: a model that can represent the behaviour of the non-linear systems, and that possibilities the prediction of possible future behaviour for the model, in order to apply an supervisor which decides the optimal and secure action to drive the system toward the goal. Specific problems can be determined by the use of this kind of hybrid models are: - The unity of order. - Control the system along a reachable path. - Control the system in a safe path. - Optimise the cost function. - Modularity of control The proposed model solves the specified problems in the switching models problem, the initial condition calculus and the unity of the order models. Continuous and discrete phenomena are represented in Linear hybrid models, defined with defined eighth-tuple parameters to model different types of hybrid phenomena. Applying a transformation over the state vector : for LTI system we obtain from a two-dimensional SS a single parameter, alpha, which still maintains the dynamical information. Combining this parameter with the system output, a complete description of the system is obtained in a form of a graph in polar representation. Using Tagaki-Sugeno type III is a fuzzy model which include linear time invariant LTI models for each local model, the fuzzyfication of different LTI local model gives as a result a non-linear time invariant model. In our case the output and the alpha measure govern the membership function. Hybrid systems control is a huge task, the processes need to be guided from the Starting point to the desired End point, passing a through of different specific states and points in the trajectory. The system can be structured in different levels of abstraction and the control in three layers for the Hybrid systems from planning the process to produce the actions, these are the planning, the process and control layer. In this case the algorithms will be applied to robotics ¡V a domain where improvements are well accepted ¡V it is expected to find a simple repetitive processes for which the extra effort in complexity can be compensated by some cost reductions. It may be also interesting to implement some control optimisation to processes such as fuel injection, DC-DC converters etc. In order to apply the RW theory of discrete event systems on a Hybrid system, we must abstract the continuous signals and to project the events generated for these signals, to obtain new sets of observable and controllable events. Ramadge & Wonham¡¦s theory along with the TCT software give a Controllable Sublanguage of the legal language generated for a Discrete Event System (DES). Continuous abstraction transforms predicates over continuous variables into controllable or uncontrollable events, and modifies the set of uncontrollable, controllable observable and unobservable events. Continuous signals produce into the system virtual events, when this crosses the bound limits. If this event is deterministic, they can be projected. It is necessary to determine the controllability of this event, in order to assign this to the corresponding set, , controllable, uncontrollable, observable and unobservable set of events. Find optimal trajectories in order to minimise some cost function is the goal of the modelling procedure. Mathematical model for the system allows the user to apply mathematical techniques over this expression. These possibilities are, to minimise a specific cost function, to obtain optimal controllers and to approximate a specific trajectory. The combination of the Dynamic Programming with Bellman Principle of optimality, give us the procedure to solve the minimum time trajectory for Hybrid systems. The problem is greater when there exists interaction between adjacent states. In Hybrid systems the problem is to determine the partial set points to be applied at the local models. Optimal controller can be implemented in each local model in order to assure the minimisation of the local costs. The solution of this problem needs to give us the trajectory to follow the system. Trajectory marked by a set of set points to force the system to passing over them. Several ways are possible to drive the system from the Starting point Xi to the End point Xf. Different ways are interesting in: dynamic sense, minimum states, approximation at set points, etc. These ways need to be safe and viable and RchW. And only one of them must to be applied, normally the best, which minimises the proposed cost function. A Reachable Way, this means the controllable way and safe, will be evaluated in order to obtain which one minimises the cost function. Contribution of this work is a complete framework to work with the majority Hybrid systems, the procedures to model, control and supervise are defined and explained and its use is demonstrated. Also explained is the procedure to model the systems to be analysed for automatic verification. Great improvements were obtained by using this methodology in comparison to using other piecewise linear approximations. It is demonstrated in particular cases this methodology can provide best approximation. The most important contribution of this work, is the Alpha approximation for non-linear systems with high dynamics While this kind of process is not typical, but in this case the Alpha approximation is the best linear approximation to use, and give a compact representation.