779 resultados para Fuzzy Boolean Nets


Relevância:

20.00% 20.00%

Publicador:

Resumo:

There is controversy regarding the use of the similarity functions proposed in the literature to compare generalized trapezoidal fuzzy numbers since conflicting similarity values are sometimes output for the same pair of fuzzy numbers. In this paper we propose a similarity function aimed at establishing a consensus. It accounts for the different approaches of all the similarity functions. It also has better properties and can easily incorporate new parameters for future improvements. The analysis is carried out on the basis of a large and representative set of pairs of trapezoidal fuzzy numbers.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La tesis doctoral "Contribución al estudio y clasificación de las funciones de implicación borrosas" constituye a la vez una revisión y un conjunto de aportaciones a la modelización de enunciados condicionales, o enunciados del tipo "Si P, entonces Q". Dentro del marco de la lógica borrosa, tradicionalmente, se considera que la mayor parte de las funciones de implicación, que modelizan los enunciados condicionales, deberían generalizar la implicación material booleana. En esta memoria se apoya el argumento de que la implicación material booleana no es siempre el modelo más adecuado para la modelización de los enunciados condicionales, por lo que se definen y estudian clases o grupos de funciones de implicación que se adecúen a las necesidades de cada caso de aplicación. Así pues, tras un capítulo introductorio, en el capítulo 2 se plantean clases de funciones de implicación que sirvan de apoyo en la definición de aplicaciones de carácter borroso más allá de las funciones de implicación borrosas derivadas de la implicación material booleana. Así se llega a la conclusión de que son necesarios, por lo menos, cuatro clases de funciones de implicación de las que se estudian sus propiedades. En el tercer capítulo se aborda el estudio de la verificación tanto de la regla del Modus Ponens como de la regla del Modus Tollens por parte de las funciones de implicación pertenecientes a cada una de las clases planteadas. Por último, en el capítulo 4 se presenta un estudio de los operadores citados como implicaciones borrosas en la literatura bajo el enfoque presentado en esta memoria, ampliándose estudios realizados por otros autores. ---ABSTRACT--- The Ph.D. Thesis "Contribution al estudio y clasificacion de las funciones de implication borrosas" is a revision and a set of contributions to the conditional statements or statements of the type "If P then Q" modelization. It has been usually considered in Fuzzy Logic, that implication functions that model conditional statements should generalize the boolean material implication. The present memory supports the argument that boolean material implication is not always the most appropiated model for conditional statements, so implication function groups or clusters are denned in order to fit in each application. Therefore, after an introductory chapter, in chapter 2 groups or clusters of implication functions are given in order to bear out fuzzy applications definitions further on from implication functions coming from boolean material implication. Then, it has been reached the conclusion that, at least, four types of implication functions are needed so their properties are studied.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Abstract. Receptive fields of retinal and other sensory neurons show a large variety of spatiotemporal linear and non linear types of responses to local stimuli. In visual neurons, these responses present either asymmetric sensitive zones or center-surround organization. In most cases, the nature of the responses suggests the existence of a kind of distributed computation prior to the integration by the final cell which is evidently supported by the anatomy. We describe a new kind of discrete and continuous filters to model the kind of computations taking place in the receptive fields of retinal cells. To show their performance in the analysis of diferent non-trivial neuron-like structures, we use a computer tool specifically programmed by the authors to that efect. This tool is also extended to study the efect of lesions on the whole performance of our model nets.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we focus on the selection of safeguards in a fuzzy risk analysis and management methodology for information systems (IS). Assets are connected by dependency relationships, and a failure of one asset may affect other assets. After computing impact and risk indicators associated with previously identified threats, we identify and apply safeguards to reduce risks in the IS by minimizing the transmission probabilities of failures throughout the asset network. However, as safeguards have associated costs, the aim is to select the safeguards that minimize costs while keeping the risk within acceptable levels. To do this, we propose a dynamic programming-based method that incorporates simulated annealing to tackle optimizations problems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Pridneprovsky Chemical Plant was a largest uranium processing enterprises, producing a huge amount of uranium residues. The Zapadnoe tailings site contains the majority of these residues. We propose a theoretical framework based on Multi-Criteria Decision Analysis and fuzzy logic to analyse different remediation alternatives for the Zapadnoe tailings, in which potentially conflicting economic, radiological, social and environmental objectives are simultaneously taken into account. An objective hierarchy is built that includes all the relevant aspects. Fuzzy rather than precise values are proposed for use to evaluate remediation alternatives against the different criteria and to quantify preferences, such as the weights representing the relative importance of criteria identified in the objective hierarchy. Finally, it is proposed that remediation alternatives should be evaluated by means of a fuzzy additive multi-attribute utility function and ranked on the basis of the respective trapezoidal fuzzy number representing their overall utility.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This work aims to develop a novel Cross-Entropy (CE) optimization-based fuzzy controller for Unmanned Aerial Monocular Vision-IMU System (UAMVIS) to solve the seeand-avoid problem using its accurate autonomous localization information. The function of this fuzzy controller is regulating the heading of this system to avoid the obstacle, e.g. wall. In the Matlab Simulink-based training stages, the Scaling Factor (SF) is adjusted according to the specified task firstly, and then the Membership Function (MF) is tuned based on the optimized Scaling Factor to further improve the collison avoidance performance. After obtained the optimal SF and MF, 64% of rules has been reduced (from 125 rules to 45 rules), and a large number of real flight tests with a quadcopter have been done. The experimental results show that this approach precisely navigates the system to avoid the obstacle. To our best knowledge, this is the first work to present the optimized fuzzy controller for UAMVIS using Cross-Entropy method in Scaling Factors and Membership Functions optimization.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

En el trabajo que aquí presentamos se incluye la base teórica (sintaxis y semántica) y una implementación de un framework para codificar el razonamiento de la representación difusa o borrosa del mundo (tal y como nosotros, seres humanos, entendemos éste). El interés en la realización de éste trabajo parte de dos fuentes: eliminar la complejidad existente cuando se realiza una implementación con un lenguaje de programación de los llamados de propósito general y proporcionar una herramienta lo suficientemente inteligente para dar respuestas de forma constructiva a consultas difusas o borrosas. El framework, RFuzzy, permite codificar reglas y consultas en una sintaxis muy cercana al lenguaje natural usado por los seres humanos para expresar sus pensamientos, pero es bastante más que eso. Permite representar conceptos muy interesantes, como fuzzificaciones (funciones usadas para convertir conceptos no difusos en difusos), valores por defecto (que se usan para devolver resultados un poco menos válidos que los que devolveríamos si tuviésemos la información necesaria para calcular los más válidos), similaridad entre atributos (característica que utilizamos para buscar aquellos individuos en la base de datos con una característica similar a la buscada), sinónimos o antónimos y, además, nos permite extender el numero de conectivas y modificadores (incluyendo modificadores de negación) que podemos usar en las reglas y consultas. La personalización de la definición de conceptos difusos (muy útil para lidiar con el carácter subjetivo de los conceptos borrosos, donde nos encontramos con que cualificar a alguien de “alto” depende de la altura de la persona que cualifica) es otra de las facilidades incluida. Además, RFuzzy implementa la semántica multi-adjunta. El interés en esta reside en que introduce la posibilidad de obtener la credibilidad de una regla a partir de un conjunto de datos y una regla dada y no solo el grado de satisfacción de una regla a partir de el universo modelado en nuestro programa. De esa forma podemos obtener automáticamente la credibilidad de una regla para una determinada situación. Aún cuando la contribución teórica de la tesis es interesante en si misma, especialmente la inclusión del modificador de negacion, sus multiples usos practicos lo son también. Entre los diferentes usos que se han dado al framework destacamos el reconocimiento de emociones, el control de robots, el control granular en computacion paralela/distribuída y las busquedas difusas o borrosas en bases de datos. ABSTRACT In this work we provide a theoretical basis (syntax and semantics) and a practical implementation of a framework for encoding the reasoning and the fuzzy representation of the world (as human beings understand it). The interest for this work comes from two sources: removing the existing complexity when doing it with a general purpose programming language (one developed without focusing in providing special constructions for representing fuzzy information) and providing a tool intelligent enough to answer, in a constructive way, expressive queries over conventional data. The framework, RFuzzy, allows to encode rules and queries in a syntax very close to the natural language used by human beings to express their thoughts, but it is more than that. It allows to encode very interesting concepts, as fuzzifications (functions to easily fuzzify crisp concepts), default values (used for providing results less adequate but still valid when the information needed to provide results is missing), similarity between attributes (used to search for individuals with a characteristic similar to the one we are looking for), synonyms or antonyms and it allows to extend the number of connectives and modifiers (even negation) we can use in the rules. The personalization of the definition of fuzzy concepts (very useful for dealing with the subjective character of fuzziness, in which a concept like tall depends on the height of the person performing the query) is another of the facilities included. Besides, RFuzzy implements the multi-adjoint semantics. The interest in them is that in addition to obtaining the grade of satisfaction of a consequent from a rule, its credibility and the grade of satisfaction of the antecedents we can determine from a set of data how much credibility we must assign to a rule to model the behaviour of the set of data. So, we can determine automatically the credibility of a rule for a particular situation. Although the theoretical contribution is interesting by itself, specially the inclusion of the negation modifier, the practical usage of it is equally important. Between the different uses given to the framework we highlight emotion recognition, robocup control, granularity control in parallel/distributed computing and flexible searches in databases.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Recently, the cross-layer design for the wireless sensor network communication protocol has become more and more important and popular. Considering the disadvantages of the traditional cross-layer routing algorithms, in this paper we propose a new fuzzy logic-based routing algorithm, named the Balanced Cross-layer Fuzzy Logic (BCFL) routing algorithm. In BCFL, we use the cross-layer parameters’ dispersion as the fuzzy logic inference system inputs. Moreover, we give each cross-layer parameter a dynamic weight according the value of the dispersion. For getting a balanced solution, the parameter whose dispersion is large will have small weight, and vice versa. In order to compare it with the traditional cross-layer routing algorithms, BCFL is evaluated through extensive simulations. The simulation results show that the new routing algorithm can handle the multiple constraints without increasing the complexity of the algorithm and can achieve the most balanced performance on selecting the next hop relay node. Moreover, the Balanced Cross-layer Fuzzy Logic routing algorithm can adapt to the dynamic changing of the network conditions and topology effectively.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The solutions to cope with new challenges that societies have to face nowadays involve providing smarter daily systems. To achieve this, technology has to evolve and leverage physical systems automatic interactions, with less human intervention. Technological paradigms like Internet of Things (IoT) and Cyber-Physical Systems (CPS) are providing reference models, architectures, approaches and tools that are to support cross-domain solutions. Thus, CPS based solutions will be applied in different application domains like e-Health, Smart Grid, Smart Transportation and so on, to assure the expected response from a complex system that relies on the smooth interaction and cooperation of diverse networked physical systems. The Wireless Sensors Networks (WSN) are a well-known wireless technology that are part of large CPS. The WSN aims at monitoring a physical system, object, (e.g., the environmental condition of a cargo container), and relaying data to the targeted processing element. The WSN communication reliability, as well as a restrained energy consumption, are expected features in a WSN. This paper shows the results obtained in a real WSN deployment, based on SunSPOT nodes, which carries out a fuzzy based control strategy to improve energy consumption while keeping communication reliability and computational resources usage among boundaries.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Long-lasting insecticide-treated nets (LLITNs) constitute a novel alternative that combines physical and chemical tactics to prevent insect access and the spread of insect-transmitted plant viruses in protected enclosures. This approach is based on a slow-release insecticide-treated net with large hole sizes that allow improved ventilation of greenhouses. The efficacy of a wide range of LLITNs was tested under laboratory conditions against Myzus persicae, Aphis gossypii and Bemisia tabaci. Two nets were selected for field tests under a high insect infestation pressure in the presence of plants infected with Cucumber mosaic virus and Cucurbit aphid-borne yellows virus. The efficacy of Aphidius colemani, a parasitoid commonly used for biological control of aphids, was studied in parallel field experiments.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Esta tesis establece los fundamentos teóricos y diseña una colección abierta de clases C++ denominada VBF (Vector Boolean Functions) para analizar funciones booleanas vectoriales (funciones que asocian un vector booleano a otro vector booleano) desde una perspectiva criptográfica. Esta nueva implementación emplea la librería NTL de Victor Shoup, incorporando nuevos módulos que complementan a las funciones de NTL, adecuándolas para el análisis criptográfico. La clase fundamental que representa una función booleana vectorial se puede inicializar de manera muy flexible mediante diferentes estructuras de datas tales como la Tabla de verdad, la Representación de traza y la Forma algebraica normal entre otras. De esta manera VBF permite evaluar los criterios criptográficos más relevantes de los algoritmos de cifra en bloque y de stream, así como funciones hash: por ejemplo, proporciona la no-linealidad, la distancia lineal, el grado algebraico, las estructuras lineales, la distribución de frecuencias de los valores absolutos del espectro Walsh o del espectro de autocorrelación, entre otros criterios. Adicionalmente, VBF puede llevar a cabo operaciones entre funciones booleanas vectoriales tales como la comprobación de igualdad, la composición, la inversión, la suma, la suma directa, el bricklayering (aplicación paralela de funciones booleanas vectoriales como la empleada en el algoritmo de cifra Rijndael), y la adición de funciones coordenada. La tesis también muestra el empleo de la librería VBF en dos aplicaciones prácticas. Por un lado, se han analizado las características más relevantes de los sistemas de cifra en bloque. Por otro lado, combinando VBF con algoritmos de optimización, se han diseñado funciones booleanas cuyas propiedades criptográficas son las mejores conocidas hasta la fecha. ABSTRACT This thesis develops the theoretical foundations and designs an open collection of C++ classes, called VBF, designed for analyzing vector Boolean functions (functions that map a Boolean vector to another Boolean vector) from a cryptographic perspective. This new implementation uses the NTL library from Victor Shoup, adding new modules which complement the existing ones making VBF better suited for cryptography. The fundamental class representing a vector Boolean function can be initialized in a flexible way via several alternative types of data structures such as Truth Table, Trace Representation, Algebraic Normal Form (ANF) among others. This way, VBF allows the evaluation of the most relevant cryptographic criteria for block and stream ciphers as well as for hash functions: for instance, it provides the nonlinearity, the linearity distance, the algebraic degree, the linear structures, the frequency distribution of the absolute values of the Walsh Spectrum or the Autocorrelation Spectrum, among others. In addition, VBF can perform operations such as equality testing, composition, inversion, sum, direct sum, bricklayering (parallel application of vector Boolean functions as employed in Rijndael cipher), and adding coordinate functions of two vector Boolean functions. This thesis also illustrates the use of VBF in two practical applications. On the one hand, the most relevant properties of the existing block ciphers have been analysed. On the other hand, by combining VBF with optimization algorithms, new Boolean functions have been designed which have the best known cryptographic properties up-to-date.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La Diabetes mellitus es una enfermedad caracterizada por la insuficiente o nula producción de insulina por parte del páncreas o la reducida sensibilidad del organismo a esta hormona, que ayuda a que la glucosa llegue a los tejidos y al sistema nervioso para suministrar energía. La Diabetes tiene una mayor prevalencia en los países desarrollados debido a múltiples factores, entre ellos la obesidad, la vida sedentaria, y disfunciones en el sistema endocrino relacionadas con el páncreas. La Diabetes Tipo 1 es una enfermedad crónica e incurable, en la que son destruidas las células beta del páncreas, que producen la insulina, haciéndose necesaria la administración de insulina de forma exógena para controlar los niveles de glucosa en sangre. El paciente debe seguir una terapia con insulina administrada por vía subcutánea, que debe estar adaptada a sus necesidades metabólicas y a sus hábitos de vida. Esta terapia intenta imitar el perfil insulínico de un páncreas sano. La tecnología actual permite abordar el desarrollo del denominado “páncreas endocrino artificial” (PEA), que aportaría precisión, eficacia y seguridad en la aplicación de las terapias con insulina y permitiría una mayor independencia de los pacientes frente a su enfermedad, que en la actualidad están sujetos a una constante toma de decisiones. El PEA consta de un sensor continuo de glucosa, una bomba de infusión de insulina y un algoritmo de control, que calcula la insulina a infusionar utilizando los niveles de glucosa del paciente como información principal. Este trabajo presenta una modificación en el método de control en lazo cerrado propuesto en un proyecto previo. El controlador del que se parte está compuesto por un controlador basal booleano y un controlador borroso postprandial basado en reglas borrosas heredadas del controlador basal. El controlador postprandial administra el 50% del bolo manual (calculado a partir de la cantidad de carbohidratos que el paciente va a consumir) en el instante del aviso de la ingesta y reparte el resto en instantes posteriores. El objetivo es conseguir una regulación óptima del nivel de glucosa en el periodo postprandial. Con el objetivo de reducir las hiperglucemias que se producen en el periodo postprandial se realiza un transporte de insulina, que es un adelanto de la insulina basal del periodo postprandial que se suministrará junto con un porcentaje variable del bolo manual. Este porcentaje estará relacionado con el estado metabólico del paciente previo a la ingesta. Además se modificará la base de conocimiento para adecuar el comportamiento del controlador al periodo postprandial. Este proyecto está enfocado en la mejora del controlador borroso postprandial previo, modificando dos aspectos: la inferencia del controlador postprandial y añadiendo una toma de decisiones automática sobre el % del bolo manual y el transporte. Se ha propuesto un controlador borroso con una nueva inferencia, que no hereda las características del controlado basal, y ha sido adaptado al periodo postprandial. Se ha añadido una inferencia borrosa que modifica la cantidad de insulina a administrar en el momento del aviso de ingesta y la cantidad de insulina basal a transportar del periodo postprandial al bolo manual. La validación del algoritmo se ha realizado mediante experimentos en simulación utilizando una población de diez pacientes sintéticos pertenecientes al Simulador de Padua/Virginia, evaluando los resultados con estadísticos para después compararlos con los obtenidos con el método de control anterior. Tras la evaluación de los resultados se puede concluir que el nuevo controlador postprandial, acompañado de la toma de decisiones automática, realiza un mejor control glucémico en el periodo postprandial, disminuyendo los niveles de las hiperglucemias. ABSTRACT. Diabetes mellitus is a disease characterized by the insufficient or null production of insulin from the pancreas or by a reduced sensitivity to this hormone, which helps glucose get to the tissues and the nervous system to provide energy. Diabetes has more prevalence in developed countries due to multiple factors, including obesity, sedentary lifestyle and endocrine dysfunctions related to the pancreas. Type 1 Diabetes is a chronic, incurable disease in which beta cells in the pancreas that produce insulin are destroyed, and exogenous insulin delivery is required to control blood glucose levels. The patient must follow a therapy with insulin administered by the subcutaneous route that should be adjusted to the metabolic needs and lifestyle of the patient. This therapy tries to imitate the insulin profile of a non-pathological pancreas. Current technology can adress the development of the so-called “endocrine artificial pancreas” (EAP) that would provide accuracy, efficacy and safety in the application of insulin therapies and will allow patients a higher level of independence from their disease. Patients are currently tied to constant decision making. The EAP consists of a continuous glucose sensor, an insulin infusion pump and a control algorithm that computes the insulin amount that has to be infused using the glucose as the main source of information. This work shows modifications to the control method in closed loop proposed in a previous project. The reference controller is composed by a boolean basal controller and a postprandial rule-based fuzzy controller which inherits the rules from the basal controller. The postprandial controller administrates 50% of the bolus (calculated from the amount of carbohydrates that the patient is going to ingest) in the moment of the intake warning, and distributes the remaining in later instants. The goal is to achieve an optimum regulation of the glucose level in the postprandial period. In order to reduce hyperglycemia in the postprandial period an insulin transport is carried out. It consists on a feedforward of the basal insulin from the postprandial period, which will be administered with a variable percentage of the manual bolus. This percentage would be linked with the metabolic state of the patient in moments previous to the intake. Furthermore, the knowledge base is going to be modified in order to fit the controller performance to the postprandial period. This project is focused on the improvement of the previous controller, modifying two aspects: the postprandial controller inference, and the automatic decision making on the percentage of the manual bolus and the transport. A fuzzy controller with a new inference has been proposed and has been adapted to the postprandial period. A fuzzy inference has been added, which modifies both the amount of manual bolus to administrate at the intake warning and the amount of basal insulin to transport to the prandial bolus. The algorithm assessment has been done through simulation experiments using a synthetic population of 10 patients in the UVA/PADOVA simulator, evaluating the results with statistical parameters for further comparison with those obtained with the previous control method. After comparing results it can be concluded that the new postprandial controller, combined with the automatic decision making, carries out a better glycemic control in the postprandial period, decreasing levels of hyperglycemia.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Stream-mining approach is defined as a set of cutting-edge techniques designed to process streams of data in real time, in order to extract knowledge. In the particular case of classification, stream-mining has to adapt its behaviour to the volatile underlying data distributions, what has been called concept drift. Moreover, it is important to note that concept drift may lead to situations where predictive models become invalid and have therefore to be updated to represent the actual concepts that data poses. In this context, there is a specific type of concept drift, known as recurrent concept drift, where the concepts represented by data have already appeared in the past. In those cases the learning process could be saved or at least minimized by applying a previously trained model. This could be extremely useful in ubiquitous environments that are characterized by the existence of resource constrained devices. To deal with the aforementioned scenario, meta-models can be used in the process of enhancing the drift detection mechanisms used by data stream algorithms, by representing and predicting when the change will occur. There are some real-world situations where a concept reappears, as in the case of intrusion detection systems (IDS), where the same incidents or an adaptation of them usually reappear over time. In these environments the early prediction of drift by means of a better knowledge of past models can help to anticipate to the change, thus improving efficiency of the model regarding the training instances needed. By means of using meta-models as a recurrent drift detection mechanism, the ability to share concepts representations among different data mining processes is open. That kind of exchanges could improve the accuracy of the resultant local model as such model may benefit from patterns similar to the local concept that were observed in other scenarios, but not yet locally. This would also improve the efficiency of training instances used during the classification process, as long as the exchange of models would aid in the application of already trained recurrent models, that have been previously seen by any of the collaborative devices. Which it is to say that the scope of recurrence detection and representation is broaden. In fact the detection, representation and exchange of concept drift patterns would be extremely useful for the law enforcement activities fighting against cyber crime. Being the information exchange one of the main pillars of cooperation, national units would benefit from the experience and knowledge gained by third parties. Moreover, in the specific scope of critical infrastructures protection it is crucial to count with information exchange mechanisms, both from a strategical and technical scope. The exchange of concept drift detection schemes in cyber security environments would aid in the process of preventing, detecting and effectively responding to threads in cyber space. Furthermore, as a complement of meta-models, a mechanism to assess the similarity between classification models is also needed when dealing with recurrent concepts. In this context, when reusing a previously trained model a rough comparison between concepts is usually made, applying boolean logic. The introduction of fuzzy logic comparisons between models could lead to a better efficient reuse of previously seen concepts, by applying not just equal models, but also similar ones. This work faces the aforementioned open issues by means of: the MMPRec system, that integrates a meta-model mechanism and a fuzzy similarity function; a collaborative environment to share meta-models between different devices; a recurrent drift generator that allows to test the usefulness of recurrent drift systems, as it is the case of MMPRec. Moreover, this thesis presents an experimental validation of the proposed contributions using synthetic and real datasets.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Las redes del futuro, incluyendo las redes de próxima generación, tienen entre sus objetivos de diseño el control sobre el consumo de energía y la conectividad de la red. Estos objetivos cobran especial relevancia cuando hablamos de redes con capacidades limitadas, como es el caso de las redes de sensores inalámbricos (WSN por sus siglas en inglés). Estas redes se caracterizan por estar formadas por dispositivos de baja o muy baja capacidad de proceso y por depender de baterías para su alimentación. Por tanto la optimización de la energía consumida se hace muy importante. Son muchas las propuestas que se han realizado para optimizar el consumo de energía en este tipo de redes. Quizás las más conocidas son las que se basan en la planificación coordinada de periodos de actividad e inactividad, siendo una de las formas más eficaces para extender el tiempo de vida de las baterías. La propuesta que se presenta en este trabajo se basa en el control de la conectividad mediante una aproximación probabilística. La idea subyacente es que se puede esperar que una red mantenga la conectividad si todos sus nodos tienen al menos un número determinado de vecinos. Empleando algún mecanismo que mantenga ese número, se espera que se pueda mantener la conectividad con un consumo energético menor que si se empleara una potencia de transmisión fija que garantizara una conectividad similar. Para que el mecanismo sea eficiente debe tener la menor huella posible en los dispositivos donde se vaya a emplear. Por eso se propone el uso de un sistema auto-adaptativo basado en control mediante lógica borrosa. En este trabajo se ha diseñado e implementado el sistema descrito, y se ha probado en un despliegue real confirmando que efectivamente existen configuraciones posibles que permiten mantener la conectividad ahorrando energía con respecto al uso de una potencia de transmisión fija. ABSTRACT. Among the design goals for future networks, including next generation networks, we can find the energy consumption and the connectivity. These two goals are of special relevance when dealing with constrained networks. That is the case of Wireless Sensors Networks (WSN). These networks consist of devices with low or very low processing capabilities. They also depend on batteries for their operation. Thus energy optimization becomes a very important issue. Several proposals have been made for optimizing the energy consumption in this kind of networks. Perhaps the best known are those based on the coordinated planning of active and sleep intervals. They are indeed one of the most effective ways to extend the lifetime of the batteries. The proposal presented in this work uses a probabilistic approach to control the connectivity of a network. The underlying idea is that it is highly probable that the network will have a good connectivity if all the nodes have a minimum number of neighbors. By using some mechanism to reach that number, we hope that we can preserve the connectivity with a lower energy consumption compared to the required one if a fixed transmission power is used to achieve a similar connectivity. The mechanism must have the smallest footprint possible on the devices being used in order to be efficient. Therefore a fuzzy control based self-adaptive system is proposed. This work includes the design and implementation of the described system. It also has been validated in a real scenario deployment. We have obtained results supporting that there exist configurations where it is possible to get a good connectivity saving energy when compared to the use of a fixed transmission power for a similar connectivity.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This work presents a systematic process for building a Fault Diagnoser (FD), based on Petri Nets (PNs) which has been applied to a small helicopter. This novel tool is able to detect both intermittent and permanent faults. The work carried out is discussed from theoretical and practical point of view. The procedure begins with a division of the whole system into subsystems, which are the devices that have to be modeled by using PN, considering both the normal and fault operations. Subsequently, the models are integrated into a global Petri Net diagnoser (PND) that is able to monitor a whole helicopter and show critical variables to the operator in order to determine the UAV health, preventing accidents in this manner. A Data Acquisition System (DAQ) has been designed for collecting data during the flights and feeding PN diagnoser with them. Several real flights (nominal or under failure) have been carried out to perform the diagnoser setup and verify its performance. A summary of the validation results obtained during real flight tests is also included. An extensive use of this tool will improve preventive maintenance protocols for UAVs (especially helicopters) and allow establishing recommendations in regulations