17 resultados para Data envelopment analysis

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

En estos tiempos de crisis se hace imperativo lograr un consumo de recursos públicos lo más racional posible. El transporte público urbano es un sector al que se dedican grandes inversiones y cuya prestación de servicios está fuertemente subvencionada. El incremento de la eficiencia técnica del sector puede ayudar a conseguir una mejor gestión de los fondos públicos. Un primer paso para que se produzca una mejora es el desarrollo de una metodología de evaluación de la eficiencia del sector. Existen diferentes métodos para la evaluación técnica de un conjunto de compañías pertenecientes a un sector, entre los que se encuentra el análisis envolvente de datos (Data Envelopment Analysis, DEA, por sus siglas en inglés). Este método permite establecer una frontera de eficiencia técnica relativa a un determinado grupo de compañías, en función de un número limitado de variables. Las variables deben cuantificar, por un lado, la prestación de servicios de las distintas compañías, y por el otro, los recursos consumidos en la producción de dichos servicios. En el presente artículo se estudian, mediante el método DEA, las variables más idóneas para la evaluación de la eficiencia técnica de los servicios de autobuses urbanos en España. Asimismo, se analiza el número de variables más adecuado para conformar los modelos con los que se obtienen las fronteras de eficiencia. Para ello, se utilizan indicadores de los servicios de autobús urbano de las principales ciudades de las áreas metropolitanas españolas, para el periodo 2004-2009. Se emplea la base de datos del Observatorio de la Movilidad Metropolitana.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract interpretation-based data-flow analysis of logic programs is at this point relatively well understood from the point of view of general frameworks and abstract domains. On the other hand, comparatively little attention has been given to the problems which arise when analysis of a full, practical dialect of the Prolog language is attempted, and only few solutions to these problems have been proposed to date. Such problems relate to dealing correctly with all builtins, including meta-logical and extra-logical predicates, with dynamic predicates (where the program is modified during execution), and with the absence of certain program text during compilation. Existing proposals for dealing with such issues generally restrict in one way or another the classes of programs which can be analyzed if the information from analysis is to be used for program optimization. This paper attempts to fill this gap by considering a full dialect of Prolog, essentially following the recently proposed ISO standard, pointing out the problems that may arise in the analysis of such a dialect, and proposing a combination of known and novel solutions that together allow the correct analysis of arbitrary programs using the full power of the language.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

During the first decade of the new millennium, fueled by the economic development in Spain, urban bus services were extended. Since the years 2008 and 2009, the root of the economic crisis, the improvement of these services is at risk due to economic problems. In this paper, the technical efficiency of the main urban bus companies in Spain during the 2004–2009 period are studied using SBM (slack-based measures) models and by establishing the slacks in the services' production inputs. The influence of a series of exogenous variables on the operation of the different services is also analyzed. It is concluded that only the 24% of the case studies are efficient, and some urban form variables can explain part of the inefficiency. The methodology used allows studying the inefficiency in a disaggregated way that other DEA (data envelopment analysis) models do not.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper analyses the relationship between productive efficiency and online-social-networks (OSN) in Spanish telecommunications firms. A data-envelopment-analysis (DEA) is used and several indicators of business ?social Media? activities are incorporated. A super-efficiency analysis and bootstrapping techniques are performed to increase the model?s robustness and accuracy. Then, a logistic regression model is applied to characterise factors and drivers of good performance in OSN. Results reveal the company?s ability to absorb and utilise OSNs as a key factor in improving the productive efficiency. This paper presents a model for assessing the strategic performance of the presence and activity in OSN.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Sentiment and Emotion Analysis strongly depend on quality language resources, especially sentiment dictionaries. These resources are usually scattered, heterogeneous and limited to specific domains of appli- cation by simple algorithms. The EUROSENTIMENT project addresses these issues by 1) developing a common language resource representation model for sentiment analysis, and APIs for sentiment analysis services based on established Linked Data formats (lemon, Marl, NIF and ONYX) 2) by creating a Language Resource Pool (a.k.a. LRP) that makes avail- able to the community existing scattered language resources and services for sentiment analysis in an interoperable way. In this paper we describe the available language resources and services in the LRP and some sam- ple applications that can be developed on top of the EUROSENTIMENT LRP.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

En estos tiempos de crisis se hace imperativo lograr un consumo de recursos públicos lo más racional posible. El transporte público urbano es un sector al que se dedican grandes inversiones y cuya prestación de servicios está fuertemente subvencionada. El incremento de la eficiencia técnica del sector, entendida como la relación entre producción de servicios y consumo de recursos, puede ayudar a conseguir una mejor gestión de los fondos públicos. Un primer paso para que se produzca una mejora es el desarrollo de una metodología de evaluación de la eficiencia técnica de las compañías de transporte público. Existen diferentes métodos para la evaluación técnica de un conjunto de compañías pertenecientes a un sector. Uno de los más utilizados es el método frontera, en el que se encuentra el análisis envolvente de datos (Data Envelopment Analysis, DEA, por sus siglas en inglés). Este método permite establecer una frontera de eficiencia técnica relativa a un determinado grupo de compañías, en función de un número limitado de variables. Las variables deben cuantificar, por un lado, la prestación de servicios de las distintas compañías (outputs), y por el otro, los recursos consumidos en la producción de dichos servicios (inputs). El objetivo de esta tesis es analizar, mediante el uso del método DEA, la eficiencia técnica de los servicios de autobuses urbanos en España. Para ello, se estudia el número de variables más adecuado para conformar los modelos con los que se obtienen las fronteras de eficiencia. En el desarrollo de la metodología se utilizan indicadores de los servicios de autobús urbano de las principales ciudades de las áreas metropolitanas españolas, para el periodo 2004-2009. In times of crisis it is imperative achieve a consumption of public resources as rational as possible. Urban public transport is a sector devoted to large investments and whose services are heavily subsidized. Increase the technical efficiency of the sector, defined as the ratio of service output and resource consumption, can help achieve a better management of public funds. One step to produce an improvement is the development of a methodology for evaluating the technical efficiency of the public transport companies. There are different methods for the technical evaluation of a set of companies within an industry. One of the most widely used methods is the frontier method, in particular the Data Envelopment Analysis (DEA). This method allows the calculation of a technical efficiency frontier on a specific group of companies, based on a limited number of variables. Variables must quantify, on the one hand, the provision of services of different companies (outputs), and on the other hand, the resources consumed in the production of such services (inputs). The objective of this thesis is to analyze, using the DEA method, the technical efficiency of urban bus services in Spain. For this purpose, it is studied the more suitable variables that can be used in the models to obtain the efficiency frontiers. In developing the methodology are used indicators of urban bus services in major cities of the Spanish metropolitan areas for the period 2004-2009.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Este artículo analiza la relación entre la eficiencia productiva y las Redes Sociales Virtuales (RSV) en las empresas de telecomunicaciones en España. En una primera etapa, se aplica el análisis envolvente de datos (DEA) incorporando varios indicadores de actividad ?Social Media?. En una segunda etapa, se utiliza una regresión logística para caracterizar las empresas eficientes. Los resultados muestran que la capacidad de absorción y utilización de las RSV es un factor determinante en la mejora de la eficiencia productiva. La utilización combinada y las distintas capacidades de gestión de las RSV permiten identificarlas como un recurso heterogéneo. Este trabajo presenta un modelo para la evaluación del desempeño estratégico al abordar su presencia y actividad en RSV. ABSTRACT. This paper analyzes the relationship between the productive efficiency and the Online Social Networks - OSN in the Spanish telecommunications firms. First, a data envelopment analysis (DEA) is used and several indicators of business "Social Media" activities are incorporated. In a second stage, a logistic regression model regression is applied to characteri ze the efficient enterprises. Results show that the company's ability to absorb and utilize this OSN is a key factor in improving the productive efficiency. These results on the combined use and different management capabilities of OSN point to a definitio n of OSN as a heterogeneous resource. This paper presents a model for assessing the strategic performance to address their presence and activity in OSN.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Abstract interpretation-based data-flow analysis of logic programs is, at this point, relatively well understood from the point of view of general frameworks and abstract domains. On the other hand, comparatively little attention has been given to the problems which arise when analysis of a full, practical dialect of the Prolog language is attempted, and only few solutions to these problems have been proposed to date. Existing proposals generally restrict in one way or another the classes of programs which can be analyzed. This paper attempts to fill this gap by considering a full dialect of Prolog, essentially the recent ISO standard, pointing out the problems that may arise in the analysis of such a dialect, and proposing a combination of known and novel solutions that together allow the correct analysis of arbitrary programs which use the full power of the language.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Global data-flow analysis of (constraint) logic programs, which is generally based on abstract interpretation [7], is reaching a comparatively high level of maturity. A natural question is whether it is time for its routine incorporation in standard compilers, something which, beyond a few experimental systems, has not happened to date. Such incorporation arguably makes good sense only if: • the range of applications of global analysis is large enough to justify the additional complication in the compiler, and • global analysis technology can deal with all the features of "practical" languages (e.g., the ISO-Prolog built-ins) and "scales up" for large programs. We present a tutorial overview of a number of concepts and techniques directly related to the issues above, with special emphasis on the first one. In particular, we concéntrate on novel uses of global analysis during program development and debugging, rather than on the more traditional application área of program optimization. The idea of using abstract interpretation for validation and diagnosis has been studied in the context of imperative programming [2] and also of logic programming. The latter work includes issues such as using approximations to reduce the burden posed on programmers by declarative debuggers [6, 3] and automatically generating and checking assertions [4, 5] (which includes the more traditional type checking of strongly typed languages, such as Gódel or Mercury [1, 8, 9]) We also review some solutions for scalability including modular analysis, incremental analysis, and widening. Finally, we discuss solutions for dealing with meta-predicates, side-effects, delay declarations, constraints, dynamic predicates, and other such features which may appear in practical languages. In the discussion we will draw both from the literature and from our experience and that of others in the development and use of the CIAO system analyzer. In order to emphasize the practical aspects of the solutions discussed, the presentation of several concepts will be illustrated by examples run on the CIAO system, which makes extensive use of global analysis and assertions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We report on a detailed study of the application and effectiveness of program analysis based on abstract interpretation to automatic program parallelization. We study the case of parallelizing logic programs using the notion of strict independence. We first propose and prove correct a methodology for the application in the parallelization task of the information inferred by abstract interpretation, using a parametric domain. The methodology is generic in the sense of allowing the use of different analysis domains. A number of well-known approximation domains are then studied and the transformation into the parametric domain defined. The transformation directly illustrates the relevance and applicability of each abstract domain for the application. Both local and global analyzers are then built using these domains and embedded in a complete parallelizing compiler. Then, the performance of the domains in this context is assessed through a number of experiments. A comparatively wide range of aspects is studied, from the resources needed by the analyzers in terms of time and memory to the actual benefits obtained from the information inferred. Such benefits are evaluated both in terms of the characteristics of the parallelized code and of the actual speedups obtained from it. The results show that data flow analysis plays an important role in achieving efficient parallelizations, and that the cost of such analysis can be reasonable even for quite sophisticated abstract domains. Furthermore, the results also offer significant insight into the characteristics of the domains, the demands of the application, and the trade-offs involved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Esta tesis doctoral se enmarca dentro de la computación con membranas. Se trata de un tipo de computación bio-inspirado, concretamente basado en las células de los organismos vivos, en las que se producen múltiples reacciones de forma simultánea. A partir de la estructura y funcionamiento de las células se han definido diferentes modelos formales, denominados P sistemas. Estos modelos no tratan de modelar el comportamiento biológico de una célula, sino que abstraen sus principios básicos con objeto de encontrar nuevos paradigmas computacionales. Los P sistemas son modelos de computación no deterministas y masivamente paralelos. De ahí el interés que en los últimos años estos modelos han suscitado para la resolución de problemas complejos. En muchos casos, consiguen resolver de forma teórica problemas NP-completos en tiempo polinómico o lineal. Por otra parte, cabe destacar también la aplicación que la computación con membranas ha tenido en la investigación de otros muchos campos, sobre todo relacionados con la biología. Actualmente, una gran cantidad de estos modelos de computación han sido estudiados desde el punto de vista teórico. Sin embargo, el modo en que pueden ser implementados es un reto de investigación todavía abierto. Existen varias líneas en este sentido, basadas en arquitecturas distribuidas o en hardware dedicado, que pretenden acercarse en lo posible a su carácter no determinista y masivamente paralelo, dentro de un contexto de viabilidad y eficiencia. En esta tesis doctoral se propone la realización de un análisis estático del P sistema, como vía para optimizar la ejecución del mismo en estas plataformas. Se pretende que la información recogida en tiempo de análisis sirva para configurar adecuadamente la plataforma donde se vaya a ejecutar posteriormente el P sistema, obteniendo como consecuencia una mejora en el rendimiento. Concretamente, en esta tesis se han tomado como referencia los P sistemas de transiciones para llevar a cabo el estudio de dicho análisis estático. De manera un poco más específica, el análisis estático propuesto en esta tesis persigue que cada membrana sea capaz de determinar sus reglas activas de forma eficiente en cada paso de evolución, es decir, aquellas reglas que reúnen las condiciones adecuadas para poder ser aplicadas. En esta línea, se afronta el problema de los estados de utilidad de una membrana dada, que en tiempo de ejecución permitirán a la misma conocer en todo momento las membranas con las que puede comunicarse, cuestión que determina las reglas que pueden aplicarse en cada momento. Además, el análisis estático propuesto en esta tesis se basa en otra serie de características del P sistema como la estructura de membranas, antecedentes de las reglas, consecuentes de las reglas o prioridades. Una vez obtenida toda esta información en tiempo de análisis, se estructura en forma de árbol de decisión, con objeto de que en tiempo de ejecución la membrana obtenga las reglas activas de la forma más eficiente posible. Por otra parte, en esta tesis se lleva a cabo un recorrido por un número importante de arquitecturas hardware y software que diferentes autores han propuesto para implementar P sistemas. Fundamentalmente, arquitecturas distribuidas, hardware dedicado basado en tarjetas FPGA y plataformas basadas en microcontroladores PIC. El objetivo es proponer soluciones que permitan implantar en dichas arquitecturas los resultados obtenidos del análisis estático (estados de utilidad y árboles de decisión para reglas activas). En líneas generales, se obtienen conclusiones positivas, en el sentido de que dichas optimizaciones se integran adecuadamente en las arquitecturas sin penalizaciones significativas. Summary Membrane computing is the focus of this doctoral thesis. It can be considered a bio-inspired computing type. Specifically, it is based on living cells, in which many reactions take place simultaneously. From cell structure and operation, many different formal models have been defined, named P systems. These models do not try to model the biological behavior of the cell, but they abstract the basic principles of the cell in order to find out new computational paradigms. P systems are non-deterministic and massively parallel computational models. This is why, they have aroused interest when dealing with complex problems nowadays. In many cases, they manage to solve in theory NP problems in polynomial or lineal time. On the other hand, it is important to note that membrane computing has been successfully applied in many researching areas, specially related to biology. Nowadays, lots of these computing models have been sufficiently characterized from a theoretical point of view. However, the way in which they can be implemented is a research challenge, that it is still open nowadays. There are some lines in this way, based on distributed architectures or dedicated hardware. All of them are trying to approach to its non-deterministic and parallel character as much as possible, taking into account viability and efficiency. In this doctoral thesis it is proposed carrying out a static analysis of the P system in order to optimize its performance in a computing platform. The general idea is that after data are collected in analysis time, they are used for getting a suitable configuration of the computing platform in which P system is going to be performed. As a consequence, the system throughput will improve. Specifically, this thesis has made use of Transition P systems for carrying out the study in static analysis. In particular, the static analysis proposed in this doctoral thesis tries to achieve that every membrane can efficiently determine its active rules in every evolution step. These rules are the ones that can be applied depending on the system configuration at each computational step. In this line, we are going to tackle the problem of the usefulness states for a membrane. This state will allow this membrane to know the set of membranes with which communication is possible at any time. This is a very important issue in determining the set of rules that can be applied. Moreover, static analysis in this thesis is carried out taking into account other properties such as membrane structure, rule antecedents, rule consequents and priorities among rules. After collecting all data in analysis time, they are arranged in a decision tree structure, enabling membranes to obtain the set of active rules as efficiently as possible in run-time system. On the other hand, in this doctoral thesis is going to carry out an overview of hardware and software architectures, proposed by different authors in order to implement P systems, such as distributed architectures, dedicated hardware based on PFGA, and computing platforms based on PIC microcontrollers. The aim of this overview is to propose solutions for implementing the results of the static analysis, that is, usefulness states and decision trees for active rules. In general, conclusions are satisfactory, because these optimizations can be properly integrated in most of the architectures without significant penalties.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

El objetivo de este trabajo es la aportación de un caso que evidencie la importancia de realizar esfuerzos tanto en el terreno económico como en el laboral para conseguir una adecuada retención y atracción del conocimiento científico. Para ello, se analizan los datos de un estudio que destaca estas cuestiones para un país como España, que necesita urgentemente mejorar su balance de movilidad investigadora. Los resultados confirman la importancia de la aplicación de estas medidas y las debilidades que presenta en este tipo de políticas el sistema nacional de investigación, desarrollo e innovación (I+D+i). Abstract: The goal of this paper is to disclose a case study which highlights the impor tance of realising effor ts at an economic and labour scale in order to reach an adequate retention and attraction of scientific knowledge. Although science has evolved during different periods through the flow of mankind and ideas, studies about international mobility of scientists are recent in time. It started mid XXth century with the brain drain phenomenon. To alleviate the loss of scientists, countries have disclosed different politics. Some countries like Singapore, Southern Korea, India and China have been successful in bringing scientists back home. Those countries have made huge effor ts in human resources investments and scientific infrastructures, so as to enforce their national R&D&I systems. Despite such experiences and despite the increasement in terms of evidences related with international scientific mobility, few studies and figures have been disclosed. The goal of this current work is to disclose data, which backs up the liability of such kind of politics that combine effor ts in economic and labour terms. For that reason, data has been analyzed from a study which highlights these key aspects for a country like Spain and which recently established a wide range of politics in order to attract scientists. Despite the increasement of financial resources, such politics have turned out unable to reach a proper balance in terms of mobility research. Such data as well as it analysis comes from the development of an annual survey to Spanish scientists during the period 2008- 2011. Based on sample data and analysis an expectations index for the coming year has been constructed. This compares the level of confidence which related the support and prestige of their national R&D&I systems in which they operate. Two groups where surveyed: young researchers which currently work in Spain and as a group are more likely to go working abroad and Spanish scientists who are currently working abroad. Samples were obtained for young researchers were the following: 2008 (217), 2009 (270), 2010 (390) and 2011 (610). In the case of Spanish scientists abroad, the following data were obtained: 2008 (218), 2009 (250), 2010 (241) and 2011 (167). Both groups assume simple random sampling, with a level of confidence of 95%. The results obtained confirm the importance of this combined effor t of labor and economic policies which presents weaknesses of Spanish national R&D&I systems and its application, so that Spain becomes a pole of excellence in terms of attraction and retention of scientific knowledge.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Tradicionalmente, el uso de técnicas de análisis de datos ha sido una de las principales vías para el descubrimiento de conocimiento oculto en grandes cantidades de datos, recopilados por expertos en diferentes dominios. Por otra parte, las técnicas de visualización también se han usado para mejorar y facilitar este proceso. Sin embargo, existen limitaciones serias en la obtención de conocimiento, ya que suele ser un proceso lento, tedioso y en muchas ocasiones infructífero, debido a la dificultad de las personas para comprender conjuntos de datos de grandes dimensiones. Otro gran inconveniente, pocas veces tenido en cuenta por los expertos que analizan grandes conjuntos de datos, es la degradación involuntaria a la que someten a los datos durante las tareas de análisis, previas a la obtención final de conclusiones. Por degradación quiere decirse que los datos pueden perder sus propiedades originales, y suele producirse por una reducción inapropiada de los datos, alterando así su naturaleza original y llevando en muchos casos a interpretaciones y conclusiones erróneas que podrían tener serias implicaciones. Además, este hecho adquiere una importancia trascendental cuando los datos pertenecen al dominio médico o biológico, y la vida de diferentes personas depende de esta toma final de decisiones, en algunas ocasiones llevada a cabo de forma inapropiada. Ésta es la motivación de la presente tesis, la cual propone un nuevo framework visual, llamado MedVir, que combina la potencia de técnicas avanzadas de visualización y minería de datos para tratar de dar solución a estos grandes inconvenientes existentes en el proceso de descubrimiento de información válida. El objetivo principal es hacer más fácil, comprensible, intuitivo y rápido el proceso de adquisición de conocimiento al que se enfrentan los expertos cuando trabajan con grandes conjuntos de datos en diferentes dominios. Para ello, en primer lugar, se lleva a cabo una fuerte disminución en el tamaño de los datos con el objetivo de facilitar al experto su manejo, y a la vez preservando intactas, en la medida de lo posible, sus propiedades originales. Después, se hace uso de efectivas técnicas de visualización para representar los datos obtenidos, permitiendo al experto interactuar de forma sencilla e intuitiva con los datos, llevar a cabo diferentes tareas de análisis de datos y así estimular visualmente su capacidad de comprensión. De este modo, el objetivo subyacente se basa en abstraer al experto, en la medida de lo posible, de la complejidad de sus datos originales para presentarle una versión más comprensible, que facilite y acelere la tarea final de descubrimiento de conocimiento. MedVir se ha aplicado satisfactoriamente, entre otros, al campo de la magnetoencefalografía (MEG), que consiste en la predicción en la rehabilitación de lesiones cerebrales traumáticas (Traumatic Brain Injury (TBI) rehabilitation prediction). Los resultados obtenidos demuestran la efectividad del framework a la hora de acelerar y facilitar el proceso de descubrimiento de conocimiento sobre conjuntos de datos reales. ABSTRACT Traditionally, the use of data analysis techniques has been one of the main ways of discovering knowledge hidden in large amounts of data, collected by experts in different domains. Moreover, visualization techniques have also been used to enhance and facilitate this process. However, there are serious limitations in the process of knowledge acquisition, as it is often a slow, tedious and many times fruitless process, due to the difficulty for human beings to understand large datasets. Another major drawback, rarely considered by experts that analyze large datasets, is the involuntary degradation to which they subject the data during analysis tasks, prior to obtaining the final conclusions. Degradation means that data can lose part of their original properties, and it is usually caused by improper data reduction, thereby altering their original nature and often leading to erroneous interpretations and conclusions that could have serious implications. Furthermore, this fact gains a trascendental importance when the data belong to medical or biological domain, and the lives of people depends on the final decision-making, which is sometimes conducted improperly. This is the motivation of this thesis, which proposes a new visual framework, called MedVir, which combines the power of advanced visualization techniques and data mining to try to solve these major problems existing in the process of discovery of valid information. Thus, the main objective is to facilitate and to make more understandable, intuitive and fast the process of knowledge acquisition that experts face when working with large datasets in different domains. To achieve this, first, a strong reduction in the size of the data is carried out in order to make the management of the data easier to the expert, while preserving intact, as far as possible, the original properties of the data. Then, effective visualization techniques are used to represent the obtained data, allowing the expert to interact easily and intuitively with the data, to carry out different data analysis tasks, and so visually stimulating their comprehension capacity. Therefore, the underlying objective is based on abstracting the expert, as far as possible, from the complexity of the original data to present him a more understandable version, thus facilitating and accelerating the task of knowledge discovery. MedVir has been succesfully applied to, among others, the field of magnetoencephalography (MEG), which consists in predicting the rehabilitation of Traumatic Brain Injury (TBI). The results obtained successfully demonstrate the effectiveness of the framework to accelerate and facilitate the process of knowledge discovery on real world datasets.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

La gestión del conocimiento (KM) es el proceso de recolectar datos en bruto para su análisis y filtrado, con la finalidad de obtener conocimiento útil a partir de dichos datos. En este proyecto se pretende hacer un estudio sobre la gestión de la información en las redes de sensores inalámbricos como inicio para sentar las bases para la gestión del conocimiento en las mismas. Las redes de sensores inalámbricos (WSN) son redes compuestas por sensores (también conocidos como motas) distribuidos sobre un área, cuya misión es monitorizar una o varias condiciones físicas del entorno. Las redes de sensores inalámbricos se caracterizan por tener restricciones de consumo para los sensores que utilizan baterías, por su capacidad para adaptarse a cambios y ser escalables, y también por su habilidad para hacer frente a fallos en los sensores. En este proyecto se hace un estudio sobre la gestión de la información en redes de sensores inalámbricos. Se comienza introduciendo algunos conceptos básicos: arquitectura, pila de protocolos, topologías de red, etc.… Después de esto, se ha enfocado el estudio hacia TinyDB, el cual puede ser considerado como parte de las tecnologías más avanzadas en el estado del arte de la gestión de la información en redes de sensores inalámbricos. TinyDB es un sistema de procesamiento de consultas para extraer información de una red de sensores. Proporciona una interfaz similar a SQL y permite trabajar con consultas contra la red de sensores inalámbricos como si se tratara de una base de datos tradicional. Además, TinyDB implementa varias optimizaciones para manejar los datos eficientemente. En este proyecto se describe también la implementación de una sencilla aplicación basada en redes de sensores inalámbricos. Las motas en la aplicación son capaces de medir la corriente a través de un cable. El objetivo de esta aplicación es monitorizar el consumo de energía en diferentes zonas de un área industrial o doméstico, utilizando redes de sensores inalámbricas. Además, se han implementado las optimizaciones más importantes que se han aprendido en el análisis de la plataforma TinyDB. Para desarrollar esta aplicación se ha utilizado como sensores la plataforma open-source de creación de prototipos electrónicos Arduino, y el ordenador de placa reducida Raspberry Pi como coordinador. ABSTRACT. Knowledge management (KM) is the process of collecting raw data for analysis and filtering, to get a useful knowledge from this data. In this project the information management in wireless sensor networks is studied as starting point before knowledge management. Wireless sensor networks (WSN) are networks which consists of sensors (also known as motes) distributed over an area, to monitor some physical conditions of the environment. Wireless sensor networks are characterized by power consumption constrains for sensors which are using batteries, by the ability to be adaptable to changes and to be scalable, and by the ability to cope sensor failures. In this project it is studied information management in wireless sensor networks. The document starts introducing basic concepts: architecture, stack of protocols, network topology… After this, the study has been focused on TinyDB, which can be considered as part of the most advanced technologies in the state of the art of information management in wireless sensor networks. TinyDB is a query processing system for extracting information from a network of sensors. It provides a SQL-like interface and it lets us to work with queries against the wireless sensor network like if it was a traditional database. In addition, TinyDB implements a lot of optimizations to manage data efficiently. In this project, it is implemented a simple wireless sensor network application too. Application’s motes are able to measure amperage through a cable. The target of the application is, by using a wireless sensor network and these sensors, to monitor energy consumption in different areas of a house. Additionally, it is implemented the most important optimizations that we have learned from the analysis of TinyDB platform. To develop this application it is used Arduino open-source electronics prototyping platform as motes, and Raspberry Pi single-board computer as coordinator.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

En España, los datos catastrales de acceso público, obtenidos a través de la Sede Electrónica del Catastro (SEC), no incluyen información sobre la genealogía o las relaciones de linaje existente entre las parcelas, de forma que la gestión de la información histórica es muy limitada. Este artículo presenta un método para obtener las relaciones de linaje más frecuentes entre las parcelas (agregación y segregación) y propone un prototipo de estructura relacional para el almacenamiento y la gestión histórica de los datos catastrales de acceso público. El proceso de análisis para deducir el linaje se basa en superposiciones espacio-temporales junto con secuencias de sentencias SQL. El método proporciona un 70% de relaciones de agregación y segregación correctas; el resto presentan errores debidos, en general, a anomalías presentes en los propios datos catastrales. ABSTRACT Public cadastral data in Spain, available through the Electronic Office of Cadastre, do not include any information on the genealogy or lineage relationship among the parcels, so the management of historical information becomes very limit ed. This paper presents a method for obtaining the most common lineage relationships between plots (aggregation and segregation) and proposes a relational structure prototype for the storage and historic management of public cadastral data. The analysis process to deduce the lineage is based on spatio-temporal overlaps coupled with SQL statements sequences. This method provides a 70% of correct aggregation and segregation relationships; the wrong cases are mainly caused by errors existing in the cadastral data themselves.