918 resultados para Information Models
Resumo:
One of the most promising areas in which probabilistic graphical models have shown an incipient activity is the field of heuristic optimization and, in particular, in Estimation of Distribution Algorithms. Due to their inherent parallelism, different research lines have been studied trying to improve Estimation of Distribution Algorithms from the point of view of execution time and/or accuracy. Among these proposals, we focus on the so-called distributed or island-based models. This approach defines several islands (algorithms instances) running independently and exchanging information with a given frequency. The information sent by the islands can be either a set of individuals or a probabilistic model. This paper presents a comparative study for a distributed univariate Estimation of Distribution Algorithm and a multivariate version, paying special attention to the comparison of two alternative methods for exchanging information, over a wide set of parameters and problems ? the standard benchmark developed for the IEEE Workshop on Evolutionary Algorithms and other Metaheuristics for Continuous Optimization Problems of the ISDA 2009 Conference. Several analyses from different points of view have been conducted to analyze both the influence of the parameters and the relationships between them including a characterization of the configurations according to their behavior on the proposed benchmark.
Resumo:
In the last decade energy utility sector has undergone major changes in terms of liberalization, increased competition, efforts in improving energy efficiency, and in new technological solution such as smart meter and grid operations. There are new information technology solutions (e.g. Advanced Metering Infrastructure /AMI ) on the horizon that will not only introduce new technical and organizational concepts, but have a very strong potential to radically change modus operandi of utility companies. Coordinated, multi-utility programs can help accelerate the development and market success of new high-efficiency technologies. These programs provide opportunities for researchers to develop new high-efficiency equipment for manufacturers to sell this new equipment with utility help, for utilities to increase the amount of energy they save from incentive programs, and for consumers to benefit from lower utility bills and a cleaner environment (as energy is reduced, pollutants produced at power plants decline).
Resumo:
The REpresentational State Transfer (REST) architectural style describes the design principles that made the World Wide Web scalable and the same principles can be applied in enterprise context to do loosely coupled and scalable application integration. In recent years, RESTful services are gaining traction in the industry and are commonly used as a simpler alternative to SOAP Web Services. However, one of the main drawbacks of RESTful services is the lack of standard mechanisms to support advanced quality-ofservice requirements that are common to enterprises. Transaction processing is one of the essential features of enterprise information systems and several transaction models have been proposed in the past years to fulfill the gap of transaction processing in RESTful services. The goal of this paper is to analyze the state-of-the-art RESTful transaction models and identify the current challenges.
Resumo:
One of the main concerns when conducting a dam test is the acute determination of the hydrograph for a specific flood event. The use of 2D direct rainfall hydraulic mathematical models on a finite elements mesh, combined with the efficiency of vector calculus that provides CUDA (Compute Unified Device Architecture) technology, enables nowadays the simulation of complex hydrological models without the need for terrain subbasin and transit splitting (as in HEC-HMS). Both the Spanish PNOA (National Plan of Aereal Orthophotography) Digital Terrain Model GRID with a 5 x 5 m accuracy and the CORINE GIS Land Cover (Coordination of INformation of the Environment) that allows assessment of the ground roughness, provide enough data to easily build these kind of models
Resumo:
El trabajo realizado en la presente tesis doctoral se debe considerar parte del proyecto UPMSat-2, que se enmarca dentro del ámbito de la tecnología aeroespacial. El UPMSat-2 es un microsatélite (de bajo coste y pequeño tamaño) diseñado, construido, probado e integrado por la Universidad Politécnica de Madrid (España), para fines de demostración tecnológica y educación. El objetivo de la presente tesis doctoral es presentar nuevos modelos analíticos para estudiar la interdependencia energética entre los subsistemas de potencia y de control de actitud de un satélite. En primer lugar, se estudia la simulación del subsistema de potencia de un microsatélite, prestando especial atención a la simulación de la fuente de potencia, esto es, los paneles solares. En la tesis se presentan métodos sencillos pero precisos para simular la producción de energía de los paneles en condiciones ambientales variables a través de su circuito equivalente. Los métodos propuestos para el cálculo de los parámetros del circuito equivalente son explícitos (o al menos, con las variables desacopladas), no iterativos y directos; no se necesitan iteraciones o valores iniciales para calcular los parámetros. La precisión de este método se prueba y se compara con métodos similares de la literatura disponible, demostrando una precisión similar para mayor simplicidad. En segundo lugar, se presenta la simulación del subsistema de control de actitud de un microsatélite, prestando especial atención a la nueva ley de control propuesta. La tesis presenta un nuevo tipo de control magnético es aplicable a la órbita baja terrestre (LEO). La ley de control propuesta es capaz de ajustar la velocidad de rotación del satélite alrededor de su eje principal de inercia máximo o mínimo. Además, en el caso de órbitas de alta inclinación, la ley de control favorece la alineación del eje de rotación con la dirección normal al plano orbital. El algoritmo de control propuesto es simple, sólo se requieren magnetopares como actuadores; sólo se requieren magnetómetros como sensores; no hace falta estimar la velocidad angular; no incluye un modelo de campo magnético de la Tierra; no tiene por qué ser externamente activado con información sobre las características orbitales y permite el rearme automático después de un apagado total del subsistema de control de actitud. La viabilidad teórica de la citada ley de control se demuestra a través de análisis de Monte Carlo. Por último, en términos de producción de energía, se demuestra que la actitud propuesto (en eje principal perpendicular al plano de la órbita, y el satélite que gira alrededor de ella con una velocidad controlada) es muy adecuado para la misión UPMSat-2, ya que permite una área superior de los paneles apuntando hacia el sol cuando se compara con otras actitudes estudiadas. En comparación con el control de actitud anterior propuesto para el UPMSat-2 resulta en un incremento de 25% en la potencia disponible. Además, la actitud propuesto mostró mejoras significativas, en comparación con otros, en términos de control térmico, como la tasa de rotación angular por satélite puede seleccionarse para conseguir una homogeneización de la temperatura más alta que apunta satélite y la antena. ABSTRACT The work carried out in the present doctoral dissertation should be considered part of the UPMSat-2 project, falling within the scope of the aerospace technology. The UPMSat-2 is a microsatellite (low cost and small size) designed, constructed integrated and tested for educational and technology demonstration purposes at the Universidad Politécnica de Madrid (Spain). The aim of the present doctoral dissertation is to present new analytical models to study the energy interdependence between the power and the attitude control subsystems of a satellite. First, the simulation of the power subsystem of a microsatellite is studied, paying particular attention to the simulation of the power supply, i.e. the solar panels. Simple but accurate methods for simulate the power production under variable ambient conditions using its equivalent circuit are presented. The proposed methods for calculate the equivalent circuit parameters are explicit (or at least, with decoupled variables), non-iterative and straight forward; no iterations or initial values for the parameters are needed. The accuracy of this method is tested and compared with similar methods from the available literature demonstrating similar precision but higher simplicity. Second, the simulation of the control subsystem of a microsatellite is presented, paying particular attention to the new control law proposed. A new type of magnetic control applied to Low Earth Orbit (LEO) satellites has been presented. The proposed control law is able to set the satellite rotation speed around its maximum or minimum inertia principal axis. Besides, the proposed control law favors the alignment of this axis with the normal direction to the orbital plane for high inclination orbits. The proposed control algorithm is simples, only magnetorquers are required as actuators; only magnetometers are required as sensors; no estimation of the angular velocity is needed; it does not include an in-orbit Earth magnetic field model; it does not need to be externally activated with information about the orbital characteristics and it allows automatic reset after a total shutdown of attitude control subsystem. The theoretical viability of the control law is demonstrated through Monte Carlo analysis. Finally, in terms of power production, it is demonstrated that the proposed attitude (on principal axis perpendicular to the orbit plane, and the satellite rotating around it with a controlled rate) is quite suitable for the UPMSat-2 mission, as it allows a higher area of the panels pointing towards the sun when compared to other studied attitudes. Compared with the previous attitude control proposed for the UPMSat-2 it results in a 25% increment in available power. Besides, the proposed attitude showed significant improvements, when compared to others, in terms of thermal control, as the satellite angular rotation rate can be selected to achieve a higher temperature homogenization of the satellite and antenna pointing.
Resumo:
Stream-mining approach is defined as a set of cutting-edge techniques designed to process streams of data in real time, in order to extract knowledge. In the particular case of classification, stream-mining has to adapt its behaviour to the volatile underlying data distributions, what has been called concept drift. Moreover, it is important to note that concept drift may lead to situations where predictive models become invalid and have therefore to be updated to represent the actual concepts that data poses. In this context, there is a specific type of concept drift, known as recurrent concept drift, where the concepts represented by data have already appeared in the past. In those cases the learning process could be saved or at least minimized by applying a previously trained model. This could be extremely useful in ubiquitous environments that are characterized by the existence of resource constrained devices. To deal with the aforementioned scenario, meta-models can be used in the process of enhancing the drift detection mechanisms used by data stream algorithms, by representing and predicting when the change will occur. There are some real-world situations where a concept reappears, as in the case of intrusion detection systems (IDS), where the same incidents or an adaptation of them usually reappear over time. In these environments the early prediction of drift by means of a better knowledge of past models can help to anticipate to the change, thus improving efficiency of the model regarding the training instances needed. By means of using meta-models as a recurrent drift detection mechanism, the ability to share concepts representations among different data mining processes is open. That kind of exchanges could improve the accuracy of the resultant local model as such model may benefit from patterns similar to the local concept that were observed in other scenarios, but not yet locally. This would also improve the efficiency of training instances used during the classification process, as long as the exchange of models would aid in the application of already trained recurrent models, that have been previously seen by any of the collaborative devices. Which it is to say that the scope of recurrence detection and representation is broaden. In fact the detection, representation and exchange of concept drift patterns would be extremely useful for the law enforcement activities fighting against cyber crime. Being the information exchange one of the main pillars of cooperation, national units would benefit from the experience and knowledge gained by third parties. Moreover, in the specific scope of critical infrastructures protection it is crucial to count with information exchange mechanisms, both from a strategical and technical scope. The exchange of concept drift detection schemes in cyber security environments would aid in the process of preventing, detecting and effectively responding to threads in cyber space. Furthermore, as a complement of meta-models, a mechanism to assess the similarity between classification models is also needed when dealing with recurrent concepts. In this context, when reusing a previously trained model a rough comparison between concepts is usually made, applying boolean logic. The introduction of fuzzy logic comparisons between models could lead to a better efficient reuse of previously seen concepts, by applying not just equal models, but also similar ones. This work faces the aforementioned open issues by means of: the MMPRec system, that integrates a meta-model mechanism and a fuzzy similarity function; a collaborative environment to share meta-models between different devices; a recurrent drift generator that allows to test the usefulness of recurrent drift systems, as it is the case of MMPRec. Moreover, this thesis presents an experimental validation of the proposed contributions using synthetic and real datasets.
Resumo:
This paper presents an overview of depth averaged modelling of fast catastrophic landslides where coupling of solid skeleton and pore fluid (air and water) is important. The first goal is to show how Biot-Zienkiewicz models can be applied to develop depth integrated, coupled models. The second objective of the paper is to consider a link which can be established between rheological and constitutive models. Perzyna´s viscoplasticity can be considered a general framework within which rheological models such as Bingham and cohesive frictional fluids can be derived. Among the several alternative numerical models, we will focus here on SPH which has not been widely applied by engineers to model landslide propagation. We propose an improvement, based on combining Finite Difference meshes associated to SPH nodes to describe pore pressure evolution inside the landslide mass. We devote a Section to analyze the performance of the models, considering three sets of tests and examples which allows to assess the model performance and limitations: (i) Problems having an analytical solution, (ii) Small scale laboratory tests, and (iii) Real cases for which we have had access to reliable information
Resumo:
The Mouse Tumor Biology (MTB) Database serves as a curated, integrated resource for information about tumor genetics and pathology in genetically defined strains of mice (i.e., inbred, transgenic and targeted mutation strains). Sources of information for the database include the published scientific literature and direct data submissions by the scientific community. Researchers access MTB using Web-based query forms and can use the database to answer such questions as ‘What tumors have been reported in transgenic mice created on a C57BL/6J background?’, ‘What tumors in mice are associated with mutations in the Trp53 gene?’ and ‘What pathology images are available for tumors of the mammary gland regardless of genetic background?’. MTB has been available on the Web since 1998 from the Mouse Genome Informatics web site (http://www.informatics.jax.org). We have recently implemented a number of enhancements to MTB including new query options, redesigned query forms and results pages for pathology and genetic data, and the addition of an electronic data submission and annotation tool for pathology data.
Resumo:
We evaluate the use of Generalized Empirical Likelihood (GEL) estimators in portfolios efficiency tests for asset pricing models in the presence of conditional information. Estimators from GEL family presents some optimal statistical properties, such as robustness to misspecification and better properties in finite samples. Unlike GMM, the bias for GEL estimators do not increase as more moment conditions are included, which is expected in conditional efficiency analysis. We found some evidences that estimators from GEL class really performs differently in small samples, where efficiency tests using GEL generate lower estimates compared to tests using the standard approach with GMM. With Monte Carlo experiments we see that GEL has better performance when distortions are present in data, especially under heavy tails and Gaussian shocks.
Resumo:
The exponential growth of the subjective information in the framework of the Web 2.0 has led to the need to create Natural Language Processing tools able to analyse and process such data for multiple practical applications. They require training on specifically annotated corpora, whose level of detail must be fine enough to capture the phenomena involved. This paper presents EmotiBlog – a fine-grained annotation scheme for subjectivity. We show the manner in which it is built and demonstrate the benefits it brings to the systems using it for training, through the experiments we carried out on opinion mining and emotion detection. We employ corpora of different textual genres –a set of annotated reported speech extracted from news articles, the set of news titles annotated with polarity and emotion from the SemEval 2007 (Task 14) and ISEAR, a corpus of real-life self-expressed emotion. We also show how the model built from the EmotiBlog annotations can be enhanced with external resources. The results demonstrate that EmotiBlog, through its structure and annotation paradigm, offers high quality training data for systems dealing both with opinion mining, as well as emotion detection.
Resumo:
The retina is a very complex neural structure, which performs spatial, temporal, and chromatic processing on visual information and converts it into a compact ‘digital’ format composed of neural impulses. This paper presents a new compiler-based framework able to describe, simulate and validate custom retina models. The framework is compatible with the most usual neural recording and analysis tools, taking advantage of the interoperability with these kinds of applications. Furthermore it is possible to compile the code to generate accelerated versions of the visual processing models compatible with COTS microprocessors, FPGAs or GPUs. The whole system represents an ongoing work to design and develop a functional visual neuroprosthesis. Several case studies are described to assess the effectiveness and usefulness of the framework.
Resumo:
Two predictive models are developed in this article: the first is designed to predict people's attitudes to alcoholic drinks, while the second sets out to predict the use of alcohol in relation to selected individual values. University students (N = 1,500) were recruited through stratified sampling based on sex and academic discipline. The questionnaire used obtained information on participants' alcohol use, attitudes and personal values. The results show that the attitudes model correctly classifies 76.3% of cases. Likewise, the model for level of alcohol use correctly classifies 82% of cases. According to our results, we can conclude that there are a series of individual values that influence drinking and attitudes to alcohol use, which therefore provides us with a potentially powerful instrument for developing preventive intervention programs.
Resumo:
The modeling of complex dynamic systems depends on the solution of a differential equations system. Some problems appear because we do not know the mathematical expressions of the said equations. Enough numerical data of the system variables are known. The authors, think that it is very important to establish a code between the different languages to let them codify and decodify information. Coding permits us to reduce the study of some objects to others. Mathematical expressions are used to model certain variables of the system are complex, so it is convenient to define an alphabet code determining the correspondence between these equations and words in the alphabet. In this paper the authors begin with the introduction to the coding and decoding of complex structural systems modeling.
Resumo:
The mathematical models of the complex reality are texts belonging to a certain literature that is written in a semi-formal language, denominated L(MT) by the authors whose laws linguistic mathematics have been previously defined. This text possesses linguistic entropy that is the reflection of the physical entropy of the processes of real world that said text describes. Through the temperature of information defined by Mandelbrot, the authors begin a text-reality thermodynamic theory that drives to the existence of information attractors, or highly structured point, settling down a heterogeneity of the space text, the same one that of ontologic space, completing the well-known law of Saint Mathew, of the General Theory of Systems and formulated by Margalef saying: “To the one that has more he will be given, and to the one that doesn't have he will even be removed it little that it possesses.
Resumo:
La recherche d'informations s'intéresse, entre autres, à répondre à des questions comme: est-ce qu'un document est pertinent à une requête ? Est-ce que deux requêtes ou deux documents sont similaires ? Comment la similarité entre deux requêtes ou documents peut être utilisée pour améliorer l'estimation de la pertinence ? Pour donner réponse à ces questions, il est nécessaire d'associer chaque document et requête à des représentations interprétables par ordinateur. Une fois ces représentations estimées, la similarité peut correspondre, par exemple, à une distance ou une divergence qui opère dans l'espace de représentation. On admet généralement que la qualité d'une représentation a un impact direct sur l'erreur d'estimation par rapport à la vraie pertinence, jugée par un humain. Estimer de bonnes représentations des documents et des requêtes a longtemps été un problème central de la recherche d'informations. Le but de cette thèse est de proposer des nouvelles méthodes pour estimer les représentations des documents et des requêtes, la relation de pertinence entre eux et ainsi modestement avancer l'état de l'art du domaine. Nous présentons quatre articles publiés dans des conférences internationales et un article publié dans un forum d'évaluation. Les deux premiers articles concernent des méthodes qui créent l'espace de représentation selon une connaissance à priori sur les caractéristiques qui sont importantes pour la tâche à accomplir. Ceux-ci nous amènent à présenter un nouveau modèle de recherche d'informations qui diffère des modèles existants sur le plan théorique et de l'efficacité expérimentale. Les deux derniers articles marquent un changement fondamental dans l'approche de construction des représentations. Ils bénéficient notamment de l'intérêt de recherche dont les techniques d'apprentissage profond par réseaux de neurones, ou deep learning, ont fait récemment l'objet. Ces modèles d'apprentissage élicitent automatiquement les caractéristiques importantes pour la tâche demandée à partir d'une quantité importante de données. Nous nous intéressons à la modélisation des relations sémantiques entre documents et requêtes ainsi qu'entre deux ou plusieurs requêtes. Ces derniers articles marquent les premières applications de l'apprentissage de représentations par réseaux de neurones à la recherche d'informations. Les modèles proposés ont aussi produit une performance améliorée sur des collections de test standard. Nos travaux nous mènent à la conclusion générale suivante: la performance en recherche d'informations pourrait drastiquement être améliorée en se basant sur les approches d'apprentissage de représentations.