868 resultados para localized algorithms


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Analysis of big amount of data is a field with many years of research. It is centred in getting significant values, to make it easier to understand and interpret data. Being the analysis of interdependence between time series an important field of research, mainly as a result of advances in the characterization of dynamical systems from the signals they produce. In the medicine sphere, it is easy to find many researches that try to understand the brain behaviour, its operation mode and its internal connections. The human brain comprises approximately 1011 neurons, each of which makes about 103 synaptic connections. This huge number of connections between individual processing elements provides the fundamental substrate for neuronal ensembles to become transiently synchronized or functionally connected. A similar complex network configuration and dynamics can also be found at the macroscopic scales of systems neuroscience and brain imaging. The emergence of dynamically coupled cell assemblies represents the neurophysiological substrate for cognitive function such as perception, learning, thinking. Understanding the complex network organization of the brain on the basis of neuroimaging data represents one of the most impervious challenges for systems neuroscience. Brain connectivity is an elusive concept that refers to diferent interrelated aspects of brain organization: structural, functional connectivity (FC) and efective connectivity (EC). Structural connectivity refers to a network of physical connections linking sets of neurons, it is the anatomical structur of brain networks. However, FC refers to the statistical dependence between the signals stemming from two distinct units within a nervous system, while EC refers to the causal interactions between them. This research opens the door to try to resolve diseases related with the brain, like Parkinson’s disease, senile dementia, mild cognitive impairment, etc. One of the most important project associated with Alzheimer’s research and other diseases are enclosed in the European project called Blue Brain. The center for Biomedical Technology (CTB) of Universidad Politecnica de Madrid (UPM) forms part of the project. The CTB researches have developed a magnetoencephalography (MEG) data processing tool that allow to visualise and analyse data in an intuitive way. This tool receives the name of HERMES, and it is presented in this document. Analysis of big amount of data is a field with many years of research. It is centred in getting significant values, to make it easier to understand and interpret data. Being the analysis of interdependence between time series an important field of research, mainly as a result of advances in the characterization of dynamical systems from the signals they produce. In the medicine sphere, it is easy to find many researches that try to understand the brain behaviour, its operation mode and its internal connections. The human brain comprises approximately 1011 neurons, each of which makes about 103 synaptic connections. This huge number of connections between individual processing elements provides the fundamental substrate for neuronal ensembles to become transiently synchronized or functionally connected. A similar complex network configuration and dynamics can also be found at the macroscopic scales of systems neuroscience and brain imaging. The emergence of dynamically coupled cell assemblies represents the neurophysiological substrate for cognitive function such as perception, learning, thinking. Understanding the complex network organization of the brain on the basis of neuroimaging data represents one of the most impervious challenges for systems neuroscience. Brain connectivity is an elusive concept that refers to diferent interrelated aspects of brain organization: structural, functional connectivity (FC) and efective connectivity (EC). Structural connectivity refers to a network of physical connections linking sets of neurons, it is the anatomical structur of brain networks. However, FC refers to the statistical dependence between the signals stemming from two distinct units within a nervous system, while EC refers to the causal interactions between them. This research opens the door to try to resolve diseases related with the brain, like Parkinson’s disease, senile dementia, mild cognitive impairment, etc. One of the most important project associated with Alzheimer’s research and other diseases are enclosed in the European project called Blue Brain. The center for Biomedical Technology (CTB) of Universidad Politecnica de Madrid (UPM) forms part of the project. The CTB researches have developed a magnetoencephalography (MEG) data processing tool that allow to visualise and analyse data in an intuitive way. This tool receives the name of HERMES, and it is presented in this document.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The design of nuclear power plant has to follow a number of regulations aimed at limiting the risks inherent in this type of installation. The goal is to prevent and to limit the consequences of any possible incident that might threaten the public or the environment. To verify that the safety requirements are met a safety assessment process is followed. Safety analysis is as key component of a safety assessment, which incorporates both probabilistic and deterministic approaches. The deterministic approach attempts to ensure that the various situations, and in particular accidents, that are considered to be plausible, have been taken into account, and that the monitoring systems and engineered safety and safeguard systems will be capable of ensuring the safety goals. On the other hand, probabilistic safety analysis tries to demonstrate that the safety requirements are met for potential accidents both within and beyond the design basis, thus identifying vulnerabilities not necessarily accessible through deterministic safety analysis alone. Probabilistic safety assessment (PSA) methodology is widely used in the nuclear industry and is especially effective in comprehensive assessment of the measures needed to prevent accidents with small probability but severe consequences. Still, the trend towards a risk informed regulation (RIR) demanded a more extended use of risk assessment techniques with a significant need to further extend PSA’s scope and quality. Here is where the theory of stimulated dynamics (TSD) intervenes, as it is the mathematical foundation of the integrated safety assessment (ISA) methodology developed by the CSN(Consejo de Seguridad Nuclear) branch of Modelling and Simulation (MOSI). Such methodology attempts to extend classical PSA including accident dynamic analysis, an assessment of the damage associated to the transients and a computation of the damage frequency. The application of this ISA methodology requires a computational framework called SCAIS (Simulation Code System for Integrated Safety Assessment). SCAIS provides accident dynamic analysis support through simulation of nuclear accident sequences and operating procedures. Furthermore, it includes probabilistic quantification of fault trees and sequences; and integration and statistic treatment of risk metrics. SCAIS comprehensively implies an intensive use of code coupling techniques to join typical thermal hydraulic analysis, severe accident and probability calculation codes. The integration of accident simulation in the risk assessment process and thus requiring the use of complex nuclear plant models is what makes it so powerful, yet at the cost of an enormous increase in complexity. As the complexity of the process is primarily focused on such accident simulation codes, the question of whether it is possible to reduce the number of required simulation arises, which will be the focus of the present work. This document presents the work done on the investigation of more efficient techniques applied to the process of risk assessment inside the mentioned ISA methodology. Therefore such techniques will have the primary goal of decreasing the number of simulation needed for an adequate estimation of the damage probability. As the methodology and tools are relatively recent, there is not much work done inside this line of investigation, making it a quite difficult but necessary task, and because of time limitations the scope of the work had to be reduced. Therefore, some assumptions were made to work in simplified scenarios best suited for an initial approximation to the problem. The following section tries to explain in detail the process followed to design and test the developed techniques. Then, the next section introduces the general concepts and formulae of the TSD theory which are at the core of the risk assessment process. Afterwards a description of the simulation framework requirements and design is given. Followed by an introduction to the developed techniques, giving full detail of its mathematical background and its procedures. Later, the test case used is described and result from the application of the techniques is shown. Finally the conclusions are presented and future lines of work are exposed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dimensionality Reduction (DR) is attracting more attention these days as a result of the increasing need to handle huge amounts of data effectively. DR methods allow the number of initial features to be reduced considerably until a set of them is found that allows the original properties of the data to be kept. However, their use entails an inherent loss of quality that is likely to affect the understanding of the data, in terms of data analysis. This loss of quality could be determinant when selecting a DR method, because of the nature of each method. In this paper, we propose a methodology that allows different DR methods to be analyzed and compared as regards the loss of quality produced by them. This methodology makes use of the concept of preservation of geometry (quality assessment criteria) to assess the loss of quality. Experiments have been carried out by using the most well-known DR algorithms and quality assessment criteria, based on the literature. These experiments have been applied on 12 real-world datasets. Results obtained so far show that it is possible to establish a method to select the most appropriate DR method, in terms of minimum loss of quality. Experiments have also highlighted some interesting relationships between the quality assessment criteria. Finally, the methodology allows the appropriate choice of dimensionality for reducing data to be established, whilst giving rise to a minimum loss of quality.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Subtraction of Ictal SPECT Co-registered to MRI (SISCOM) is an imaging technique used to localize the epileptogenic focus in patients with intractable partial epilepsy. The aim of this study was to determine the accuracy of registration algorithms involved in SISCOM analysis using FocusDET, a new user-friendly application. To this end, Monte Carlo simulation was employed to generate realistic SPECT studies. Simulated sinograms were reconstructed by using the Filtered BackProjection (FBP) algorithm and an Ordered Subsets Expectation Maximization (OSEM) reconstruction method that included compensation for all degradations. Registration errors in SPECT-SPECT and SPECT-MRI registration were evaluated by comparing the theoretical and actual transforms. Patient studies with well-localized epilepsy were also included in the registration assessment. Global registration errors including SPECT-SPECT and SPECT-MRI registration errors were less than 1.2 mm on average, exceeding the voxel size (3.32 mm) of SPECT studies in no case. Although images reconstructed using OSEM led to lower registration errors than images reconstructed with FBP, differences after using OSEM or FBP in reconstruction were less than 0.2 mm on average. This indicates that correction for degradations does not play a major role in the SISCOM process, thereby facilitating the application of the methodology in centers where OSEM is not implemented with correction of all degradations. These findings together with those obtained by clinicians from patients via MRI, interictal and ictal SPECT and video-EEG, show that FocusDET is a robust application for performing SISCOM analysis in clinical practice.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

At present, all methods in Evolutionary Computation are bioinspired by the fundamental principles of neo-Darwinism, as well as by a vertical gene transfer. Virus transduction is one of the key mechanisms of horizontal gene propagation in microorganisms (e.g. bacteria). In the present paper, we model and simulate a transduction operator, exploring the possible role and usefulness of transduction in a genetic algorithm. The genetic algorithm including transduction has been named PETRI (abbreviation of Promoting Evolution Through Reiterated Infection). Our results showed how PETRI approaches higher fitness values as transduction probability comes close to 100%. The conclusion is that transduction improves the performance of a genetic algorithm, assuming a population divided among several sub-populations or ?bacterial colonies?.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El artículo aborda el problema del encaje de diversas imágenes de una misma escena capturadas por escáner 3d para generar un único modelo tridimensional. Para ello se utilizaron algoritmos genéticos. ABSTRACT: This work introduces a solution based on genetic algorithms to find the overlapping area between two point cloud captures obtained from a three-dimensional scanner. Considering three translation coordinates and three rotation angles, the genetic algorithm evaluates the matching points in the overlapping area between the two captures given that transformation. Genetic simulated annealing is used to improve the accuracy of the results obtained by the genetic algorithm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

8th International Conference on Fracture Mechanics of Concrete and Concrete Structures (FraMCoS8).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Apples can be considered as having a complex system formed by several structures at different organization levels: macroscale (mayor que100 ?m) and microscale (menor que100 ?m). This work implements 2D T1/T2 global and localized relaxometry sequences on whole apples to be able to perform an intensive non-destructive and non-invasive microstructure study. The 2D T1/T2 cross-correlation spectroscopy allows the extraction of quantitative information about the water compartmentation in different subcellular organelles. A clear difference is found as sound apples show neat peaks for water in different subcellular compartments, such as vacuolar, cytoplasmatic and extracellular water, while in watercore-affected tissues such compartments appear merged. Localized relaxometry allows for the predefinition of slices in order to understand the microstructure of a particular region of the fruit, providing information that cannot be derived from global 2D T1/T2 relaxometry.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this article, a novel approach to deal with the design of in-building wireless networks deployments is proposed. This approach known as MOQZEA (Multiobjective Quality Zone Based Evolutionary Algorithm) is a hybr id evolutionary algorithm adapted to use a novel fitness function, based on the definition of quality zones for the different objective functions considered. This approach is conceived to solve wireless network design problems without previous information of the required number of transmitters, considering simultaneously a high number of objective functions and optimizing multiple configuration parameters of the transmitters.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Este trabajo es una contribución a los sistemas fotovoltaicos (FV) con seguimiento distribuido del punto de máxima potencia (DMPPT), una topología que se caracteriza porque lleva a cabo el MPPT a nivel de módulo, al contrario de las topologías más tradicionales que llevan a cabo el MPPT para un número más elevado de módulos, pudiendo ser hasta cientos de módulos. Las dos tecnologías DMPPT que existen en el mercado son conocidos como microinversores y optimizadores de potencia, y ofrecen ciertas ventajas sobre sistemas de MPPT central como: mayor producción en situaciones de mismatch, monitorización individual de cada módulo, flexibilidad de diseño, mayor seguridad del sistema, etc. Aunque los sistemas DMPPT no están limitados a los entornos urbanos, se ha enfatizado en el título ya que es su mercado natural, siendo difícil una justificación de su sobrecoste en grandes huertas solares en suelo. Desde el año 2010 el mercado de estos sistemas ha incrementado notablemente y sigue creciendo de una forma continuada. Sin embargo, todavía falta un conocimiento profundo de cómo funcionan estos sistemas, especialmente en el caso de los optimizadores de potencia, de las ganancias energéticas esperables en condiciones de mismatch y de las posibilidades avanzadas de diagnóstico de fallos. El principal objetivo de esta tesis es presentar un estudio completo de cómo funcionan los sistemas DMPPT, sus límites y sus ventajas, así como experimentos varios que verifican la teoría y el desarrollo de herramientas para valorar las ventajas de utilizar DMPPT en cada instalación. Las ecuaciones que modelan el funcionamiento de los sistemas FVs con optimizadores de potencia se han desarrollado y utilizado para resaltar los límites de los mismos a la hora de resolver ciertas situaciones de mismatch. Se presenta un estudio profundo sobre el efecto de las sombras en los sistemas FVs: en la curva I-V y en los algoritmos MPPT. Se han llevado a cabo experimentos sobre el funcionamiento de los algoritmos MPPT en situaciones de sombreado, señalando su ineficiencia en estas situaciones. Un análisis de la ventaja del uso de DMPPT frente a los puntos calientes es presentado y verificado. También se presenta un análisis sobre las posibles ganancias en potencia y energía con el uso de DMPPT en condiciones de sombreado y este también es verificado experimentalmente, así como un breve estudio de su viabilidad económica. Para ayudar a llevar a cabo todos los análisis y experimentos descritos previamente se han desarrollado una serie de herramientas software. Una siendo un programa en LabView para controlar un simulador solar y almacenar las medidas. También se ha desarrollado un programa que simula curvas I-V de módulos y generador FVs afectados por sombras y este se ha verificado experimentalmente. Este mismo programa se ha utilizado para desarrollar un programa todavía más completo que estima las pérdidas anuales y las ganancias obtenidas con DMPPT en instalaciones FVs afectadas por sombras. Finalmente, se han desarrollado y verificado unos algoritmos para diagnosticar fallos en sistemas FVs con DMPPT. Esta herramienta puede diagnosticar los siguientes fallos: sombras debido a objetos fijos (con estimación de la distancia al objeto), suciedad localizada, suciedad general, posible punto caliente, degradación de módulos y pérdidas en el cableado de DC. Además, alerta al usuario de las pérdidas producidas por cada fallo y no requiere del uso de sensores de irradiancia y temperatura. ABSTRACT This work is a contribution to photovoltaic (PV) systems with distributed maximum power point tracking (DMPPT), a system topology characterized by performing the MPPT at module level, instead of the more traditional topologies which perform MPPT for a larger number of modules. The two DMPPT technologies available at the moment are known as microinverters and power optimizers, also known as module level power electronics (MLPE), and they provide certain advantages over central MPPT systems like: higher energy production in mismatch situations, monitoring of each individual module, system design flexibility, higher system safety, etc. Although DMPPT is not limited to urban environments, it has been emphasized in the title as it is their natural market, since in large ground-mounted PV plants the extra cost is difficult to justify. Since 2010 MLPE have increased their market share steadily and continuing to grow steadily. However, there still lacks a profound understanding of how they work, especially in the case of power optimizers, the achievable energy gains with their use and the possibilities in failure diagnosis. The main objective of this thesis is to provide a complete understanding of DMPPT technologies: how they function, their limitations and their advantages. A series of equations used to model PV arrays with power optimizers have been derived and used to point out limitations in solving certain mismatch situation. Because one of the most emphasized benefits of DMPPT is their ability to mitigate shading losses, an extensive study on the effects of shadows on PV systems is presented; both on the I-V curve and on MPPT algorithms. Experimental tests have been performed on the MPPT algorithms of central inverters and MLPE, highlighting their inefficiency in I-V curves with local maxima. An analysis of the possible mitigation of hot-spots with DMPPT is discussed and experimentally verified. And a theoretical analysis of the possible power and energy gains is presented as well as experiments in real PV systems. A short economic analysis of the benefits of DMPPT has also been performed. In order to aide in the previous task, a program which simulates I-V curves under shaded conditions has been developed and experimentally verified. This same program has been used to develop a software tool especially designed for PV systems affected by shading, which estimates the losses due to shading and the energy gains obtained with DMPPT. Finally, a set of algorithms for diagnosing system faults in PV systems with DMPPT has been developed and experimentally verified. The tool can diagnose the following failures: fixed object shading (with distance estimation), localized dirt, generalized dirt, possible hot-spots, module degradation and excessive losses in DC cables. In addition, it alerts the user of the power losses produced by each failure and classifies the failures by their severity and it does not require the use of irradiance or temperature sensors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper describes the objectives, content, learning methodology and results of an online course on the History of Algorithms for engineering students at Polytechnic University of Madrid (UPM). This course is conducted in a virtual environment based on Moodle, with a student-centred educational model which includes a detailed planning of learning activities. Our experience indicates that this subject is highly motivating for students and the virtual environment facilitates competencies development

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Este trabajo es una contribución a los sistemas fotovoltaicos (FV) con seguimiento distribuido del punto de máxima potencia (DMPPT), una topología que se caracteriza porque lleva a cabo el MPPT a nivel de módulo, al contrario de las topologías más tradicionales que llevan a cabo el MPPT para un número más elevado de módulos, pudiendo ser hasta cientos de módulos. Las dos tecnologías DMPPT que existen en el mercado son conocidos como microinversores y optimizadores de potencia, y ofrecen ciertas ventajas sobre sistemas de MPPT central como: mayor producción en situaciones de mismatch, monitorización individual de cada módulo, flexibilidad de diseño, mayor seguridad del sistema, etc. Aunque los sistemas DMPPT no están limitados a los entornos urbanos, se ha enfatizado en el título ya que es su mercado natural, siendo difícil una justificación de su sobrecoste en grandes huertas solares en suelo. Desde el año 2010 el mercado de estos sistemas ha incrementado notablemente y sigue creciendo de una forma continuada. Sin embargo, todavía falta un conocimiento profundo de cómo funcionan estos sistemas, especialmente en el caso de los optimizadores de potencia, de las ganancias energéticas esperables en condiciones de mismatch y de las posibilidades avanzadas de diagnóstico de fallos. El principal objetivo de esta tesis es presentar un estudio completo de cómo funcionan los sistemas DMPPT, sus límites y sus ventajas, así como experimentos varios que verifican la teoría y el desarrollo de herramientas para valorar las ventajas de utilizar DMPPT en cada instalación. Las ecuaciones que modelan el funcionamiento de los sistemas FVs con optimizadores de potencia se han desarrollado y utilizado para resaltar los límites de los mismos a la hora de resolver ciertas situaciones de mismatch. Se presenta un estudio profundo sobre el efecto de las sombras en los sistemas FVs: en la curva I-V y en los algoritmos MPPT. Se han llevado a cabo experimentos sobre el funcionamiento de los algoritmos MPPT en situaciones de sombreado, señalando su ineficiencia en estas situaciones. Un análisis de la ventaja del uso de DMPPT frente a los puntos calientes es presentado y verificado. También se presenta un análisis sobre las posibles ganancias en potencia y energía con el uso de DMPPT en condiciones de sombreado y este también es verificado experimentalmente, así como un breve estudio de su viabilidad económica. Para ayudar a llevar a cabo todos los análisis y experimentos descritos previamente se han desarrollado una serie de herramientas software. Una siendo un programa en LabView para controlar un simulador solar y almacenar las medidas. También se ha desarrollado un programa que simula curvas I-V de módulos y generador FVs afectados por sombras y este se ha verificado experimentalmente. Este mismo programa se ha utilizado para desarrollar un programa todavía más completo que estima las pérdidas anuales y las ganancias obtenidas con DMPPT en instalaciones FVs afectadas por sombras. Finalmente, se han desarrollado y verificado unos algoritmos para diagnosticar fallos en sistemas FVs con DMPPT. Esta herramienta puede diagnosticar los siguientes fallos: sombras debido a objetos fijos (con estimación de la distancia al objeto), suciedad localizada, suciedad general, posible punto caliente, degradación de módulos y pérdidas en el cableado de DC. Además, alerta al usuario de las pérdidas producidas por cada fallo y no requiere del uso de sensores de irradiancia y temperatura. ABSTRACT This work is a contribution to photovoltaic (PV) systems with distributed maximum power point tracking (DMPPT), a system topology characterized by performing the MPPT at module level, instead of the more traditional topologies which perform MPPT for a larger number of modules. The two DMPPT technologies available at the moment are known as microinverters and power optimizers, also known as module level power electronics (MLPE), and they provide certain advantages over central MPPT systems like: higher energy production in mismatch situations, monitoring of each individual module, system design flexibility, higher system safety, etc. Although DMPPT is not limited to urban environments, it has been emphasized in the title as it is their natural market, since in large ground-mounted PV plants the extra cost is difficult to justify. Since 2010 MLPE have increased their market share steadily and continuing to grow steadily. However, there still lacks a profound understanding of how they work, especially in the case of power optimizers, the achievable energy gains with their use and the possibilities in failure diagnosis. The main objective of this thesis is to provide a complete understanding of DMPPT technologies: how they function, their limitations and their advantages. A series of equations used to model PV arrays with power optimizers have been derived and used to point out limitations in solving certain mismatch situation. Because one of the most emphasized benefits of DMPPT is their ability to mitigate shading losses, an extensive study on the effects of shadows on PV systems is presented; both on the I-V curve and on MPPT algorithms. Experimental tests have been performed on the MPPT algorithms of central inverters and MLPE, highlighting their inefficiency in I-V curves with local maxima. An analysis of the possible mitigation of hot-spots with DMPPT is discussed and experimentally verified. And a theoretical analysis of the possible power and energy gains is presented as well as experiments in real PV systems. A short economic analysis of the benefits of DMPPT has also been performed. In order to aide in the previous task, a program which simulates I-V curves under shaded conditions has been developed and experimentally verified. This same program has been used to develop a software tool especially designed for PV systems affected by shading, which estimates the losses due to shading and the energy gains obtained with DMPPT. Finally, a set of algorithms for diagnosing system faults in PV systems with DMPPT has been developed and experimentally verified. The tool can diagnose the following failures: fixed object shading (with distance estimation), localized dirt, generalized dirt, possible hot-spots, module degradation and excessive losses in DC cables. In addition, it alerts the user of the power losses produced by each failure and classifies the failures by their severity and it does not require the use of irradiance or temperature sensors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Plant diseases represent a major economic and environmental problem in agriculture and forestry. Upon infection, a plant develops symptoms that affect different parts of the plant causing a significant agronomic impact. As many such diseases spread in time over the whole crop, a system for early disease detection can aid to mitigate the losses produced by the plant diseases and can further prevent their spread [1]. In recent years, several mathematical algorithms of search have been proposed [2,3] that could be used as a non-invasive, fast, reliable and cost-effective methods to localize in space infectious focus by detecting changes in the profile of volatile organic compounds. Tracking scents and locating odor sources is a major challenge in robotics, on one hand because odour plumes consists of non-uniform intermittent odour patches dispersed by the wind and on the other hand because of the lack of precise and reliable odour sensors. Notwithstanding, we have develop a simple robotic platform to study the robustness and effectiveness of different search algorithms [4], with respect to specific problems to be found in their further application in agriculture, namely errors committed in the motion and sensing and to the existence of spatial constraints due to land topology or the presence of obstacles.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The diversity of bibliometric indices today poses the challenge of exploiting the relationships among them. Our research uncovers the best core set of relevant indices for predicting other bibliometric indices. An added difficulty is to select the role of each variable, that is, which bibliometric indices are predictive variables and which are response variables. This results in a novel multioutput regression problem where the role of each variable (predictor or response) is unknown beforehand. We use Gaussian Bayesian networks to solve the this problem and discover multivariate relationships among bibliometric indices. These networks are learnt by a genetic algorithm that looks for the optimal models that best predict bibliometric data. Results show that the optimal induced Gaussian Bayesian networks corroborate previous relationships between several indices, but also suggest new, previously unreported interactions. An extended analysis of the best model illustrates that a set of 12 bibliometric indices can be accurately predicted using only a smaller predictive core subset composed of citations, g-index, q2-index, and hr-index. This research is performed using bibliometric data on Spanish full professors associated with the computer science area.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

One of the most promising areas in which probabilistic graphical models have shown an incipient activity is the field of heuristic optimization and, in particular, in Estimation of Distribution Algorithms. Due to their inherent parallelism, different research lines have been studied trying to improve Estimation of Distribution Algorithms from the point of view of execution time and/or accuracy. Among these proposals, we focus on the so-called distributed or island-based models. This approach defines several islands (algorithms instances) running independently and exchanging information with a given frequency. The information sent by the islands can be either a set of individuals or a probabilistic model. This paper presents a comparative study for a distributed univariate Estimation of Distribution Algorithm and a multivariate version, paying special attention to the comparison of two alternative methods for exchanging information, over a wide set of parameters and problems ? the standard benchmark developed for the IEEE Workshop on Evolutionary Algorithms and other Metaheuristics for Continuous Optimization Problems of the ISDA 2009 Conference. Several analyses from different points of view have been conducted to analyze both the influence of the parameters and the relationships between them including a characterization of the configurations according to their behavior on the proposed benchmark.