856 resultados para Population set-based methods


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper we propose a flexible Multi-Agent Architecture together with a methodology for indoor location which allows us to locate any mobile station (MS) such as a Laptop, Smartphone, Tablet or a robotic system in an indoor environment using wireless technology. Our technology is complementary to the GPS location finder as it allows us to locate a mobile system in a specific room on a specific floor using the Wi-Fi networks. The idea is that any MS will have an agent known at a Fuzzy Location Software Agent (FLSA) with a minimum capacity processing at its disposal which collects the power received at different Access Points distributed around the floor and establish its location on a plan of the floor of the building. In order to do so it will have to communicate with the Fuzzy Location Manager Software Agent (FLMSA). The FLMSAs are local agents that form part of the management infrastructure of the Wi-Fi network of the Organization. The FLMSA implements a location estimation methodology divided into three phases (measurement, calibration and estimation) for locating mobile stations (MS). Our solution is a fingerprint-based positioning system that overcomes the problem of the relative effect of doors and walls on signal strength and is independent of the network device manufacturer. In the measurement phase, our system collects received signal strength indicator (RSSI) measurements from multiple access points. In the calibration phase, our system uses these measurements in a normalization process to create a radio map, a database of RSS patterns. Unlike traditional radio map-based methods, our methodology normalizes RSS measurements collected at different locations on a floor. In the third phase, we use Fuzzy Controllers to locate an MS on the plan of the floor of a building. Experimental results demonstrate the accuracy of the proposed method. From these results it is clear that the system is highly likely to be able to locate an MS in a room or adjacent room.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Leaf nitrogen and leaf surface area influence the exchange of gases between terrestrial ecosystems and the atmosphere, and play a significant role in the global cycles of carbon, nitrogen and water. The purpose of this study is to use field-based and satellite remote-sensing-based methods to assess leaf nitrogen pools in five diverse European agricultural landscapes located in Denmark, Scotland (United Kingdom), Poland, the Netherlands and Italy. REGFLEC (REGularized canopy reFLECtance) is an advanced image-based inverse canopy radiative transfer modelling system which has shown proficiency for regional mapping of leaf area index (LAI) and leaf chlorophyll (CHLl) using remote sensing data. In this study, high spatial resolution (10–20 m) remote sensing images acquired from the multispectral sensors aboard the SPOT (Satellite For Observation of Earth) satellites were used to assess the capability of REGFLEC for mapping spatial variations in LAI, CHLland the relation to leaf nitrogen (Nl) data in five diverse European agricultural landscapes. REGFLEC is based on physical laws and includes an automatic model parameterization scheme which makes the tool independent of field data for model calibration. In this study, REGFLEC performance was evaluated using LAI measurements and non-destructive measurements (using a SPAD meter) of leaf-scale CHLl and Nl concentrations in 93 fields representing crop- and grasslands of the five landscapes. Furthermore, empirical relationships between field measurements (LAI, CHLl and Nl and five spectral vegetation indices (the Normalized Difference Vegetation Index, the Simple Ratio, the Enhanced Vegetation Index-2, the Green Normalized Difference Vegetation Index, and the green chlorophyll index) were used to assess field data coherence and to serve as a comparison basis for assessing REGFLEC model performance. The field measurements showed strong vertical CHLl gradient profiles in 26% of fields which affected REGFLEC performance as well as the relationships between spectral vegetation indices (SVIs) and field measurements. When the range of surface types increased, the REGFLEC results were in better agreement with field data than the empirical SVI regression models. Selecting only homogeneous canopies with uniform CHLl distributions as reference data for evaluation, REGFLEC was able to explain 69% of LAI observations (rmse = 0.76), 46% of measured canopy chlorophyll contents (rmse = 719 mg m−2) and 51% of measured canopy nitrogen contents (rmse = 2.7 g m−2). Better results were obtained for individual landscapes, except for Italy, where REGFLEC performed poorly due to a lack of dense vegetation canopies at the time of satellite recording. Presence of vegetation is needed to parameterize the REGFLEC model. Combining REGFLEC- and SVI-based model results to minimize errors for a "snap-shot" assessment of total leaf nitrogen pools in the five landscapes, results varied from 0.6 to 4.0 t km−2. Differences in leaf nitrogen pools between landscapes are attributed to seasonal variations, extents of agricultural area, species variations, and spatial variations in nutrient availability. In order to facilitate a substantial assessment of variations in Nl pools and their relation to landscape based nitrogen and carbon cycling processes, time series of satellite data are needed. The upcoming Sentinel-2 satellite mission will provide new multiple narrowband data opportunities at high spatio-temporal resolution which are expected to further improve remote sensing capabilities for mapping LAI, CHLl and Nl.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background Magnetoencephalography (MEG) provides a direct measure of brain activity with high combined spatiotemporal resolution. Preprocessing is necessary to reduce contributions from environmental interference and biological noise. New method The effect on the signal-to-noise ratio of different preprocessing techniques is evaluated. The signal-to-noise ratio (SNR) was defined as the ratio between the mean signal amplitude (evoked field) and the standard error of the mean over trials. Results Recordings from 26 subjects obtained during and event-related visual paradigm with an Elekta MEG scanner were employed. Two methods were considered as first-step noise reduction: Signal Space Separation and temporal Signal Space Separation, which decompose the signal into components with origin inside and outside the head. Both algorithm increased the SNR by approximately 100%. Epoch-based methods, aimed at identifying and rejecting epochs containing eye blinks, muscular artifacts and sensor jumps provided an SNR improvement of 5–10%. Decomposition methods evaluated were independent component analysis (ICA) and second-order blind identification (SOBI). The increase in SNR was of about 36% with ICA and 33% with SOBI. Comparison with existing methods No previous systematic evaluation of the effect of the typical preprocessing steps in the SNR of the MEG signal has been performed. Conclusions The application of either SSS or tSSS is mandatory in Elekta systems. No significant differences were found between the two. While epoch-based methods have been routinely applied the less often considered decomposition methods were clearly superior and therefore their use seems advisable.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Scaling is becoming an increasingly important topic in the earth and environmental sciences as researchers attempt to understand complex natural systems through the lens of an ever-increasing set of methods and scales. The guest editors introduce the papers in this issue’s special section and present an overview of some of the work being done. Scaling remains one of the most challenging topics in earth and environmental sciences, forming a basis for our understanding of process development across the multiple scales that make up the subsurface environment. Tremendous progress has been made in discovery, explanation, and applications of scaling. And yet much more needs to be done and is being done as part of the modern quest to quantify, analyze, and manage the complexity of natural systems. Understanding and succinct representation of scaling properties can unveil underlying relationships between system structure and response functions, improve parameterization of natural variability and heterogeneity, and help us address societal needs by effectively merging knowledge acquired at different scales.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The cyclic compression of several granular systems has been simulated with a molecular dynamics code. All the samples consisted of bidimensional, soft, frictionless and equal-sized particles that were initially arranged according to a squared lattice and were compressed by randomly generated irregular walls. The compression protocols can be described by some control variables (volume or external force acting on the walls) and by some dimensionless factors, that relate stiffness, density, diameter, damping ratio and water surface tension to the external forces, displacements and periods. Each protocol, that is associated to a dynamic process, results in an arrangement with its own macroscopic features: volume (or packing ratio), coordination number, and stress; and the differences between packings can be highly significant. The statistical distribution of the force-moment state of the particles (i.e. the equivalent average stress multiplied by the volume) is analyzed. In spite of the lack of a theoretical framework based on statistical mechanics specific for these protocols, it is shown how the obtained distributions of mean and relative deviatoric force-moment are. Then it is discussed on the nature of these distributions and on their relation to specific protocols.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La energía basada en turbinas hidráulicas de reducida potencia es, a menudo, un ejemplo- modelo dentro del campo de las energías renovables desde su aparición a finales del S. XIX, aunque los ingenios precursores surgen mucho antes. Entre los testimonios más antiguos destacan la saqia o rueda persa y la rueda hidráulica romana que había sido previamente implementada en Extremo Oriente, y que después llega a Europa a través de Egipto. Más tarde, durante la Edad Media y el Renacimiento, se generaliza el uso de los molinos hidráulicos, además de los eólicos. Ejemplos de ello son las norias de Alepo (Siria) y de Córdoba (España). Otro caso interesante es el de los molinos de regolfo en la Península Ibérica e Iberoamérica, muy cercanos en su forma y fundamentos a las turbinas hidráulicas. Algunos de estos ingenios siguen activos en los ríos de España. Posteriormente los estudios de Euler, Burdin y Fourneyron prepararon las bases para el definitivo avance de Pelton, Kaplan, Francis, y otros, pero ya como máquinas motrices de los generadores eléctricos. En la actualidad, se admite como límite superior para minihidráulica aquellas centrales con una potencia instalada de 5000 kW, y considerando que cuando las potencias son inferiores a 100 kW se denomina micro hidráulica, aunque en Latinoamérica este límite se fija en 20 kW. El estudio del recurso hídrico, ayudado del geotécnico, constituyen la base sobre la que podremos proyectar el aprovechamiento hidroeléctrico: selección del tipo de central dentro de la tipología disponible, diseño y cálculos de la turbina, obra civil necesaria (azud, presa, canal, tubería forzada, edificio, aspiración, desagüe, etc.) y equipo electromecánico. El proyecto tecnológico se complementa con el estudio de impacto ambiental y de viabilidad económica. Muchos de estos proyectos tratan de reducir la falta de acceso a la energía en poblaciones desfavorecidas, entendida esta carencia como un factor determinante de la pobreza. Así la energía mini y micro-hidráulica adquiere un nuevo valor como tecnología para el desarrollo humano.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La educación y los entornos educativos están en constante evolución. Tanto alumnos como educadores cambian de hábitos, de maneras de aprender, de gustos, de dispositivos que manejan y de aplicaciones que usan regularmente entre otras cosas. Todos estos cambios vienen acompañados y fomentados en gran medida por la evolución paralela que experimenta la tecnología, tanto software en los programas utilizados, como hardware en los dispositivos y capacidades de éstos. La educación debe también adaptarse a estos cambios, tanto personales como tecnológicos y sacar el mayor provecho de ellos. El uso de Sistemas de Gestión del Aprendizaje está muy extendido en todos los centros educativos. Estos sistemas poseen un gran número de características y funcionalidades que permiten desde la aplicación de un modelo didáctico totalmente tradicional en el que el profesor imparte un contenido y los alumnos lo reciben a uno totalmente innovador en el que ocurren procesos totalmente diferentes. Por otro lado, el potencial que ofrecen los recursos multimedia no ha sido completamente aprovechado en la educación y supone una gran oportunidad. Esta tesis doctoral propone un conjunto de métodos y herramientas para la creación y el uso de recursos multimedia en la educación. Para ello el desarrollo de esta tesis parte de la definición de un modelo didáctico social, colaborativo y centrado en el alumno que servirá de hilo conductor y que integrará los diferentes y métodos y herramientas estudiados y desarrollados. En un primer paso se identifican varias herramientas y métodos para el aula, tales son la grabación de clases, donde se crea y posteriormente se mejora un carrito portátil de grabación que da muy buen resultado, las herramientas de grabación de screencast y la videoconferencia. Estas herramientas además se integran en una plataforma colaborativa dando lugar a una arquitectura completa y escalable que permite la realización de dichas actividades y la interconexión sencilla con el Sistema de Gestión del Aprendizaje. A continuación y ya en un entorno totalmente online se desarrolla una nueva plataforma de e-learning llamada Virtual Science Hub (ViSH) que consta de cuatro funcionalidades principales, red social, videoconferencia, repositorio educativo y herramienta de autor. En esta plataforma se aplicaron técnicas de recomendación proactiva tanto de recursos educativos como de otros usuarios similares. Por último se validó el modelo educativo completo usando algunas de las herramientas identificadas y desarrolladas en dos escenarios diferentes con gran éxito. Finalmente, esta tesis discute las conclusiones obtenidas a lo largo de la extensa investigación llevada a cabo y que ha propiciado la consecución de una buena base teórica y práctica para la creación de herramientas y métodos para la generación y el uso de recursos multimedia en la educación. ABSTRACT Education and learning environments are constantly evolving. Students and educators change the things their habits, their ways of learning, the things they like or the devices and applications that they use regularly among other things. All these changes are accompanied and fostered by the parallel evolution that technology experiences, both in the software programs used as in the hardware and capabilities of these devices. Education must also adapt to these changes, both personal and technological and get the most out of them. Learning Management Systems are widely used in all educational centers. These systems have a large number of features and functionalities. They allow from the implementation of a traditional teaching model in which the teacher gives content and students receive it to one absolutely innovative teaching model where totally different processes occur. Furthermore, the potential of multimedia resources has not been fully exploited in education and can be a great opportunity. This thesis proposes a set of methods and tools for the creation and use of multimedia in education. The development of this thesis starts with the definition of a social, collaborative and learner-centered model, that serves as a common thread and that integrates different tools and methods studied and developed. In a first step, several tools and methods for the classroom are identified, such as recording, where a portable kit is created and then improved giving very good results, screencast recording and videoconferencing. These tools also are integrated into a collaborative platform resulting in a complete, scalable architecture that enables the execution of such activities and a simple interconnection with the Learning Management System. In an fully online environment a new e-learning platform called Virtual Science Hub (ViSH) is created. It consists of four main features that combine and complement each other, social network, videoconferencing, educational repository and authoring tool. In this platform proactive recommendation of both educational resources and similar users is applied. In a last step the entire educational model using some of the tools identified and developed is successfully validated in two different scenarios. Finally, this thesis discusses the findings obtained during the extensive research carried out and has led to the achievement of a good theoretical and practical basis for the development of tools and methods for the generation and use of multimedia in education.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Project you are about to see it is based on the technologies used on object detection and recognition, especially on leaves and chromosomes. To do so, this document contains the typical parts of a scientific paper, as it is what it is. It is composed by an Abstract, an Introduction, points that have to do with the investigation area, future work, conclusions and references used for the elaboration of the document. The Abstract talks about what are we going to find in this paper, which is technologies employed on pattern detection and recognition for leaves and chromosomes and the jobs that are already made for cataloguing these objects. In the introduction detection and recognition meanings are explained. This is necessary as many papers get confused with these terms, specially the ones talking about chromosomes. Detecting an object is gathering the parts of the image that are useful and eliminating the useless parts. Summarizing, detection would be recognizing the objects borders. When talking about recognition, we are talking about the computers or the machines process, which says what kind of object we are handling. Afterwards we face a compilation of the most used technologies in object detection in general. There are two main groups on this category: Based on derivatives of images and based on ASIFT points. The ones that are based on derivatives of images have in common that convolving them with a previously created matrix does the treatment of them. This is done for detecting borders on the images, which are changes on the intensity of the pixels. Within these technologies we face two groups: Gradian based, which search for maximums and minimums on the pixels intensity as they only use the first derivative. The Laplacian based methods search for zeros on the pixels intensity as they use the second derivative. Depending on the level of details that we want to use on the final result, we will choose one option or the other, because, as its logic, if we used Gradian based methods, the computer will consume less resources and less time as there are less operations, but the quality will be worse. On the other hand, if we use the Laplacian based methods we will need more time and resources as they require more operations, but we will have a much better quality result. After explaining all the derivative based methods, we take a look on the different algorithms that are available for both groups. The other big group of technologies for object recognition is the one based on ASIFT points, which are based on 6 image parameters and compare them with another image taking under consideration these parameters. These methods disadvantage, for our future purposes, is that it is only valid for one single object. So if we are going to recognize two different leaves, even though if they refer to the same specie, we are not going to be able to recognize them with this method. It is important to mention these types of technologies as we are talking about recognition methods in general. At the end of the chapter we can see a comparison with pros and cons of all technologies that are employed. Firstly comparing them separately and then comparing them all together, based on our purposes. Recognition techniques, which are the next chapter, are not really vast as, even though there are general steps for doing object recognition, every single object that has to be recognized has its own method as the are different. This is why there is not a general method that we can specify on this chapter. We now move on into leaf detection techniques on computers. Now we will use the technique explained above based on the image derivatives. Next step will be to turn the leaf into several parameters. Depending on the document that you are referring to, there will be more or less parameters. Some papers recommend to divide the leaf into 3 main features (shape, dent and vein] and doing mathematical operations with them we can get up to 16 secondary features. Next proposition is dividing the leaf into 5 main features (Diameter, physiological length, physiological width, area and perimeter] and from those, extract 12 secondary features. This second alternative is the most used so it is the one that is going to be the reference. Following in to leaf recognition, we are based on a paper that provides a source code that, clicking on both leaf ends, it automatically tells to which specie belongs the leaf that we are trying to recognize. To do so, it only requires having a database. On the tests that have been made by the document, they assure us a 90.312% of accuracy over 320 total tests (32 plants on the database and 10 tests per specie]. Next chapter talks about chromosome detection, where we shall pass the metaphasis plate, where the chromosomes are disorganized, into the karyotype plate, which is the usual view of the 23 chromosomes ordered by number. There are two types of techniques to do this step: the skeletonization process and swiping angles. Skeletonization progress consists on suppressing the inside pixels of the chromosome to just stay with the silhouette. This method is really similar to the ones based on the derivatives of the image but the difference is that it doesnt detect the borders but the interior of the chromosome. Second technique consists of swiping angles from the beginning of the chromosome and, taking under consideration, that on a single chromosome we cannot have more than an X angle, it detects the various regions of the chromosomes. Once the karyotype plate is defined, we continue with chromosome recognition. To do so, there is a technique based on the banding that chromosomes have (grey scale bands] that make them unique. The program then detects the longitudinal axis of the chromosome and reconstructs the band profiles. Then the computer is able to recognize this chromosome. Concerning the future work, we generally have to independent techniques that dont reunite detection and recognition, so our main focus would be to prepare a program that gathers both techniques. On the leaf matter we have seen that, detection and recognition, have a link as both share the option of dividing the leaf into 5 main features. The work that would have to be done is to create an algorithm that linked both methods, as in the program, which recognizes leaves, it has to be clicked both leaf ends so it is not an automatic algorithm. On the chromosome side, we should create an algorithm that searches for the beginning of the chromosome and then start to swipe angles, to later give the parameters to the program that searches for the band profiles. Finally, on the summary, we explain why this type of investigation is needed, and that is because with global warming, lots of species (animals and plants] are beginning to extinguish. That is the reason why a big database, which gathers all the possible species, is needed. For recognizing animal species, we just only have to have the 23 chromosomes. While recognizing a plant, there are several ways of doing it, but the easiest way to input a computer is to scan the leaf of the plant. RESUMEN. El proyecto que se puede ver a continuación trata sobre las tecnologías empleadas en la detección y reconocimiento de objetos, especialmente de hojas y cromosomas. Para ello, este documento contiene las partes típicas de un paper de investigación, puesto que es de lo que se trata. Así, estará compuesto de Abstract, Introducción, diversos puntos que tengan que ver con el área a investigar, trabajo futuro, conclusiones y biografía utilizada para la realización del documento. Así, el Abstract nos cuenta qué vamos a poder encontrar en este paper, que no es ni más ni menos que las tecnologías empleadas en el reconocimiento y detección de patrones en hojas y cromosomas y qué trabajos hay existentes para catalogar a estos objetos. En la introducción se explican los conceptos de qué es la detección y qué es el reconocimiento. Esto es necesario ya que muchos papers científicos, especialmente los que hablan de cromosomas, confunden estos dos términos que no podían ser más sencillos. Por un lado tendríamos la detección del objeto, que sería simplemente coger las partes que nos interesasen de la imagen y eliminar aquellas partes que no nos fueran útiles para un futuro. Resumiendo, sería reconocer los bordes del objeto de estudio. Cuando hablamos de reconocimiento, estamos refiriéndonos al proceso que tiene el ordenador, o la máquina, para decir qué clase de objeto estamos tratando. Seguidamente nos encontramos con un recopilatorio de las tecnologías más utilizadas para la detección de objetos, en general. Aquí nos encontraríamos con dos grandes grupos de tecnologías: Las basadas en las derivadas de imágenes y las basadas en los puntos ASIFT. El grupo de tecnologías basadas en derivadas de imágenes tienen en común que hay que tratar a las imágenes mediante una convolución con una matriz creada previamente. Esto se hace para detectar bordes en las imágenes que son básicamente cambios en la intensidad de los píxeles. Dentro de estas tecnologías nos encontramos con dos grupos: Los basados en gradientes, los cuales buscan máximos y mínimos de intensidad en la imagen puesto que sólo utilizan la primera derivada; y los Laplacianos, los cuales buscan ceros en la intensidad de los píxeles puesto que estos utilizan la segunda derivada de la imagen. Dependiendo del nivel de detalles que queramos utilizar en el resultado final nos decantaremos por un método u otro puesto que, como es lógico, si utilizamos los basados en el gradiente habrá menos operaciones por lo que consumirá más tiempo y recursos pero por la contra tendremos menos calidad de imagen. Y al revés pasa con los Laplacianos, puesto que necesitan más operaciones y recursos pero tendrán un resultado final con mejor calidad. Después de explicar los tipos de operadores que hay, se hace un recorrido explicando los distintos tipos de algoritmos que hay en cada uno de los grupos. El otro gran grupo de tecnologías para el reconocimiento de objetos son los basados en puntos ASIFT, los cuales se basan en 6 parámetros de la imagen y la comparan con otra imagen teniendo en cuenta dichos parámetros. La desventaja de este método, para nuestros propósitos futuros, es que sólo es valido para un objeto en concreto. Por lo que si vamos a reconocer dos hojas diferentes, aunque sean de la misma especie, no vamos a poder reconocerlas mediante este método. Aún así es importante explicar este tipo de tecnologías puesto que estamos hablando de técnicas de reconocimiento en general. Al final del capítulo podremos ver una comparación con los pros y las contras de todas las tecnologías empleadas. Primeramente comparándolas de forma separada y, finalmente, compararemos todos los métodos existentes en base a nuestros propósitos. Las técnicas de reconocimiento, el siguiente apartado, no es muy extenso puesto que, aunque haya pasos generales para el reconocimiento de objetos, cada objeto a reconocer es distinto por lo que no hay un método específico que se pueda generalizar. Pasamos ahora a las técnicas de detección de hojas mediante ordenador. Aquí usaremos la técnica explicada previamente explicada basada en las derivadas de las imágenes. La continuación de este paso sería diseccionar la hoja en diversos parámetros. Dependiendo de la fuente a la que se consulte pueden haber más o menos parámetros. Unos documentos aconsejan dividir la morfología de la hoja en 3 parámetros principales (Forma, Dentina y ramificación] y derivando de dichos parámetros convertirlos a 16 parámetros secundarios. La otra propuesta es dividir la morfología de la hoja en 5 parámetros principales (Diámetro, longitud fisiológica, anchura fisiológica, área y perímetro] y de ahí extraer 12 parámetros secundarios. Esta segunda propuesta es la más utilizada de todas por lo que es la que se utilizará. Pasamos al reconocimiento de hojas, en la cual nos hemos basado en un documento que provee un código fuente que cucando en los dos extremos de la hoja automáticamente nos dice a qué especie pertenece la hoja que estamos intentando reconocer. Para ello sólo hay que formar una base de datos. En los test realizados por el citado documento, nos aseguran que tiene un índice de acierto del 90.312% en 320 test en total (32 plantas insertadas en la base de datos por 10 test que se han realizado por cada una de las especies]. El siguiente apartado trata de la detección de cromosomas, en el cual se debe de pasar de la célula metafásica, donde los cromosomas están desorganizados, al cariotipo, que es como solemos ver los 23 cromosomas de forma ordenada. Hay dos tipos de técnicas para realizar este paso: Por el proceso de esquelotonización y barriendo ángulos. El proceso de esqueletonización consiste en eliminar los píxeles del interior del cromosoma para quedarse con su silueta; Este proceso es similar a los métodos de derivación de los píxeles pero se diferencia en que no detecta bordes si no que detecta el interior de los cromosomas. La segunda técnica consiste en ir barriendo ángulos desde el principio del cromosoma y teniendo en cuenta que un cromosoma no puede doblarse más de X grados detecta las diversas regiones de los cromosomas. Una vez tengamos el cariotipo, se continua con el reconocimiento de cromosomas. Para ello existe una técnica basada en las bandas de blancos y negros que tienen los cromosomas y que son las que los hacen únicos. Para ello el programa detecta los ejes longitudinales del cromosoma y reconstruye los perfiles de las bandas que posee el cromosoma y que lo identifican como único. En cuanto al trabajo que se podría desempeñar en el futuro, tenemos por lo general dos técnicas independientes que no unen la detección con el reconocimiento por lo que se habría de preparar un programa que uniese estas dos técnicas. Respecto a las hojas hemos visto que ambos métodos, detección y reconocimiento, están vinculados debido a que ambos comparten la opinión de dividir las hojas en 5 parámetros principales. El trabajo que habría que realizar sería el de crear un algoritmo que conectase a ambos ya que en el programa de reconocimiento se debe clicar a los dos extremos de la hoja por lo que no es una tarea automática. En cuanto a los cromosomas, se debería de crear un algoritmo que busque el inicio del cromosoma y entonces empiece a barrer ángulos para después poder dárselo al programa que busca los perfiles de bandas de los cromosomas. Finalmente, en el resumen se explica el por qué hace falta este tipo de investigación, esto es que con el calentamiento global, muchas de las especies (tanto animales como plantas] se están empezando a extinguir. Es por ello que se necesitará una base de datos que contemple todas las posibles especies tanto del reino animal como del reino vegetal. Para reconocer a una especie animal, simplemente bastará con tener sus 23 cromosomas; mientras que para reconocer a una especie vegetal, existen diversas formas. Aunque la más sencilla de todas es contar con la hoja de la especie puesto que es el elemento más fácil de escanear e introducir en el ordenador.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Remote sensing information from spaceborne and airborne platforms continues to provide valuable data for different environmental monitoring applications. In this sense, high spatial resolution im-agery is an important source of information for land cover mapping. For the processing of high spa-tial resolution images, the object-based methodology is one of the most commonly used strategies. However, conventional pixel-based methods, which only use spectral information for land cover classification, are inadequate for classifying this type of images. This research presents a method-ology to characterise Mediterranean land covers in high resolution aerial images by means of an object-oriented approach. It uses a self-calibrating multi-band region growing approach optimised by pre-processing the image with a bilateral filtering. The obtained results show promise in terms of both segmentation quality and computational efficiency.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Probabilistic graphical models are a huge research field in artificial intelligence nowadays. The scope of this work is the study of directed graphical models for the representation of discrete distributions. Two of the main research topics related to this area focus on performing inference over graphical models and on learning graphical models from data. Traditionally, the inference process and the learning process have been treated separately, but given that the learned models structure marks the inference complexity, this kind of strategies will sometimes produce very inefficient models. With the purpose of learning thinner models, in this master thesis we propose a new model for the representation of network polynomials, which we call polynomial trees. Polynomial trees are a complementary representation for Bayesian networks that allows an efficient evaluation of the inference complexity and provides a framework for exact inference. We also propose a set of methods for the incremental compilation of polynomial trees and an algorithm for learning polynomial trees from data using a greedy score+search method that includes the inference complexity as a penalization in the scoring function.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the last decade, the research community has focused on new classification methods that rely on statistical characteristics of Internet traffic, instead of pre-viously popular port-number-based or payload-based methods, which are under even bigger constrictions. Some research works based on statistical characteristics generated large fea-ture sets of Internet traffic; however, nowadays it?s impossible to handle hun-dreds of features in big data scenarios, only leading to unacceptable processing time and misleading classification results due to redundant and correlative data. As a consequence, a feature selection procedure is essential in the process of Internet traffic characterization. In this paper a survey of feature selection methods is presented: feature selection frameworks are introduced, and differ-ent categories of methods are briefly explained and compared; several proposals on feature selection in Internet traffic characterization are shown; finally, future application of feature selection to a concrete project is proposed.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La edificación residencial existente en España y en Europa se encuentra abocada a una rehabilitación profunda para cumplir los objetivos marcados en la estrategia europea para el año 2050. Estos, para el sector de la edificación, se proponen una reducción del 90% de emisiones de gases de efecto invernadero (GEI) respecto a niveles del año 1990. Este plan a largo plazo establece hitos intermedios de control, con objetivos parciales para el año 2020 y 2030. El objetivo último es aprovechar el potencial de reducción de demanda energética del sector de la edificación, del cual la edificación residencial supone el 85% en España. Dentro de estos requerimientos, de reducción de demanda energética en la edificación, la ventilación en la edificación residencial se convierte en uno de los retos a resolver por su vinculación directa a la salud y el confort de los ocupantes de la misma, y al mismo tiempo su relación proporcional con la demanda energética que presenta el edificio asociada al acondicionamiento térmico. Gran parte de las pérdidas térmicas de la edificación residencial se producen por el aire de renovación y la infiltración de aire a través de la envolvente. La directiva europea de eficiencia energética de la edificación (EPBD), que establece las directrices necesarias para alcanzar los objetivos de este sector en cuanto a emisiones de CO2 y gases de efecto invernadero (GEI), contempla la ventilación con aire limpio como un requisito fundamental a tener en cuenta de cara a las nuevas construcciones y a la rehabilitación energética de los edificios existentes. El síndrome del edificio enfermo, un conjunto de molestias y síntomas asociados a la baja calidad del aire de edificios no residenciales que surgió a raíz de la crisis del petróleo de 1973, tuvo su origen en una ventilación deficiente y una renovación del aire interior insuficiente de estos edificios, producto del intento de ahorro en la factura energética. Teniendo en cuenta que, de media, pasamos un 58% de nuestro tiempo en las viviendas, es fundamental cuidar la calidad del aire interior y no empeorarla aplicando medidas de “eficiencia energética” con efectos no esperados. Para conseguir esto es fundamental conocer en profundidad cómo se produce la ventilación en la edificación en bloque en España en sus aspectos de calidad del aire interior y demanda energética asociada a la ventilación. El objetivo de esta tesis es establecer una metodología de caracterización y de optimización de las necesidades de ventilación para los espacios residenciales existentes en España que aúne el doble objetivo de garantizar la calidad ambiental y reducir la demanda energética de los mismos. La caracterización del parque edificatorio residencial español en cuanto a ventilación es concluyente: La vivienda en España se distribuye principalmente en tres periodos en los que se encuentran más del 80% del total de las viviendas construidas. El periodo anterior a las normas básicas de la edificación (NBE), de 1960 a 1980, el periodo desde 1980 al año 2005, con el mayor número total de viviendas construidas, guiado por la NTE ISV 75, y el periodo correspondiente a la edificación construida a partir del Código Técnico de la Edificación, en 2006, cuyo documento básico de condiciones de salubridad (DB HS3) es la primera norma de obligado cumplimiento en diseño y dimensionamiento de ventilación residencial en España. La selección de un modelo de bloque de viviendas de referencia, un valor medio y representativo, seleccionado de entre estos periodos, pero con cualidades que se extienden más allá de uno de ellos, nos permite realizar un intensivo análisis comparativo de las condiciones de calidad de aire interior y la demanda energética del mismo, aplicando las distintas configuraciones que presenta la ventilación en viviendas dependiendo del escenario o época constructiva (o normativa) en que esta fuera construida. Este análisis se lleva a cabo apoyándose en un doble enfoque: el modelado numérico de simulaciones y el análisis de datos experimentales, para comprobar y afinar los modelos y observar la situación real de las viviendas en estos dos aspectos. Gracias a las conclusiones del análisis previo, se define una estrategia de optimización de la ventilación basada fundamentalmente en dos medidas: 1) La introducción de un sistema de extracción mecánica y recuperación de calor que permita reducir la demanda energética debida a la renovación del aire y a la vez diluir los contaminantes interiores más eficazmente para mejorar, de esta forma, la calidad del ambiente interior. 2) La racionalización del horario de utilización de estos sistemas, no malgastando la energía en periodos de no ocupación, permitiendo una leve ventilación de fondo, debida a la infiltración, que no incida en pérdidas energéticas cuantiosas. A esta optimización, además de aplicar la metodología de análisis previo, en cuanto a demanda energética y calidad del aire, se aplica una valoración económica integradora y comparativa basada en el reglamento delegado EU244/2012 de coste óptimo (Cost Optimal Methodology). Los resultados principales de esta tesis son: • Un diagnóstico de la calidad del aire interior de la edificación residencial en España y su demanda energética asociada, imprescindible para lograr una rehabilitación energética profunda garantizando la calidad del aire interior. • Un indicador de la relación directa entre calidad de aire y demanda energética, para evaluar la adecuación de los sistemas de ventilación, respecto de las nuevas normativas de eficiencia energética y ventilación. • Una estrategia de optimización, que ofrece una alternativa de intervención, y la aplicación de un método de valoración que permite evaluar la amortización comparada de la instalación de los sistemas. ABSTRACT The housing building stock already built in Spain and Europe faces a deep renovation in the present and near future to accomplish with the objectives agreed in the European strategy for 2050. These objectives, for the building sector, are set in a 90% of Green House Gases (GHG) reduction compared to levels in 1990. This long‐term plan has set milestones to control the correct advance of achievement in 2020 and 2030. The main objective is to take advantage of the great potential to reduce energy demand from the building sector, in which housing represents 85% share in Spain. Among this reduction on building energy demand requirements, ventilation of dwellings becomes one of the challenges to solve as it’s directly connected to the indoor air quality (IAQ) and comfort conditions for the users, as well as proportional to the building energy demand on thermal conditioning. A big share of thermal losses in housing is caused by air renovation and infiltration through the envelope leaks. The European Directive on Building energy performance (EPBD), establishes the roots needed to reach the building sector objectives in terms of CO2 and GHG emissions. This directive sets the ventilation and renovation with clean air of the new and existing buildings as a fundamental requirement. The Sick Building Syndrome (SBS), an aggregation of symptoms and annoys associated to low air quality in non residential buildings, appeared as common after the 1973 oil crisis. It is originated in defective ventilation systems and deficient air renovation rates, as a consequence of trying to lower the energy bill. Accounting that we spend 58% of our time in dwellings, it becomes crucial to look after the indoor air quality and focus in not worsening it by applying “energy efficient” measures, with not expected side effects. To do so, it is primary to research in deep how the ventilation takes place in the housing blocks in Spain, in the aspects related to IAQ and ventilation energy demand. This thesis main objective is to establish a characterization and optimization methodology regarding the ventilation needs for existing housing in Spain, considering the twofold objective of guaranteeing the air quality as reducing the energy demand. The characterization of the existing housing building stock in Spain regarding ventilation is conclusive. More of 80% of the housing stock is distributed in 3 main periods: before the implementation of the firsts regulations on building comfort conditions (Normas Básicas de la Edificación), from 1960 to 1980; the period after the first recommendations on ventilation (NTE ISV 75) for housing were set, around 1980 until 2005 and; the period corresponding to the housing built after the existing mandatory regulation in terms of indoor sanity conditions and ventilation (Spanish Building Code, DB HS3) was set, in 2006. Selecting a representative blueprint of a housing block in Spain, which has medium characteristics not just within the 3 periods mention, but which qualities extent beyond the 3 of them, allows the next step, analyzing. This comparative and intense analyzing phase is focused on the air indoor conditions and the related energy demand, applying different configurations to the ventilation systems according to the different constructive or regulation period in which the building is built. This analysis is also twofold: 1) Numerical modeling with computer simulations and 2) experimental data collection from existing housing in real conditions to check and refine the models to be tested. Thanks to the analyzing phase conclusions, an optimization strategy on the ventilation of the housing stock is set, based on two actions to take: 1) To introduce a mechanical exhaust and intake ventilation system with heat recovery that allows reducing energy demand, as improves the capacity of the system to dilute the pollutant load. This way, the environmental quality is improved. 2) To optimize the schedule of the system use, avoids waste of energy in no occupancy periods, relying ventilation during this time in a light infiltration ventilation, intended not to become large and not causing extra energy losses. Apart from applying the previous analyzing methodology to the optimization strategy, regarding energy demand and air quality, a ROI valorization is performed, based on the cost optimal methodology (delegated regulation EU244/2012). The main results from the thesis are: • To obtain a through diagnose regarding air quality and energy demand for the existing housing stock in Spain, unavoidable to reach a energy deep retrofitting scheme with no air quality worsening. • To obtain a marker to relate air quality and energy demand and evaluate adequateness of ventilation systems, for the new regulations to come. • To establish an optimization strategy to improve both air quality and energy demand, applying a compared valorization methodology to obtain the Return On Investment (ROI).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Gene expression profiling provides powerful analyses of transcriptional responses to cellular perturbation. In contrast to DNA array-based methods, reporter gene technology has been underused for this application. Here we describe a genomewide, genome-registered collection of Escherichia coli bioluminescent reporter gene fusions. DNA sequences from plasmid-borne, random fusions of E. coli chromosomal DNA to a Photorhabdus luminescens luxCDABE reporter allowed precise mapping of each fusion. The utility of this collection covering about 30% of the transcriptional units was tested by analyzing individual fusions representative of heat shock, SOS, OxyR, SoxRS, and cya/crp stress-responsive regulons. Each fusion strain responded as anticipated to environmental conditions known to activate the corresponding regulatory circuit. Thus, the collection mirrors E. coli's transcriptional wiring diagram. This genomewide collection of gene fusions provides an independent test of results from other gene expression analyses. Accordingly, a DNA microarray-based analysis of mitomycin C-treated E. coli indicated elevated expression of expected and unanticipated genes. Selected luxCDABE fusions corresponding to these up-regulated genes were used to confirm or contradict the DNA microarray results. The power of partnering gene fusion and DNA microarray technology to discover promoters and define operons was demonstrated when data from both suggested that a cluster of 20 genes encoding production of type I extracellular polysaccharide in E. coli form a single operon.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Phyllosphere microbial communities were evaluated on leaves of field-grown plant species by culture-dependent and -independent methods. Denaturing gradient gel electrophoresis (DGGE) with 16S rDNA primers generally indicated that microbial community structures were similar on different individuals of the same plant species, but unique on different plant species. Phyllosphere bacteria were identified from Citrus sinesis (cv. Valencia) by using DGGE analysis followed by cloning and sequencing of the dominant rDNA bands. Of the 17 unique sequences obtained, database queries showed only four strains that had been described previously as phyllosphere bacteria. Five of the 17 sequences had 16S similarities lower than 90% to database entries, suggesting that they represent previously undescribed species. In addition, three fungal species were also identified. Very different 16S rDNA DGGE banding profiles were obtained when replicate cv. Valencia leaf samples were cultured in BIOLOG EcoPlates for 4.5 days. All of these rDNA sequences had 97–100% similarity to those of known phyllosphere bacteria, but only two of them matched those identified by the culture independent DGGE analysis. Like other studied ecosystems, microbial phyllosphere communities therefore are more complex than previously thought, based on conventional culture-based methods.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The field of natural language processing (NLP) has seen a dramatic shift in both research direction and methodology in the past several years. In the past, most work in computational linguistics tended to focus on purely symbolic methods. Recently, more and more work is shifting toward hybrid methods that combine new empirical corpus-based methods, including the use of probabilistic and information-theoretic techniques, with traditional symbolic methods. This work is made possible by the recent availability of linguistic databases that add rich linguistic annotation to corpora of natural language text. Already, these methods have led to a dramatic improvement in the performance of a variety of NLP systems with similar improvement likely in the coming years. This paper focuses on these trends, surveying in particular three areas of recent progress: part-of-speech tagging, stochastic parsing, and lexical semantics.