966 resultados para Crowd density estimation


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Neuronal morphology is a key feature in the study of brain circuits, as it is highly related to information processing and functional identification. Neuronal morphology affects the process of integration of inputs from other neurons and determines the neurons which receive the output of the neurons. Different parts of the neurons can operate semi-independently according to the spatial location of the synaptic connections. As a result, there is considerable interest in the analysis of the microanatomy of nervous cells since it constitutes an excellent tool for better understanding cortical function. However, the morphologies, molecular features and electrophysiological properties of neuronal cells are extremely variable. Except for some special cases, this variability makes it hard to find a set of features that unambiguously define a neuronal type. In addition, there are distinct types of neurons in particular regions of the brain. This morphological variability makes the analysis and modeling of neuronal morphology a challenge. Uncertainty is a key feature in many complex real-world problems. Probability theory provides a framework for modeling and reasoning with uncertainty. Probabilistic graphical models combine statistical theory and graph theory to provide a tool for managing domains with uncertainty. In particular, we focus on Bayesian networks, the most commonly used probabilistic graphical model. In this dissertation, we design new methods for learning Bayesian networks and apply them to the problem of modeling and analyzing morphological data from neurons. The morphology of a neuron can be quantified using a number of measurements, e.g., the length of the dendrites and the axon, the number of bifurcations, the direction of the dendrites and the axon, etc. These measurements can be modeled as discrete or continuous data. The continuous data can be linear (e.g., the length or the width of a dendrite) or directional (e.g., the direction of the axon). These data may follow complex probability distributions and may not fit any known parametric distribution. Modeling this kind of problems using hybrid Bayesian networks with discrete, linear and directional variables poses a number of challenges regarding learning from data, inference, etc. In this dissertation, we propose a method for modeling and simulating basal dendritic trees from pyramidal neurons using Bayesian networks to capture the interactions between the variables in the problem domain. A complete set of variables is measured from the dendrites, and a learning algorithm is applied to find the structure and estimate the parameters of the probability distributions included in the Bayesian networks. Then, a simulation algorithm is used to build the virtual dendrites by sampling values from the Bayesian networks, and a thorough evaluation is performed to show the model’s ability to generate realistic dendrites. In this first approach, the variables are discretized so that discrete Bayesian networks can be learned and simulated. Then, we address the problem of learning hybrid Bayesian networks with different kinds of variables. Mixtures of polynomials have been proposed as a way of representing probability densities in hybrid Bayesian networks. We present a method for learning mixtures of polynomials approximations of one-dimensional, multidimensional and conditional probability densities from data. The method is based on basis spline interpolation, where a density is approximated as a linear combination of basis splines. The proposed algorithms are evaluated using artificial datasets. We also use the proposed methods as a non-parametric density estimation technique in Bayesian network classifiers. Next, we address the problem of including directional data in Bayesian networks. These data have some special properties that rule out the use of classical statistics. Therefore, different distributions and statistics, such as the univariate von Mises and the multivariate von Mises–Fisher distributions, should be used to deal with this kind of information. In particular, we extend the naive Bayes classifier to the case where the conditional probability distributions of the predictive variables given the class follow either of these distributions. We consider the simple scenario, where only directional predictive variables are used, and the hybrid case, where discrete, Gaussian and directional distributions are mixed. The classifier decision functions and their decision surfaces are studied at length. Artificial examples are used to illustrate the behavior of the classifiers. The proposed classifiers are empirically evaluated over real datasets. We also study the problem of interneuron classification. An extensive group of experts is asked to classify a set of neurons according to their most prominent anatomical features. A web application is developed to retrieve the experts’ classifications. We compute agreement measures to analyze the consensus between the experts when classifying the neurons. Using Bayesian networks and clustering algorithms on the resulting data, we investigate the suitability of the anatomical terms and neuron types commonly used in the literature. Additionally, we apply supervised learning approaches to automatically classify interneurons using the values of their morphological measurements. Then, a methodology for building a model which captures the opinions of all the experts is presented. First, one Bayesian network is learned for each expert, and we propose an algorithm for clustering Bayesian networks corresponding to experts with similar behaviors. Then, a Bayesian network which represents the opinions of each group of experts is induced. Finally, a consensus Bayesian multinet which models the opinions of the whole group of experts is built. A thorough analysis of the consensus model identifies different behaviors between the experts when classifying the interneurons in the experiment. A set of characterizing morphological traits for the neuronal types can be defined by performing inference in the Bayesian multinet. These findings are used to validate the model and to gain some insights into neuron morphology. Finally, we study a classification problem where the true class label of the training instances is not known. Instead, a set of class labels is available for each instance. This is inspired by the neuron classification problem, where a group of experts is asked to individually provide a class label for each instance. We propose a novel approach for learning Bayesian networks using count vectors which represent the number of experts who selected each class label for each instance. These Bayesian networks are evaluated using artificial datasets from supervised learning problems. Resumen La morfología neuronal es una característica clave en el estudio de los circuitos cerebrales, ya que está altamente relacionada con el procesado de información y con los roles funcionales. La morfología neuronal afecta al proceso de integración de las señales de entrada y determina las neuronas que reciben las salidas de otras neuronas. Las diferentes partes de la neurona pueden operar de forma semi-independiente de acuerdo a la localización espacial de las conexiones sinápticas. Por tanto, existe un interés considerable en el análisis de la microanatomía de las células nerviosas, ya que constituye una excelente herramienta para comprender mejor el funcionamiento de la corteza cerebral. Sin embargo, las propiedades morfológicas, moleculares y electrofisiológicas de las células neuronales son extremadamente variables. Excepto en algunos casos especiales, esta variabilidad morfológica dificulta la definición de un conjunto de características que distingan claramente un tipo neuronal. Además, existen diferentes tipos de neuronas en regiones particulares del cerebro. La variabilidad neuronal hace que el análisis y el modelado de la morfología neuronal sean un importante reto científico. La incertidumbre es una propiedad clave en muchos problemas reales. La teoría de la probabilidad proporciona un marco para modelar y razonar bajo incertidumbre. Los modelos gráficos probabilísticos combinan la teoría estadística y la teoría de grafos con el objetivo de proporcionar una herramienta con la que trabajar bajo incertidumbre. En particular, nos centraremos en las redes bayesianas, el modelo más utilizado dentro de los modelos gráficos probabilísticos. En esta tesis hemos diseñado nuevos métodos para aprender redes bayesianas, inspirados por y aplicados al problema del modelado y análisis de datos morfológicos de neuronas. La morfología de una neurona puede ser cuantificada usando una serie de medidas, por ejemplo, la longitud de las dendritas y el axón, el número de bifurcaciones, la dirección de las dendritas y el axón, etc. Estas medidas pueden ser modeladas como datos continuos o discretos. A su vez, los datos continuos pueden ser lineales (por ejemplo, la longitud o la anchura de una dendrita) o direccionales (por ejemplo, la dirección del axón). Estos datos pueden llegar a seguir distribuciones de probabilidad muy complejas y pueden no ajustarse a ninguna distribución paramétrica conocida. El modelado de este tipo de problemas con redes bayesianas híbridas incluyendo variables discretas, lineales y direccionales presenta una serie de retos en relación al aprendizaje a partir de datos, la inferencia, etc. En esta tesis se propone un método para modelar y simular árboles dendríticos basales de neuronas piramidales usando redes bayesianas para capturar las interacciones entre las variables del problema. Para ello, se mide un amplio conjunto de variables de las dendritas y se aplica un algoritmo de aprendizaje con el que se aprende la estructura y se estiman los parámetros de las distribuciones de probabilidad que constituyen las redes bayesianas. Después, se usa un algoritmo de simulación para construir dendritas virtuales mediante el muestreo de valores de las redes bayesianas. Finalmente, se lleva a cabo una profunda evaluaci ón para verificar la capacidad del modelo a la hora de generar dendritas realistas. En esta primera aproximación, las variables fueron discretizadas para poder aprender y muestrear las redes bayesianas. A continuación, se aborda el problema del aprendizaje de redes bayesianas con diferentes tipos de variables. Las mixturas de polinomios constituyen un método para representar densidades de probabilidad en redes bayesianas híbridas. Presentamos un método para aprender aproximaciones de densidades unidimensionales, multidimensionales y condicionales a partir de datos utilizando mixturas de polinomios. El método se basa en interpolación con splines, que aproxima una densidad como una combinación lineal de splines. Los algoritmos propuestos se evalúan utilizando bases de datos artificiales. Además, las mixturas de polinomios son utilizadas como un método no paramétrico de estimación de densidades para clasificadores basados en redes bayesianas. Después, se estudia el problema de incluir información direccional en redes bayesianas. Este tipo de datos presenta una serie de características especiales que impiden el uso de las técnicas estadísticas clásicas. Por ello, para manejar este tipo de información se deben usar estadísticos y distribuciones de probabilidad específicos, como la distribución univariante von Mises y la distribución multivariante von Mises–Fisher. En concreto, en esta tesis extendemos el clasificador naive Bayes al caso en el que las distribuciones de probabilidad condicionada de las variables predictoras dada la clase siguen alguna de estas distribuciones. Se estudia el caso base, en el que sólo se utilizan variables direccionales, y el caso híbrido, en el que variables discretas, lineales y direccionales aparecen mezcladas. También se estudian los clasificadores desde un punto de vista teórico, derivando sus funciones de decisión y las superficies de decisión asociadas. El comportamiento de los clasificadores se ilustra utilizando bases de datos artificiales. Además, los clasificadores son evaluados empíricamente utilizando bases de datos reales. También se estudia el problema de la clasificación de interneuronas. Desarrollamos una aplicación web que permite a un grupo de expertos clasificar un conjunto de neuronas de acuerdo a sus características morfológicas más destacadas. Se utilizan medidas de concordancia para analizar el consenso entre los expertos a la hora de clasificar las neuronas. Se investiga la idoneidad de los términos anatómicos y de los tipos neuronales utilizados frecuentemente en la literatura a través del análisis de redes bayesianas y la aplicación de algoritmos de clustering. Además, se aplican técnicas de aprendizaje supervisado con el objetivo de clasificar de forma automática las interneuronas a partir de sus valores morfológicos. A continuación, se presenta una metodología para construir un modelo que captura las opiniones de todos los expertos. Primero, se genera una red bayesiana para cada experto y se propone un algoritmo para agrupar las redes bayesianas que se corresponden con expertos con comportamientos similares. Después, se induce una red bayesiana que modela la opinión de cada grupo de expertos. Por último, se construye una multired bayesiana que modela las opiniones del conjunto completo de expertos. El análisis del modelo consensuado permite identificar diferentes comportamientos entre los expertos a la hora de clasificar las neuronas. Además, permite extraer un conjunto de características morfológicas relevantes para cada uno de los tipos neuronales mediante inferencia con la multired bayesiana. Estos descubrimientos se utilizan para validar el modelo y constituyen información relevante acerca de la morfología neuronal. Por último, se estudia un problema de clasificación en el que la etiqueta de clase de los datos de entrenamiento es incierta. En cambio, disponemos de un conjunto de etiquetas para cada instancia. Este problema está inspirado en el problema de la clasificación de neuronas, en el que un grupo de expertos proporciona una etiqueta de clase para cada instancia de manera individual. Se propone un método para aprender redes bayesianas utilizando vectores de cuentas, que representan el número de expertos que seleccionan cada etiqueta de clase para cada instancia. Estas redes bayesianas se evalúan utilizando bases de datos artificiales de problemas de aprendizaje supervisado.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

La evolución de los teléfonos móviles inteligentes, dotados de cámaras digitales, está provocando una creciente demanda de aplicaciones cada vez más complejas que necesitan algoritmos de visión artificial en tiempo real; puesto que el tamaño de las señales de vídeo no hace sino aumentar y en cambio el rendimiento de los procesadores de un solo núcleo se ha estancado, los nuevos algoritmos que se diseñen para visión artificial han de ser paralelos para poder ejecutarse en múltiples procesadores y ser computacionalmente escalables. Una de las clases de procesadores más interesantes en la actualidad se encuentra en las tarjetas gráficas (GPU), que son dispositivos que ofrecen un alto grado de paralelismo, un excelente rendimiento numérico y una creciente versatilidad, lo que los hace interesantes para llevar a cabo computación científica. En esta tesis se exploran dos aplicaciones de visión artificial que revisten una gran complejidad computacional y no pueden ser ejecutadas en tiempo real empleando procesadores tradicionales. En cambio, como se demuestra en esta tesis, la paralelización de las distintas subtareas y su implementación sobre una GPU arrojan los resultados deseados de ejecución con tasas de refresco interactivas. Asimismo, se propone una técnica para la evaluación rápida de funciones de complejidad arbitraria especialmente indicada para su uso en una GPU. En primer lugar se estudia la aplicación de técnicas de síntesis de imágenes virtuales a partir de únicamente dos cámaras lejanas y no paralelas—en contraste con la configuración habitual en TV 3D de cámaras cercanas y paralelas—con información de color y profundidad. Empleando filtros de mediana modificados para la elaboración de un mapa de profundidad virtual y proyecciones inversas, se comprueba que estas técnicas son adecuadas para una libre elección del punto de vista. Además, se demuestra que la codificación de la información de profundidad con respecto a un sistema de referencia global es sumamente perjudicial y debería ser evitada. Por otro lado se propone un sistema de detección de objetos móviles basado en técnicas de estimación de densidad con funciones locales. Este tipo de técnicas es muy adecuada para el modelado de escenas complejas con fondos multimodales, pero ha recibido poco uso debido a su gran complejidad computacional. El sistema propuesto, implementado en tiempo real sobre una GPU, incluye propuestas para la estimación dinámica de los anchos de banda de las funciones locales, actualización selectiva del modelo de fondo, actualización de la posición de las muestras de referencia del modelo de primer plano empleando un filtro de partículas multirregión y selección automática de regiones de interés para reducir el coste computacional. Los resultados, evaluados sobre diversas bases de datos y comparados con otros algoritmos del estado del arte, demuestran la gran versatilidad y calidad de la propuesta. Finalmente se propone un método para la aproximación de funciones arbitrarias empleando funciones continuas lineales a tramos, especialmente indicada para su implementación en una GPU mediante el uso de las unidades de filtraje de texturas, normalmente no utilizadas para cómputo numérico. La propuesta incluye un riguroso análisis matemático del error cometido en la aproximación en función del número de muestras empleadas, así como un método para la obtención de una partición cuasióptima del dominio de la función para minimizar el error. ABSTRACT The evolution of smartphones, all equipped with digital cameras, is driving a growing demand for ever more complex applications that need to rely on real-time computer vision algorithms. However, video signals are only increasing in size, whereas the performance of single-core processors has somewhat stagnated in the past few years. Consequently, new computer vision algorithms will need to be parallel to run on multiple processors and be computationally scalable. One of the most promising classes of processors nowadays can be found in graphics processing units (GPU). These are devices offering a high parallelism degree, excellent numerical performance and increasing versatility, which makes them interesting to run scientific computations. In this thesis, we explore two computer vision applications with a high computational complexity that precludes them from running in real time on traditional uniprocessors. However, we show that by parallelizing subtasks and implementing them on a GPU, both applications attain their goals of running at interactive frame rates. In addition, we propose a technique for fast evaluation of arbitrarily complex functions, specially designed for GPU implementation. First, we explore the application of depth-image–based rendering techniques to the unusual configuration of two convergent, wide baseline cameras, in contrast to the usual configuration used in 3D TV, which are narrow baseline, parallel cameras. By using a backward mapping approach with a depth inpainting scheme based on median filters, we show that these techniques are adequate for free viewpoint video applications. In addition, we show that referring depth information to a global reference system is ill-advised and should be avoided. Then, we propose a background subtraction system based on kernel density estimation techniques. These techniques are very adequate for modelling complex scenes featuring multimodal backgrounds, but have not been so popular due to their huge computational and memory complexity. The proposed system, implemented in real time on a GPU, features novel proposals for dynamic kernel bandwidth estimation for the background model, selective update of the background model, update of the position of reference samples of the foreground model using a multi-region particle filter, and automatic selection of regions of interest to reduce computational cost. The results, evaluated on several databases and compared to other state-of-the-art algorithms, demonstrate the high quality and versatility of our proposal. Finally, we propose a general method for the approximation of arbitrarily complex functions using continuous piecewise linear functions, specially formulated for GPU implementation by leveraging their texture filtering units, normally unused for numerical computation. Our proposal features a rigorous mathematical analysis of the approximation error in function of the number of samples, as well as a method to obtain a suboptimal partition of the domain of the function to minimize approximation error.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Rock mass characterization requires a deep geometric understanding of the discontinuity sets affecting rock exposures. Recent advances in Light Detection and Ranging (LiDAR) instrumentation currently allow quick and accurate 3D data acquisition, yielding on the development of new methodologies for the automatic characterization of rock mass discontinuities. This paper presents a methodology for the identification and analysis of flat surfaces outcropping in a rocky slope using the 3D data obtained with LiDAR. This method identifies and defines the algebraic equations of the different planes of the rock slope surface by applying an analysis based on a neighbouring points coplanarity test, finding principal orientations by Kernel Density Estimation and identifying clusters by the Density-Based Scan Algorithm with Noise. Different sources of information —synthetic and 3D scanned data— were employed, performing a complete sensitivity analysis of the parameters in order to identify the optimal value of the variables of the proposed method. In addition, raw source files and obtained results are freely provided in order to allow to a more straightforward method comparison aiming to a more reproducible research.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper examines the determinants of foreign direct investment (FDI) under free trade agreements (FTAs) from a new institutional perspective. First, the determinants of FDI are theoretically discussed from a new institutional perspective. Then, FDI is statistically analyzed at the aggregate level. Kernel density estimation of firm-size reveals some evidence of "structural changes" after FTAs, as characterized by the investing firms' paid-up capital stock. Statistical tests of the average and variance of the size distribution confirm this in the case of FTAs with Asian partner countries. For FTAs with South American partner countries, the presence of FTAs seems to promote larger-scale FDIs. These results remain correlational instead of causal, and more statistical analyses would be needed to infer causality. Policy implications suggest that participants should consider "institutional" aspects of FTAs, that is, the size matters as a determinant of FDI. Future work along this line is needed to study "firm heterogeneity."

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Acoustic estimates of herring and blue whiting abundance were obtained during the surveys using the Simrad ER60 scientific echosounder. The allocation of NASC-values to herring, blue whiting and other acoustic targets were based on the composition of the trawl catches and the appearance of echo recordings. To estimate the abundance, the allocated NASC -values were averaged for ICES-squares (0.5° latitude by 1° longitude). For each statistical square, the unit area density of fish (rA) in number per square nautical mile (N*nm-2) was calculated using standard equations (Foote et al., 1987; Toresen et al., 1998). To estimate the total abundance of fish, the unit area abundance for each statistical square was multiplied by the number of square nautical miles in each statistical square and then summed for all the statistical squares within defined subareas and over the total area. Biomass estimation was calculated by multiplying abundance in numbers by the average weight of the fish in each statistical square then summing all squares within defined subareas and over the total area. The Norwegian BEAM soft-ware (Totland and Godø 2001) was used to make estimates of total biomass.