348 resultados para STREAMING


Relevância:

10.00% 10.00%

Publicador:

Resumo:

This is a news video produced by Journalism student Jackie Marsiglia. This is to test streaming functionality in the UConn Digital Commons. The RealPlayer movies have been shortened.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A primary objective of Leg 175 was to investigate the upwelling history of the Benguela Current. Upwelling along the coast is found over the shelf in several well-established cells, as well as along the shelf-slope break, and extends over the 1000-m isobath. Streaming filaments along the coast also carry upwelled water off shore (Shannon, 1985). The upwelled nutrient-rich waters are sourced from the South Atlantic central water mass, which is a mixture of subtropical and subantarctic water masses. Below the central water mass lies Antarctic intermediate water (Shannon and Hunter, 1988, doi:10.2989/025776188784480735; Stramma and Peterson, 1989, doi:10.1175/1520-0485(1989)019<1440:GTITBC>2.0.CO;2). The upwelling system supports a robust marine community (Shannon and Pillar, 1986) where radiolarians are abundant (Bishop et al., 1978, doi:10.1016/0146-6291(78)90010-3). The endemic nature of radiolarians makes them useful in reconstructing the paleocirculation patterns. The biogeographic distribution of many species is limited by water-mass distribution. In a given geographic region, species may also have discrete depth habitats. However, their depth of occurrence can change worldwide because the depths of water masses vary with latitude (Boltovskoy, 1999). Consequently, species found at shallow depths at high latitudes (cold-water fauna) are observed deeper in the water column at lower latitudes. The low-latitude submergence of cold-water species broadens their distribution, resulting in species distributions that can cover multiple geographic regions (Kling, 1976, doi:10.1016/0011-7471(76)90880-9; Casey, doi:10.1016/0031-0182(89)90017-5; 1971; Boltovskoy, 1987, doi:10.1016/0377-8398(87)90014-4). Since radiolarian distribution is closely related to water-mass distribution and controlled by climatic conditions rather than geographic regions, similar assemblages characterize the equatorial, subtropical, transition, subpolar, and polar regions of ocean basins (Petrushevskaya, 1971a; Casey, 1989, doi:10.1016/0031-0182(89)90017-5; Boltovskoy, 1999). Numerous radiolarian species found in water masses in the Angola and Benguela Current systems have also been observed in plankton net samples, sediment traps, and surface-sediment studies in the Atlantic sector of the Southern Ocean, where they exhibited particular water-mass affinities (Abelmann, 1992a, doi:10.1007/BF00243107; Abelmann 1992b, doi:10.1007/BF00243108; Abelmann and Gowing, 1997, doi:10.1016/S0377-8398(96)00021-7). This report presents data on the radiolarian fauna recovered from Site 1082 sediments in the form of a survey of species reflecting the latitudinal migration of the Angola-Benguela Front and upwelling. The data constitute a time series of relative radiolarian abundances at very high resolution (every 20 cm) of the upper 12 m of Hole 1082A.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Fast-flowing ice streams discharge most of the ice from the interior of the Antarctic Ice Sheet coastward. Understanding how their tributary organisation is governed and evolves is essential for developing reliable models of the ice sheet's response to climate change. Despite much research on ice-stream mechanics, this problem is unsolved, because the complexity of flow within and across the tributary networks has hardly been interrogated. Here I present the first map of planimetric flow convergence across the ice sheet, calculated from satellite measurements of ice surface velocity, and use it to explore this complexity. The convergence map of Antarctica elucidates how ice-stream tributaries draw ice from the interior. It also reveals curvilinear zones of convergence along lateral shear margins of streaming, and abundant convergence ripples associated with nonlinear ice rheology and changes in bed topography and friction. Flow convergence on ice-stream tributaries and their feeding zones is markedly uneven, and interspersed with divergence at distances of the order of kilometres. For individual drainage basins as well as the ice sheet as a whole, the range of convergence and divergence decreases systematically with flow speed, implying that fast flow cannot converge or diverge as much as slow flow. I therefore deduce that flow in ice-stream networks is subject to mechanical regulation that limits flow-orthonormal strain rates. These properties and the gridded data of convergence and flow-orthonormal strain rate in this archive provide targets for ice- sheet simulations and motivate more research into the origin and dynamics of tributarization.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Fluxes of organic carbon normalised to a depth of 1000 m from 18 sites in the Atlantic and the Southern Ocean are presented, comprising nine biogeochemical provinces as defined by Longhurst et al. (1995. Journal of Plankton Research 17, 1245-1271). For comparison with primary production, we used a recent compilation of primary production values derived from CZCS data (Antoine et al., 1996. Global Biogeochemical Cycles 10, 57-69). In most cases, the seasonal patterns stood reasonably well in accordance with the carbon fluxes. Particularly, organic carbon flux records from two coastal sites off northwest and southwest Africa displayed a more distinct correlation to the primary production in sectors (1 x 1°) which are situated closer to the coastal environments. This was primarily caused by large upwelling filaments streaming far offshore, resulting in a cross-shelf carbon transport. With respect to primary production, organic carbon export to a water depth of 1000 m, and the fraction of primary production exported to a depth of 1000 m (export fraction=EF1000), we were able to distinguish between: (1) the coastal environments with highest values (EF1000=1.75-2.0%), (2) the eastern equatorial upwelling area with moderately high values (EF1000=0.8-1.1%), (3) and the subtropical oligotrophic gyres that yielded lowest values (EF1000=0.6%). Carbon export in the Southern Ocean was low to moderate, and the EF1000 value seems to be quite low in general. Annual organic carbon fluxes were proportional to primary production, and the export fraction EF1000 increased with primary production up to 350 gCm**-2 yr**-1. Latitudinal variations in primary production were reflected in the carbon flux pattern. A high temporal variability of primary production rates and a pronounced seasonality of carbon export were observed in the polar environments, in particular in coastal domains, although primary production (according to Antoine et al., 1996. Global Biogeochemical Cycles 10, 57-69), carbon fluxes, and the export fraction remained at low.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In Video over IP services, perceived video quality heavily depends on parameters such as video coding and network Quality of Service. This paper proposes a model for the estimation of perceived video quality in video streaming and broadcasting services that combines the aforementioned parameters with other that depend mainly on the information contents of the video sequences. These fitting parameters are derived from the Spatial and Temporal Information contents of the sequences. This model does not require reference to the original video sequence so it can be used for online, real-time monitoring of perceived video quality in Video over IP services. Furthermore, this paper proposes a measurement workbench designed to acquire both training data for model fitting and test data for model validation. Preliminary results show good correlation between measured and predicted values.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper introduces the experience of using videoconferencing and recording as a mechanism to support courses which need to be promoted or discontinued within the framework of the European convergence process. Our objective is to make these courses accessible as live streaming during the lessons as well as recorded lectures and associated documents available to the students as soon as the lesson has finished. The technology used has been developed in our university and it is all open source. Although this is a technical project the key is the human factor involved. The people managing the virtual sessions are students of the courses being recorded. However, they lack technical knowledge, so we had to train them in audiovisuals and enhance the usability of the videoconferencing tool and platform. The validation process is being carried out in five real scenarios at our university. During the whole period we are evaluating technical and pedagogical issues of this experience for both students and teachers to guide the future development of the service. Depending on the final results, the service of lectures recording will be available as educational resource for all of the teaching staff of our university.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This article proposes a MAS architecture for network diagnosis under uncertainty. Network diagnosis is divided into two inference processes: hypothesis generation and hypothesis confirmation. The first process is distributed among several agents based on a MSBN, while the second one is carried out by agents using semantic reasoning. A diagnosis ontology has been defined in order to combine both inference processes. To drive the deliberation process, dynamic data about the influence of observations are taken during diagnosis process. In order to achieve quick and reliable diagnoses, this influence is used to choose the best action to perform. This approach has been evaluated in a P2P video streaming scenario. Computational and time improvements are highlight as conclusions.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La Internet de las Cosas (IoT), como parte de la Futura Internet, se ha convertido en la actualidad en uno de los principales temas de investigación; en parte gracias a la atención que la sociedad está poniendo en el desarrollo de determinado tipo de servicios (telemetría, generación inteligente de energía, telesanidad, etc.) y por las recientes previsiones económicas que sitúan a algunos actores, como los operadores de telecomunicaciones (que se encuentran desesperadamente buscando nuevas oportunidades), al frente empujando algunas tecnologías interrelacionadas como las comunicaciones Máquina a Máquina (M2M). En este contexto, un importante número de actividades de investigación a nivel mundial se están realizando en distintas facetas: comunicaciones de redes de sensores, procesado de información, almacenamiento de grandes cantidades de datos (big--‐data), semántica, arquitecturas de servicio, etc. Todas ellas, de forma independiente, están llegando a un nivel de madurez que permiten vislumbrar la realización de la Internet de las Cosas más que como un sueño, como una realidad tangible. Sin embargo, los servicios anteriormente mencionados no pueden esperar a desarrollarse hasta que las actividades de investigación obtengan soluciones holísticas completas. Es importante proporcionar resultados intermedios que eviten soluciones verticales realizadas para desarrollos particulares. En este trabajo, nos hemos focalizado en la creación de una plataforma de servicios que pretende facilitar, por una parte la integración de redes de sensores y actuadores heterogéneas y geográficamente distribuidas, y por otra lado el desarrollo de servicios horizontales utilizando dichas redes y la información que proporcionan. Este habilitador se utilizará para el desarrollo de servicios y para la experimentación en la Internet de las Cosas. Previo a la definición de la plataforma, se ha realizado un importante estudio focalizando no sólo trabajos y proyectos de investigación, sino también actividades de estandarización. Los resultados se pueden resumir en las siguientes aseveraciones: a) Los modelos de datos definidos por el grupo “Sensor Web Enablement” (SWE™) del “Open Geospatial Consortium (OGC®)” representan hoy en día la solución más completa para describir las redes de sensores y actuadores así como las observaciones. b) Las interfaces OGC, a pesar de las limitaciones que requieren cambios y extensiones, podrían ser utilizadas como las bases para acceder a sensores y datos. c) Las redes de nueva generación (NGN) ofrecen un buen sustrato que facilita la integración de redes de sensores y el desarrollo de servicios. En consecuencia, una nueva plataforma de Servicios, llamada Ubiquitous Sensor Networks (USN), se ha definido en esta Tesis tratando de contribuir a rellenar los huecos previamente mencionados. Los puntos más destacados de la plataforma USN son: a) Desde un punto de vista arquitectónico, sigue una aproximación de dos niveles (Habilitador y Gateway) similar a otros habilitadores que utilizan las NGN (como el OMA Presence). b) Los modelos de datos están basado en los estándares del OGC SWE. iv c) Está integrado en las NGN pero puede ser utilizado sin ellas utilizando infraestructuras IP abiertas. d) Las principales funciones son: Descubrimiento de sensores, Almacenamiento de observaciones, Publicacion--‐subscripcion--‐notificación, ejecución remota homogénea, seguridad, gestión de diccionarios de datos, facilidades de monitorización, utilidades de conversión de protocolos, interacciones síncronas y asíncronas, soporte para el “streaming” y arbitrado básico de recursos. Para demostrar las funcionalidades que la Plataforma USN propuesta pueden ofrecer a los futuros escenarios de la Internet de las Cosas, se presentan resultados experimentales de tres pruebas de concepto (telemetría, “Smart Places” y monitorización medioambiental) reales a pequeña escala y un estudio sobre semántica (sistema de información vehicular). Además, se está utilizando actualmente como Habilitador para desarrollar tanto experimentación como servicios reales en el proyecto Europeo SmartSantander (que aspira a integrar alrededor de 20.000 dispositivos IoT). v Abstract Internet of Things, as part of the Future Internet, has become one of the main research topics nowadays; in part thanks to the pressure the society is putting on the development of a particular kind of services (Smart metering, Smart Grids, eHealth, etc.), and by the recent business forecasts that situate some players, like Telecom Operators (which are desperately seeking for new opportunities), at the forefront pushing for some interrelated technologies like Machine--‐to--‐Machine (M2M) communications. Under this context, an important number of research activities are currently taking place worldwide at different levels: sensor network communications, information processing, big--‐ data storage, semantics, service level architectures, etc. All of them, isolated, are arriving to a level of maturity that envision the achievement of Internet of Things (IoT) more than a dream, a tangible goal. However, the aforementioned services cannot wait to be developed until the holistic research actions bring complete solutions. It is important to come out with intermediate results that avoid vertical solutions tailored for particular deployments. In the present work, we focus on the creation of a Service--‐level platform intended to facilitate, from one side the integration of heterogeneous and geographically disperse Sensors and Actuator Networks (SANs), and from the other the development of horizontal services using them and the information they provide. This enabler will be used for horizontal service development and for IoT experimentation. Prior to the definition of the platform, we have realized an important study targeting not just research works and projects, but also standardization topics. The results can be summarized in the following assertions: a) Open Geospatial Consortium (OGC®) Sensor Web Enablement (SWE™) data models today represent the most complete solution to describe SANs and observations. b) OGC interfaces, despite the limitations that require changes and extensions, could be used as the bases for accessing sensors and data. c) Next Generation Networks (NGN) offer a good substrate that facilitates the integration of SANs and the development of services. Consequently a new Service Layer platform, called Ubiquitous Sensor Networks (USN), has been defined in this Thesis trying to contribute to fill in the previous gaps. The main highlights of the proposed USN Platform are: a) From an architectural point of view, it follows a two--‐layer approach (Enabler and Gateway) similar to other enablers that run on top of NGN (like the OMA Presence). b) Data models and interfaces are based on the OGC SWE standards. c) It is integrated in NGN but it can be used without it over open IP infrastructures. d) Main functions are: Sensor Discovery, Observation Storage, Publish--‐Subscribe--‐Notify, homogeneous remote execution, security, data dictionaries handling, monitoring facilities, authorization support, protocol conversion utilities, synchronous and asynchronous interactions, streaming support and basic resource arbitration. vi In order to demonstrate the functionalities that the proposed USN Platform can offer to future IoT scenarios, some experimental results have been addressed in three real--‐life small--‐scale proofs--‐of concepts (Smart Metering, Smart Places and Environmental monitoring) and a study for semantics (in--‐vehicle information system). Furthermore we also present the current use of the proposed USN Platform as an Enabler to develop experimentation and real services in the SmartSantander EU project (that aims at integrating around 20.000 IoT devices).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Métrica de calidad de video de alta definición construida a partir de ratios de referencia completa. La medida de calidad de video, en inglés Visual Quality Assessment (VQA), es uno de los mayores retos por solucionar en el entorno multimedia. La calidad de vídeo tiene un impacto altísimo en la percepción del usuario final (consumidor) de los servicios sustentados en la provisión de contenidos multimedia y, por tanto, factor clave en la valoración del nuevo paradigma denominado Calidad de la Experiencia, en inglés Quality of Experience (QoE). Los modelos de medida de calidad de vídeo se pueden agrupar en varias ramas según la base técnica que sustenta el sistema de medida, destacando en importancia los que emplean modelos psicovisuales orientados a reproducir las características del sistema visual humano, en inglés Human Visual System, del que toman sus siglas HVS, y los que, por el contrario, optan por una aproximación ingenieril en la que el cálculo de calidad está basado en la extracción de parámetros intrínsecos de la imagen y su comparación. A pesar de los avances recogidos en este campo en los últimos años, la investigación en métricas de calidad de vídeo, tanto en presencia de referencia (los modelos denominados de referencia completa), como en presencia de parte de ella (modelos de referencia reducida) e incluso los que trabajan en ausencia de la misma (denominados sin referencia), tiene un amplio camino de mejora y objetivos por alcanzar. Dentro de ellos, la medida de señales de alta definición, especialmente las utilizadas en las primeras etapas de la cadena de valor que son de muy alta calidad, son de especial interés por su influencia en la calidad final del servicio y no existen modelos fiables de medida en la actualidad. Esta tesis doctoral presenta un modelo de medida de calidad de referencia completa que hemos llamado PARMENIA (PArallel Ratios MEtric from iNtrInsic features Analysis), basado en la ponderación de cuatro ratios de calidad calculados a partir de características intrínsecas de la imagen. Son: El Ratio de Fidelidad, calculado mediante el gradiente morfológico o gradiente de Beucher. El Ratio de Similitud Visual, calculado mediante los puntos visualmente significativos de la imagen a través de filtrados locales de contraste. El Ratio de Nitidez, que procede de la extracción del estadístico de textura de Haralick contraste. El Ratio de Complejidad, obtenido de la definición de homogeneidad del conjunto de estadísticos de textura de Haralick PARMENIA presenta como novedad la utilización de la morfología matemática y estadísticos de Haralick como base de una métrica de medida de calidad, pues esas técnicas han estado tradicionalmente más ligadas a la teledetección y la segmentación de objetos. Además, la aproximación de la métrica como un conjunto ponderado de ratios es igualmente novedosa debido a que se alimenta de modelos de similitud estructural y otros más clásicos, basados en la perceptibilidad del error generado por la degradación de la señal asociada a la compresión. PARMENIA presenta resultados con una altísima correlación con las valoraciones MOS procedentes de las pruebas subjetivas a usuarios que se han realizado para la validación de la misma. El corpus de trabajo seleccionado procede de conjuntos de secuencias validados internacionalmente, de modo que los resultados aportados sean de la máxima calidad y el máximo rigor posible. La metodología de trabajo seguida ha consistido en la generación de un conjunto de secuencias de prueba de distintas calidades a través de la codificación con distintos escalones de cuantificación, la obtención de las valoraciones subjetivas de las mismas a través de pruebas subjetivas de calidad (basadas en la recomendación de la Unión Internacional de Telecomunicaciones BT.500), y la validación mediante el cálculo de la correlación de PARMENIA con estos valores subjetivos, cuantificada a través del coeficiente de correlación de Pearson. Una vez realizada la validación de los ratios y optimizada su influencia en la medida final y su alta correlación con la percepción, se ha realizado una segunda revisión sobre secuencias del hdtv test dataset 1 del Grupo de Expertos de Calidad de Vídeo (VQEG, Video Quality Expert Group) mostrando los resultados obtenidos sus claras ventajas. Abstract Visual Quality Assessment has been so far one of the most intriguing challenges on the media environment. Progressive evolution towards higher resolutions while increasing the quality needed (e.g. high definition and better image quality) aims to redefine models for quality measuring. Given the growing interest in multimedia services delivery, perceptual quality measurement has become a very active area of research. First, in this work, a classification of objective video quality metrics based on their underlying methodologies and approaches for measuring video quality has been introduced to sum up the state of the art. Then, this doctoral thesis describes an enhanced solution for full reference objective quality measurement based on mathematical morphology, texture features and visual similarity information that provides a normalized metric that we have called PARMENIA (PArallel Ratios MEtric from iNtrInsic features Analysis), with a high correlated MOS score. The PARMENIA metric is based on the pooling of different quality ratios that are obtained from three different approaches: Beucher’s gradient, local contrast filtering, and contrast and homogeneity Haralick’s texture features. The metric performance is excellent, and improves the current state of the art by providing a wide dynamic range that make easier to discriminate between very close quality coded sequences, especially for very high bit rates whose quality, currently, is transparent for quality metrics. PARMENIA introduces a degree of novelty against other working metrics: on the one hand, exploits the structural information variation to build the metric’s kernel, but complements the measure with texture information and a ratio of visual meaningful points that is closer to typical error sensitivity based approaches. We would like to point out that PARMENIA approach is the only metric built upon full reference ratios, and using mathematical morphology and texture features (typically used in segmentation) for quality assessment. On the other hand, it gets results with a wide dynamic range that allows measuring the quality of high definition sequences from bit rates of hundreds of Megabits (Mbps) down to typical distribution rates (5-6 Mbps), even streaming rates (1- 2 Mbps). Thus, a direct correlation between PARMENIA and MOS scores are easily constructed. PARMENIA may further enhance the number of available choices in objective quality measurement, especially for very high quality HD materials. All this results come from validation that has been achieved through internationally validated datasets on which subjective tests based on ITU-T BT.500 methodology have been carried out. Pearson correlation coefficient has been calculated to verify the accuracy of PARMENIA and its reliability.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Esta memoria está basada en el crecimiento y caracterización de heteroestructuras Al(Ga)N/GaN y nanocolumnas ordenadas de GaN, y su aplicación en sensores químicos. El método de crecimiento ha sido la epitaxia de haces moleculares asistida por plasma (PAMBE). En el caso de las heteroestructuras Al(Ga)N/GaN, se han crecido barreras de distinto espesor y composición, desde AlN de 5 nm, hasta AlGaN de 35 nm. Además de una caracterización morfológica, estructural y eléctrica básica de las capas, también se han fabricado a partir de ellas dispositivos tipo HEMTs. La caracterización eléctrica de dichos dispositivos (carga y movilidad de en el canal bidimensional) indica que las mejores heteroestructuras son aquellas con un espesor de barrera intermedio (alrededor de 20 nm). Sin embargo, un objetivo importante de esta Tesis ha sido verificar las ventajas que podían tener los sensores basados en heteroestructuras AlN/GaN (frente a los típicos basados en AlGaN/GaN), con espesores de barrera muy finos (alrededor de 5 nm), ya que el canal de conducción que se modula por efecto de cambios químicos está más cerca de la superficie en donde ocurren dichos cambios químicos. De esta manera, se han utilizado los dispositivos tipo HEMTs como sensores químicos de pH (ISFETs), y se ha comprobado la mayor sensibilidad (variación de corriente frente a cambios de pH, Ids/pH) en los sensores basados en AlN/GaN frente a los basados en AlGaN/GaN. La mayor sensibilidad es incluso más patente en aplicaciones en las que no se utiliza un electrodo de referencia. Se han fabricado y caracterizado dispositivos ISFET similares utilizando capas compactas de InN. Estos sensores presentan peor estabilidad que los basados en Al(Ga)N/GaN, aunque la sensibilidad superficial al pH era la misma (Vgs/pH), y su sensibilidad en terminos de corriente de canal (Ids/pH) arroja valores intermedios entre los ISFET basados en AlN/GaN y los valores de los basados en AlGaN/GaN. Para continuar con la comparación entre dispositivos basados en Al(Ga)N/GaN, se fabricaron ISFETs con el área sensible más pequeña (35 x 35 m2), de tamaño similar a los dispositivos destinados a las medidas de actividad celular. Sometiendo los dispositivos a pulsos de voltaje en su área sensible, la respuesta de los dispositivos de AlN presentaron menor ruido que los basados en AlGaN. El ruido en la corriente para dispositivos de AlN, donde el encapsulado no ha sido optimizado, fue tan bajo como 8.9 nA (valor rms), y el ruido equivalente en el potencial superficial 38.7 V. Estos valores son más bajos que los encontrados en los dispositivos típicos para la detección de actividad celular (basados en Si), y del orden de los mejores resultados encontrados en la literatura sobre AlGaN/GaN. Desde el punto de vista de la caracterización electro-química de las superficies de GaN e InN, se ha determinado su punto isoeléctrico. Dicho valor no había sido reportado en la literatura hasta el momento. El valor, determinado por medidas de “streaming potential”, es de 4.4 y 4 respectivamente. Este valor es una importante característica a tener en cuenta en sensores, en inmovilización electrostática o en la litografía coloidal. Esta última técnica se discute en esta memoria, y se aplica en el último bloque de investigación de esta Tesis (i.e. crecimiento ordenado). El último apartado de resultados experimentales de esta Tesis analiza el crecimiento selectivo de nanocolumnas ordenadas de GaN por MBE, utilizando mascaras de Ti con nanoagujeros. Se ha estudiado como los distintos parámetros de crecimiento (i.e. flujos de los elementos Ga y N, temperatura de crecimiento y diseño de la máscara) afectan a la selectividad y a la morfología de las nanocolumnas. Se ha conseguido con éxito el crecimiento selectivo sobre pseudosustratos de GaN con distinta orientación cristalina o polaridad; templates de GaN(0001)/zafiro, GaN(0001)/AlN/Si, GaN(000-1)/Si y GaN(11-20)/zafiro. Se ha verificado experimentalmente la alta calidad cristalina de las nanocolumnas ordenadas, y su mayor estabilidad térmica comparada con las capas compactas del mismo material. Las nanocolumnas ordenadas de nitruros del grupo III tienen una clara aplicación en el campo de la optoelectrónica, principalmente para nanoemisores de luz blanca. Sin embargo, en esta Tesis se proponen como alternativa a la utilización de capas compactas o nanocolumnas auto-ensambladas en sensores. Las nanocolumnas auto-ensambladas de GaN, debido a su alta razón superficie/volumen, son muy prometedoras en el campo de los sensores, pero su amplia dispersión en dimensiones (altura y diámetro) supone un problema para el procesado y funcionamiento de dispositivos reales. En ese aspecto, las nanocolumnas ordenadas son más robustas y homogéneas, manteniendo una alta relación superficie/volumen. Como primer experimento en el ámbito de los sensores, se ha estudiado como se ve afectada la emisión de fotoluminiscencia de las NCs ordenadas al estar expuestas al aire o al vacio. Se observa una fuerte caída en la intensidad de la fotoluminiscencia cuando las nanocolumnas están expuestas al aire (probablemente por la foto-adsorción de oxigeno en la superficie), como ya había sido documentado anteriormente en nanocolumnas auto-ensambladas. Este experimento abre el camino para futuros sensores basados en nanocolumnas ordenadas. Abstract This manuscript deals with the growth and characterization of Al(Ga)N/GaN heterostructures and GaN ordered nanocolumns, and their application in chemical sensors. The growth technique has been the plasma-assisted molecular beam epitaxy (PAMBE). In the case of Al(Ga)N/GaN heterostructures, barriers of different thickness and composition, from AlN (5 nm) to AlGaN (35 nm) have been grown. Besides the basic morphological, structural and electrical characterization of the layers, HEMT devices have been fabricated based on these layers. The best electrical characteristics (larger carriers concentration and mobility in the two dimensional electron gas) are those in AlGaN/GaN heterostructures with a medium thickness (around 20 nm). However, one of the goals of this Thesis has been to verify the advantages that sensors based on AlN/GaN (thickness around 7 nm) have compared to standard AlGaN/GaN, because the conduction channel to be modulated by chemical changes is closer to the sensitive area. In this way, HEMT devices have been used as chemical pH sensors (ISFETs), and the higher sensitivity (conductance change related to pH changes, Ids/pH) of AlN/GaN based sensors has been proved. The higher sensibility is even more obvious in application without reference electrode. Similar ISFETs devices have been fabricated based on InN compact layers. These devices show a poor stability, but its surface sensitivity to pH (Vgs/pH) and its sensibility (Ids/pH) yield values between the corresponding ones of AlN/GaN and AlGaN/GaN heterostructures. In order to a further comparison between Al(Ga)N/GaN based devices, ISFETs with smaller sensitive area (35 x 35 m2), similar to the ones used in cellular activity record, were fabricated and characterized. When the devices are subjected to a voltage pulse through the sensitive area, the response of AlN based devices shows lower noise than the ones based on AlGaN. The noise in the current of such a AlN based device, where the encapsulation has not been optimized, is as low as 8.9 nA (rms value), and the equivalent noise to the surface potential is 38.7 V. These values are lower than the found in typical devices used for cellular activity recording (based on Si), and in the range of the best published results on AlGaN/GaN. From the point of view of the electrochemical characterization of GaN and InN surfaces, their isoelectric point has been experimentally determined. Such a value is the first time reported for GaN and InN surfaces. These values are determined by “streaming potential”, being pH 4.4 and 4, respectively. Isoelectric point value is an important characteristic in sensors, electrostatic immobilization or in colloidal lithography. In particular, colloidal lithography has been optimized in this Thesis for GaN surfaces, and applied in the last part of experimental results (i.e. ordered growth). The last block of this Thesis is focused on the selective area growth of GaN nanocolumns by MBE, using Ti masks decorated with nanoholes. The effect of the different growth parameters (Ga and N fluxes, growth temperature and mask design) is studied, in particular their impact in the selectivity and in the morphology of the nanocolumns. Selective area growth has been successful performed on GaN templates with different orientation or polarity; GaN(0001)/sapphire, GaN(0001)/AlN/Si, GaN(000- 1)/Si and GaN(11-20)/sapphire. Ordered nanocolumns exhibit a high crystal quality, and a higher thermal stability (lower thermal decomposition) than the compact layers of the same material. Ordered nanocolumns based on III nitrides have a clear application in optoelectronics, mainly for white light nanoemitters. However, this Thesis proposes them as an alternative to compact layers and self-assembled nanocolumns in sensor applications. Self-assembled GaN nanocolumns are very appealing for sensor applications, due to their large surface/volume ratio. However, their large dispersion in heights and diameters are a problem in terms of processing and operation of real devices. In this aspect, ordered nanocolumns are more robust and homogeneous, keeping the large surface/volume ratio. As first experimental evidence of their sensor capabilities, ordered nanocolumns have been studied regarding their photoluminiscence on air and vacuum ambient. A big drop in the intensity is observed when the nanocolumns are exposed to air (probably because of the oxygen photo-adsortion), as was already reported in the case of self-assembled nanocolumns. This opens the way to future sensors based on ordered III nitrides nanocolumns.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Sensor networks are increasingly becoming one of the main sources of Big Data on the Web. However, the observations that they produce are made available with heterogeneous schemas, vocabularies and data formats, making it difficult to share and reuse these data for other purposes than those for which they were originally set up. In this thesis we address these challenges, considering how we can transform streaming raw data to rich ontology-based information that is accessible through continuous queries for streaming data. Our main contribution is an ontology-based approach for providing data access and query capabilities to streaming data sources, allowing users to express their needs at a conceptual level, independent of implementation and language-specific details. We introduce novel query rewriting and data translation techniques that rely on mapping definitions relating streaming data models to ontological concepts. Specific contributions include: • The syntax and semantics of the SPARQLStream query language for ontologybased data access, and a query rewriting approach for transforming SPARQLStream queries into streaming algebra expressions. • The design of an ontology-based streaming data access engine that can internally reuse an existing data stream engine, complex event processor or sensor middleware, using R2RML mappings for defining relationships between streaming data models and ontology concepts. Concerning the sensor metadata of such streaming data sources, we have investigated how we can use raw measurements to characterize streaming data, producing enriched data descriptions in terms of ontological models. Our specific contributions are: • A representation of sensor data time series that captures gradient information that is useful to characterize types of sensor data. • A method for classifying sensor data time series and determining the type of data, using data mining techniques, and a method for extracting semantic sensor metadata features from the time series.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Sensor networks are increasingly being deployed in the environment for many different purposes. The observations that they produce are made available with heterogeneous schemas, vocabularies and data formats, making it difficult to share and reuse this data, for other purposes than those for which they were originally set up. The authors propose an ontology-based approach for providing data access and query capabilities to streaming data sources, allowing users to express their needs at a conceptual level, independent of implementation and language-specific details. In this article, the authors describe the theoretical foundations and technologies that enable exposing semantically enriched sensor metadata, and querying sensor observations through SPARQL extensions, using query rewriting and data translation techniques according to mapping languages, and managing both pull and push delivery modes.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

ATM, SDH or satellite have been used in the last century as the contribution network of Broadcasters. However the attractive price of IP networks is changing the infrastructure of these networks in the last decade. Nowadays, IP networks are widely used, but their characteristics do not offer the level of performance required to carry high quality video under certain circumstances. Data transmission is always subject to errors on line. In the case of streaming, correction is attempted at destination, while on transfer of files, retransmissions of information are conducted and a reliable copy of the file is obtained. In the latter case, reception time is penalized because of the low priority this type of traffic on the networks usually has. While in streaming, image quality is adapted to line speed, and line errors result in a decrease of quality at destination, in the file copy the difference between coding speed vs line speed and errors in transmission are reflected in an increase of transmission time. The way news or audiovisual programs are transferred from a remote office to the production centre depends on the time window and the type of line available; in many cases, it must be done in real time (streaming), with the resulting image degradation. The main purpose of this work is the workflow optimization and the image quality maximization, for that reason a transmission model for multimedia files adapted to JPEG2000, is described based on the combination of advantages of file transmission and those of streaming transmission, putting aside the disadvantages that these models have. The method is based on two patents and consists of the safe transfer of the headers and data considered to be vital for reproduction. Aside, the rest of the data is sent by streaming, being able to carry out recuperation operations and error concealment. Using this model, image quality is maximized according to the time window. In this paper, we will first give a briefest overview of the broadcasters requirements and the solutions with IP networks. We will then focus on a different solution for video file transfer. We will take the example of a broadcast center with mobile units (unidirectional video link) and regional headends (bidirectional link), and we will also present a video file transfer file method that satisfies the broadcaster requirements.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, an innovative approach to perform distributed Bayesian inference using a multi-agent architecture is presented. The final goal is dealing with uncertainty in network diagnosis, but the solution can be of applied in other fields. The validation testbed has been a P2P streaming video service. An assessment of the work is presented, in order to show its advantages when it is compared with traditional manual processes and other previous systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

En los últimos años, debido al notable desarrollo de los terminales portátiles, que han pasado de ser “simples” teléfonos o reproductores a puros ordenadores, ha crecido el número de servicios que ofrecen cada vez mayor cantidad de contenido multimedia a través de internet. Además, la distinta evolución de estos terminales hace que nos encontremos en el mercado con una amplísima gama de productos de diferentes tamaños y capacidades de procesamiento, lo que hace necesario encontrar una fórmula que permita satisfacer la demanda de dichos servicios sea cual sea la naturaleza de nuestro dispositivo. Para poder ofrecer una solución adecuada se ha optado por la integración de un protocolo como RTP y un estándar de video como SVC. RTP (Real-time Transport Protocol), en contraposición a los protocolos de propósito general fue diseñado para aplicaciones de tiempo real por lo que es ideal para el streaming de contenido multimedia. Por su parte, SVC es un estándar de video escalable que permite transmitir en un mismo stream una capa base y múltiples capas de mejora, por lo que podremos adaptar la calidad y tamaño del contenido a la capacidad y tamaño de nuestro dispositivo. El objetivo de este proyecto consiste en integrar y modificar tanto el reproductor MPlayer como la librería RTP live555 de tal forma que sean capaces de soportar el formato SVC sobre el protocolo RTP y montar un sistema servidorcliente para comprobar su funcionamiento. Aunque este proceso esté orientado a llevarse a cabo en un dispositivo móvil, para este proyecto se ha optado por realizarlo en el escenario más sencillo posible, para lo cual, se emitirán secuencias a una máquina virtual alojada en el mismo ordenador que el servidor. ABSTRACT In recent years, due to the remarkable development of mobile devices, which have evolved from "simple" phones or players to computers, the amount of services that offer multimedia content over the internet have shot up. Furthermore, the different evolution of these terminals causes that we can find in the market a wide range of different sizes and processing capabilities, making necessary to find a formula that will satisfy the demand for such services regardless of the nature of our device. In order to provide a suitable solution we have chosen to integrate a protocol as RTP and a video standard as SVC. RTP (Real-time Transport Protocol), in opposition to general purpose protocols was designed for real-time applications making it ideal for media streaming. Meanwhile, SVC is a scalable video standard which can transmit a single stream in a base layer and multiple enhancement layers, so that we can adapt the quality and size of the content to the capacity and size of our device. The objective of this project is to integrate and modify both MPlayer and RTP library live555 so that they support the SVC format over RTP protocol and set up a client-server system to check its behavior. Although this process has been designed to be done on a mobile device, for this project we have chosen to do it in the simplest possible scenario so we will stream to a virtual machine hosted on the same computer where we have the server.