28 resultados para available bandwidth

em Universidad Politécnica de Madrid


Relevância:

60.00% 60.00%

Publicador:

Resumo:

Multi-user videoconferencing systems offer communication between more than two users, who are able to interact through their webcams, microphones and other components. The use of these systems has been increased recently due to, on the one hand, improvements in Internet access, networks of companies, universities and houses, whose available bandwidth has been increased whilst the delay in sending and receiving packets has decreased. On the other hand, the advent of Rich Internet Applications (RIA) means that a large part of web application logic and control has started to be implemented on the web browsers. This has allowed developers to create web applications with a level of complexity comparable to traditional desktop applications, running on top of the Operating Systems. More recently the use of Cloud Computing systems has improved application scalability and involves a reduction in the price of backend systems. This offers the possibility of implementing web services on the Internet with no need to spend a lot of money when deploying infrastructures and resources, both hardware and software. Nevertheless there are not many initiatives that aim to implement videoconferencing systems taking advantage of Cloud systems. This dissertation proposes a set of techniques, interfaces and algorithms for the implementation of videoconferencing systems in public and private Cloud Computing infrastructures. The mechanisms proposed here are based on the implementation of a basic videoconferencing system that runs on the web browser without any previous installation requirements. To this end, the development of this thesis starts from a RIA application with current technologies that allow users to access their webcams and microphones from the browser, and to send captured data through their Internet connections. Furthermore interfaces have been implemented to allow end users to participate in videoconferencing rooms that are managed in different Cloud provider servers. To do so this dissertation starts from the results obtained from the previous techniques and backend resources were implemented in the Cloud. A traditional videoconferencing service which was implemented in the department was modified to meet typical Cloud Computing infrastructure requirements. This allowed us to validate whether Cloud Computing public infrastructures are suitable for the traffic generated by this kind of system. This analysis focused on the network level and processing capacity and stability of the Cloud Computing systems. In order to improve this validation several other general considerations were taken in order to cover more cases, such as multimedia data processing in the Cloud, as research activity has increased in this area in recent years. The last stage of this dissertation is the design of a new methodology to implement these kinds of applications in hybrid clouds reducing the cost of videoconferencing systems. Finally, this dissertation opens up a discussion about the conclusions obtained throughout this study, resulting in useful information from the different stages of the implementation of videoconferencing systems in Cloud Computing systems. RESUMEN Los sistemas de videoconferencia multiusuario permiten la comunicación entre más de dos usuarios que pueden interactuar a través de cámaras de video, micrófonos y otros elementos. En los últimos años el uso de estos sistemas se ha visto incrementado gracias, por un lado, a la mejora de las redes de acceso en las conexiones a Internet en empresas, universidades y viviendas, que han visto un aumento del ancho de banda disponible en dichas conexiones y una disminución en el retardo experimentado por los datos enviados y recibidos. Por otro lado también ayudó la aparación de las Aplicaciones Ricas de Internet (RIA) con las que gran parte de la lógica y del control de las aplicaciones web comenzó a ejecutarse en los mismos navegadores. Esto permitió a los desarrolladores la creación de aplicaciones web cuya complejidad podía compararse con la de las tradicionales aplicaciones de escritorio, ejecutadas directamente por los sistemas operativos. Más recientemente el uso de sistemas de Cloud Computing ha mejorado la escalabilidad y el abaratamiento de los costes para sistemas de backend, ofreciendo la posibilidad de implementar servicios Web en Internet sin la necesidad de grandes desembolsos iniciales en las áreas de infraestructuras y recursos tanto hardware como software. Sin embargo no existen aún muchas iniciativas con el objetivo de realizar sistemas de videoconferencia que aprovechen las ventajas del Cloud. Esta tesis doctoral propone un conjunto de técnicas, interfaces y algoritmos para la implentación de sistemas de videoconferencia en infraestructuras tanto públicas como privadas de Cloud Computing. Las técnicas propuestas en la tesis se basan en la realización de un servicio básico de videoconferencia que se ejecuta directamente en el navegador sin la necesidad de instalar ningún tipo de aplicación de escritorio. Para ello el desarrollo de esta tesis parte de una aplicación RIA con tecnologías que hoy en día permiten acceder a la cámara y al micrófono directamente desde el navegador, y enviar los datos que capturan a través de la conexión de Internet. Además se han implementado interfaces que permiten a usuarios finales la participación en salas de videoconferencia que se ejecutan en servidores de proveedores de Cloud. Para ello se partió de los resultados obtenidos en las técnicas anteriores de ejecución de aplicaciones en el navegador y se implementaron los recursos de backend en la nube. Además se modificó un servicio ya existente implementado en el departamento para adaptarlo a los requisitos típicos de las infraestructuras de Cloud Computing. Alcanzado este punto se procedió a analizar si las infraestructuras propias de los proveedores públicos de Cloud Computing podrían soportar el tráfico generado por los sistemas que se habían adaptado. Este análisis se centró tanto a nivel de red como a nivel de capacidad de procesamiento y estabilidad de los sistemas. Para los pasos de análisis y validación de los sistemas Cloud se tomaron consideraciones más generales para abarcar casos como el procesamiento de datos multimedia en la nube, campo en el que comienza a haber bastante investigación en los últimos años. Como último paso se ideó una metodología de implementación de este tipo de aplicaciones para que fuera posible abaratar los costes de los sistemas de videoconferencia haciendo uso de clouds híbridos. Finalmente en la tesis se abre una discusión sobre las conclusiones obtenidas a lo largo de este amplio estudio, obteniendo resultados útiles en las distintas etapas de implementación de los sistemas de videoconferencia en la nube.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Satellite operators are starting to use the Ka-band (30/20 GHz) for communications systems requiring higher traffic capacity. The use of this band is expected to experience a significant growth in the next few years, as several operators have reported plans to launch new satellites with Ka-band capacity. It is worth mentioning the Ka-Sat satellite in Europe, launched in 2010, and ViaSat-1, of 2011, with coverage of USA1. Some other examples can be found in other parts of the World. Recent satellite communications standards, such as DVB-S22 or DVB-RCS3, which provide means to mitigate propagation impairments, have been developed with the objective of improving the use of the Ka-band, in comparison with previous technical standards. In the next years, the ALPHASAT satellite will bring about new opportunities4 for carrying out propagation and telecommunication experiments in the Ka- and Q/V-bands. Commercial uses are focused on the provision of high speed data communications, for Internet access and other applications. In the near future, it is expected that higher and higher data rates will also be needed to broadcast richer multimedia contents, including HD-TV, interactive content or 3D-TV. All of these services may be provided in the future by satellites of the current generation, whose life span can extend up to 2025 in some cases. Depending on local regulations, the available bandwidth for the satellite fixed and broadcasting services in the Ka-band is in excess of several hundred MHz, bidirectional, comprising more than 1 GHz for each sub-band in some cases. In this paper, the results of a propagation experiment that is being carried out at Universidad Politécnica de Madrid (UPM), Spain, are presented5. The objective of the experiment is twofold: gathering experimental time series of attenuation and analyzing them in order to characterize the propagation channel at these frequencies6. The experiment and statistical results correspond to five complete years of measurements. The experiment is described in more detail in Section II. Yearly characteristics of rain attenuation are presented in Section III, whereas Section IV is dedicated to the monthly, seasonal, and hourly characteristics. Section V covers the dynamic characteristics of this propagation effect, just before the conclusions are described in Section VI.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Nowadays, HTTP adaptive streaming (HAS) has become a reliable distribution technology offering significant advantages in terms of both user perceived Quality of Experience (QoE) and resource utilization for content and network service providers. By trading-off the video quality, HAS is able to adapt to the available bandwidth and display requirements so that it can deliver the video content to a variety of devices over the Internet. However, until now there is not enough knowledge of how the adaptation techniques affect the end user's visual experience. Therefore, this paper presents a comparative analysis of different bitrate adaptation strategies in adaptive streaming of monoscopic and stereoscopic video. This has been done through a subjective experiment of testing the end-user response to the video quality variations, considering the visual comfort issue. The experimental outcomes have made a good insight into the factors that can influence on the QoE of different adaptation strategies.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Poder clasificar de manera precisa la aplicación o programa del que provienen los flujos que conforman el tráfico de uso de Internet dentro de una red permite tanto a empresas como a organismos una útil herramienta de gestión de los recursos de sus redes, así como la posibilidad de establecer políticas de prohibición o priorización de tráfico específico. La proliferación de nuevas aplicaciones y de nuevas técnicas han dificultado el uso de valores conocidos (well-known) en puertos de aplicaciones proporcionados por la IANA (Internet Assigned Numbers Authority) para la detección de dichas aplicaciones. Las redes P2P (Peer to Peer), el uso de puertos no conocidos o aleatorios, y el enmascaramiento de tráfico de muchas aplicaciones en tráfico HTTP y HTTPS con el fin de atravesar firewalls y NATs (Network Address Translation), entre otros, crea la necesidad de nuevos métodos de detección de tráfico. El objetivo de este estudio es desarrollar una serie de prácticas que permitan realizar dicha tarea a través de técnicas que están más allá de la observación de puertos y otros valores conocidos. Existen una serie de metodologías como Deep Packet Inspection (DPI) que se basa en la búsqueda de firmas, signatures, en base a patrones creados por el contenido de los paquetes, incluido el payload, que caracterizan cada aplicación. Otras basadas en el aprendizaje automático de parámetros de los flujos, Machine Learning, que permite determinar mediante análisis estadísticos a qué aplicación pueden pertenecer dichos flujos y, por último, técnicas de carácter más heurístico basadas en la intuición o el conocimiento propio sobre tráfico de red. En concreto, se propone el uso de alguna de las técnicas anteriormente comentadas en conjunto con técnicas de minería de datos como son el Análisis de Componentes Principales (PCA por sus siglas en inglés) y Clustering de estadísticos extraídos de los flujos procedentes de ficheros de tráfico de red. Esto implicará la configuración de diversos parámetros que precisarán de un proceso iterativo de prueba y error que permita dar con una clasificación del tráfico fiable. El resultado ideal sería aquel en el que se pudiera identificar cada aplicación presente en el tráfico en un clúster distinto, o en clusters que agrupen grupos de aplicaciones de similar naturaleza. Para ello, se crearán capturas de tráfico dentro de un entorno controlado e identificando cada tráfico con su aplicación correspondiente, a continuación se extraerán los flujos de dichas capturas. Tras esto, parámetros determinados de los paquetes pertenecientes a dichos flujos serán obtenidos, como por ejemplo la fecha y hora de llagada o la longitud en octetos del paquete IP. Estos parámetros serán cargados en una base de datos MySQL y serán usados para obtener estadísticos que ayuden, en un siguiente paso, a realizar una clasificación de los flujos mediante minería de datos. Concretamente, se usarán las técnicas de PCA y clustering haciendo uso del software RapidMiner. Por último, los resultados obtenidos serán plasmados en una matriz de confusión que nos permitirá que sean valorados correctamente. ABSTRACT. Being able to classify the applications that generate the traffic flows in an Internet network allows companies and organisms to implement efficient resource management policies such as prohibition of specific applications or prioritization of certain application traffic, looking for an optimization of the available bandwidth. The proliferation of new applications and new technics in the last years has made it more difficult to use well-known values assigned by the IANA (Internet Assigned Numbers Authority), like UDP and TCP ports, to identify the traffic. Also, P2P networks and data encapsulation over HTTP and HTTPS traffic has increased the necessity to improve these traffic analysis technics. The aim of this project is to develop a number of techniques that make us able to classify the traffic with more than the simple observation of the well-known ports. There are some proposals that have been created to cover this necessity; Deep Packet Inspection (DPI) tries to find signatures in the packets reading the information contained in them, the payload, looking for patterns that can be used to characterize the applications to which that traffic belongs; Machine Learning procedures work with statistical analysis of the flows, trying to generate an automatic process that learns from those statistical parameters and calculate the likelihood of a flow pertaining to a certain application; Heuristic Techniques, finally, are based in the intuition or the knowledge of the researcher himself about the traffic being analyzed that can help him to characterize the traffic. Specifically, the use of some of the techniques previously mentioned in combination with data mining technics such as Principal Component Analysis (PCA) and Clustering (grouping) of the flows extracted from network traffic captures are proposed. An iterative process based in success and failure will be needed to configure these data mining techniques looking for a reliable traffic classification. The perfect result would be the one in which the traffic flows of each application is grouped correctly in each cluster or in clusters that contain group of applications of similar nature. To do this, network traffic captures will be created in a controlled environment in which every capture is classified and known to pertain to a specific application. Then, for each capture, all the flows will be extracted. These flows will be used to extract from them information such as date and arrival time or the IP length of the packets inside them. This information will be then loaded to a MySQL database where all the packets defining a flow will be classified and also, each flow will be assigned to its specific application. All the information obtained from the packets will be used to generate statistical parameters in order to describe each flow in the best possible way. After that, data mining techniques previously mentioned (PCA and Clustering) will be used on these parameters making use of the software RapidMiner. Finally, the results obtained from the data mining will be compared with the real classification of the flows that can be obtained from the database. A Confusion Matrix will be used for the comparison, letting us measure the veracity of the developed classification process.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Este proyecto versa sobre un modelo de evaluación de calidad de imagen aplicado a la optimización del ancho de banda del multiplex de televisión, mediante la realización de ensayos con distintas configuraciones de cabecera. Dicho modelo se basa en las medidas PQR y DMOS de Tektronix, destinadas a medir la percepción de las diferencias entre una secuencia antes y después de sufrir un procesado digital. Dado que actualmente, el modo de trabajo de una cabecera de televisión digital, es la multiplexación estadística (consistente en la codificación de diferentes servicios de vídeo con anchos de banda variables en función de la complejidad de las señales), las medidas estarán enfocadas a sacar conclusiones acerca de la cantidad de canales, complejidad de contenidos, y organización de los mismos en el ancho de banda disponible para emitir, manteniendo niveles de calidad Broadcast. Las medidas serán aplicadas en el proyecto para comparar el rendimiento de dos modelos de cabecera. En primer lugar serán configuradas en régimen binario constante, comparando el rendimiento de los codificadores en el área de trabajo habitual. Posteriormente se configuraran en régimen binario variable probando múltiples configuraciones, con el objetivo dar con el modelo y configuración óptima para su posterior implementación. ABSTRACT. This project concerns a picture quality assessment model applied to television multiplex bandwidth optimization by conducting test with different headend settings. This model is based on the PQR and DMOS Tektronix measures, designed to measure the differences between a sequence before and after a digital processing. Given that nowadays the working way of television headend is by statistical multiplexing (based on coding the different video services with variable bitrate depending on the complexity of the signals) measures will be focused to reach conclusions about the number of channels, complexity of content, and the way to organize them through the available bandwidth, keeping broadcast quality ratios. The measures will be applied comparing the performance of two headend models. First of all encoders will be set on constant bitrate, comparing the performance through the working bandwidth. Later on, both models will be set on variable bitrate testing multiple configurations, in order to find the optimal model/configuration for later implementation.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

El estándar LTE se ha posicionado como una de las claves para que los operadores de telecomunicación puedan abordar de manera eficiente en costes el crecimiento de la demanda de tráfico móvil que se prevé para los próximos años, al ser una tecnología más escalable en el núcleo de la red y más flexible en la interfaz radio que sus predecesoras. En este sentido, es necesario también que los reguladores garanticen un acceso al espectro radioeléctrico adecuado, equitativo y no discriminatorio, que permita un entorno estable para el despliegue de redes de comunicaciones móviles avanzadas. Además de la flexibilización del marco regulador del espectro radioeléctrico en Europa, que ha permitido el despliegue de nuevas tecnologías en las bandas de frecuencia históricas de GSM, se ha puesto a disposición espectro adicional para sistemas IMT en nuevas bandas de frecuencia, lo que ha planteando a su vez nuevos retos para la tecnología y la regulación. La fragmentación del espectro disponible para comunicaciones móviles ha impulsado el desarrollo de técnicas de agregación de portadoras en las nuevas versiones del estándar LTE, que permiten explotar mejor los recursos radio en su conjunto. No obstante, el espectro inferior a 1 GHz sigue siendo escaso, ya que el tráfico móvil aumenta y la banda de 900 MHz aún se utiliza para servicios GSM, lo que no ha conseguido sino agravar la disputa entre los servicios de radiodifusión terrestre y de comunicaciones móviles por la parte superior de la banda UHF. En concreto, la banda de 700 MHz se perfila como una de las próximas para aumentar el espectro disponible para los servicios en movilidad, si bien su liberación por parte de las actuales redes de Televisión Digital Terrestre presenta no pocas dificultades en los Estados miembros en los que ésta es la principal plataforma audiovisual de acceso gratuito, abriendo un debate sobre el modelo audiovisual a largo plazo en Europa. Por otro lado, las políticas públicas de promoción del acceso a la banda ancha rápida y ultrarrápida de la presente década han establecido objetivos ambiciosos para el año 2020, tanto en el ámbito europeo como en los diferentes Estados miembros. La universalización del acceso a redes de banda ancha de al menos 30 Mbps constituye uno de los principales retos. Las expectativas generadas por la tecnología LTE y la puesta a disposición de nuevas bandas de frecuencia hace posible que los servicios de acceso fijo inalámbrico adquieran especial relevancia ante los objetivos de política pública establecidos que, como ha sido reconocido en diversas ocasiones, no podrán lograrse sino con un compendio de diferente tecnologías. Para esta Tesis Doctoral se han desarrollado una serie modelos tecnoeconómicos con el objetivo de realizar un análisis prospectivo que evalúa tres casos de especial relevancia en el despliegue de redes LTE: en primer lugar, la valoración económica de la banda de 700 MHz; en segundo lugar, la evaluación de modelos de negocio y reducción de costes considerando tecnologías femtocelulares; y finalmente, la viabilidad de las redes LTE de acceso fijo inalámbrico para el cierre de la brecha digital en el acceso a la banda ancha de 30 Mbps. En relación con la aplicación del análisis tecnoeconómico para la valoración del espectro de 700 MHz, los resultados obtenidos ponen de manifiesto dos cuestiones fundamentales. En primer lugar, la necesidad de asignar a los operadores más espectro para satisfacer las previsiones de demanda de tráfico móvil a medio plazo. En segundo, existe una diferencia notable en los costes de despliegue de una red LTE cuando se dispone de espectro en frecuencias inferiores a 1 GHz y cuando no, pero esta diferencia de costes disminuye a medida que se añade nuevo espectro sub-1GHz. De esta manera, la atribución de la banda de 700 MHz a servicios de comunicaciones móviles supone una reducción relevante en los costes de despliegue si el operador no dispone de espectro en la banda de 800 MHz, pero no así si ya dispone de espectro en bandas bajas para el despliegue. En este sentido, puede concluirse que el precio que los operadores estarán dispuestos a pagar por el espectro de la banda de 700 MHz dependerá de si ya tienen disponible espectro en la banda de 800 MHz. Sin embargo, dado que la competencia por ese espectro será menor, los ingresos esperables en las licitaciones de esta nueva banda serán en general menores, a pesar de que para algunos operadores este espectro sería tan valioso como el de 800 MHz. En segundo lugar, en relación con el despliegue de femtoceldas pueden extraerse algunas conclusiones en términos de ahorro de costes de despliegue y también de cara a la viabilidad de los modelos de negocio que posibilitan. El ahorro que supone la introducción de femtoceldas en el despliegue de una red LTE frente al caso de un despliegue exclusivamente macrocelular se ha demostrado que es mayor cuanto menor es el ancho de banda disponible para la red macrocelular. En esta línea, para un operador convergente el despliegue de femtoceldas tiene sentido económico si el ancho de banda disponible es escaso (en torno a 2x10 MHz), que, en el caso de España, puede reflejar el caso de los operadores del segmento fijo que son nuevos entrantes en el móvil. Por otro lado, los modelos de acceso abierto son interesantes para operadores exclusivamente móviles, porque consiguen flexibilizar los costes sustituyendo estaciones base macrocelulares por el despliegue de femtoceldas, pero necesitan desplegarse en zonas con una densidad de población relativamente elevada para que éstas descarguen tráfico de varios usuarios de la red macrocelular simultáneamente. No obstante, las femtoceldas son beneficiosas en todo caso si es el usuario quien asume los costes de la femtocelda y el backhaul, lo que sólo parece probable si se integran en el modelo de negocio de comercialización de nuevos servicios. Por tanto, el despliegue de femtoceldas en buena parte de la casuística estudiada sólo tiene sentido si consiguen aumentar los ingresos por usuario comercializando servicios de valor añadido que necesiten calidad de servicio garantizada y exploten a la vez de esa forma su principal ventaja competitiva respecto a la tecnología WiFi. Finalmente, en relación con el papel de la tecnología LTE para la provisión de servicios de acceso fijo inalámbrico para la banda ancha de 30 Mbps, se ha desarrollado un modelo TD-LTE y mediante la metodología de análisis tecnoeconómico se ha realizado un estudio prospectivo para el caso de España. Los resultados obtenidos preciden una huella de cobertura de FTTH del 74% para 2020, y demuestran que una red TD-LTE en la banda de 3,5 GHz resulta viable para aumentar la cobertura de servicios de 30 Mbps en 14 puntos porcentuales. Junto con la consideración de la cobertura de otras redes, la cobertura de 30 Mbps de acuerdo a la viabilidad de los despliegues alcanzaría el 95% en España en el año 2020. Como resumen, los resultados obtenidos muestran en todos los casos la capacidad de la tecnología LTE para afrontar nuevos desafíos en relación con el aumento del tráfico móvil, especialmente crítico en las zonas más urbanas, y el cierre de la brecha digital en el acceso a la banda ancha rápida en las zonas más rurales. ABSTRACT The LTE standard has been pointed out as one of the keys for telecom operators to address the demand growth in mobile traffic foreseen for the next years in a cost-efficient way, since its core network is more scalable and its radio interface more flexible than those of its predecessor technologies. On the other hand, regulators need to guarantee an adequate, equitable and non-discriminatory access to radio spectrum, which enable a favorable environment for the deployment of advanced mobile communication networks. Despite the reform of the spectrum regulatory framework in Europe, which allowed for the deployment of new technologies in the historic GSM bands, additional spectrum has been allocated to IMT systems in new frequency bands, what in turn has set out new challenges for technology and regulation. The current fragmentation of available spectrum in very different frequency bands has boosted the development of carrier aggregation techniques in most recent releases of the LTE standard, which permit a better exploitation of radio resources as a whole. Nonetheless, spectrum below 1 GHz is still scarce for mobile networks, since mobile traffic increases at a more rapid pace than spectral efficiency and spectrum resources. The 900 MHz frequency band is still being used for GSM services, what has worsen the dispute between mobile communication services and terrestrial broadcasting services for the upper part of the UHF band. Concretely, the 700 MHz frequency band has been pointed out as one of the next bands to be allocated to mobile in order to increase available spectrum. However, its release by current Digital Terrestrial Television networks is challenging in Member States where it constitutes the main free access audiovisual platform, opening up a new debate around the audiovisual model in the long term in Europe. On the other hand, public policies of the present decade to promote fast and ultrafast broadband access has established very ambitious objectives for the year 2020, both at European and national levels. Universalization of 30 Mbps broadband access networks constitutes one of the main challenges. Expectations raised by LTE technology and the allocation of new frequency bands has lead fixed wireless access (FWA) services to acquire special relevance in light of public policy objectives, which will not be met but with a compendium of different technologies, as different involved stakeholders have acknowledged. This PhD Dissertation develops techno-economic models to carry out a prospective analysis for three cases of special relevance in LTE networks’ deployment: the spectrum pricing of the 700 MHz frequency band, an assessment of new business models and cost reduction considering femtocell technologies, and the feasibility of LTE fixed wireless access networks to close the 30 Mbps broadband access gap in rural areas. In the first place and regarding the application of techno-economic analysis for 700 MHz spectrum pricing, obtained results reveal two core issues. First of all, the need to allocate more spectrum for operators in order to fulfill mobile traffic demand in the mid-term. Secondly, there is a substantial difference in deployment costs for a LTE network when there is sub-1GHz spectrum available and when there is not, but this difference decreases as additional sub-1GHz spectrum is added. Thus, the allocation of 700 MHz band to mobile communication services would cause a relevant reduction in deployment costs if the operator does not count on spectrum in the 800 MHz, but not if it already has been assigned spectrum in low frequencies for the deployment. In this regard, the price operators will be willing to pay for 700 MHz spectrum will depend on them having already spectrum in the 800 MHz frequency band or not. However, since competition for the new spectrum will not be so strong, expected incomes from 700 MHz spectrum awards will be generally lower than those from the digital dividend, despite this spectrum being as valuable as 800 MHz spectrum for some operators. In the second place, regarding femtocell deployment, some conclusions can be drawn in terms of deployment cost savings and also with reference to the business model they enable. Savings provided by a joint macro-femto LTE network as compared to an exclusively macrocellular deployment increase as the available bandwidth for the macrocells decreases. Therefore, for a convergent operator the deployment of femtocells can only have economic sense if the available bandwidth is scarce (around 2x10 MHz), which might be the case of fix market operators which are new entrant in mobile market. Besides, open access models are interesting for exclusively mobile operators, since they make costs more flexible by substituting macrocell base stations by femtocells, but they need to be deployed relatively densely populated areas so that they can offload traffic from several macrocell users simultaneously. Nonetheless, femtocells are beneficial in all cases if the user assumes both femtocell and backhaul costs, which only seems probable if they are integrated in a business model commercializing new services. Therefore, in many of the cases analyzed femtocell deployment only makes sense if they increase revenues per user through new added value services which need from guaranteed quality of service, thus exploiting its main competitive advantage compared to WiFi. Finally, regarding the role of LTE technology in the provision of fixed wireless access services for 30 Mbps broadband, a TD-LTE model has been developed and a prospective study has been carried out through techno-economic methodology for the Spanish case. Obtained results foresee a FTTH coverage footprint of 74% households for 2020, and prove that a TD-LTE network in the 3.5 GHz band results feasible to increase 30 Mbps service coverage in additional 14 percentage points. To sum up, obtained results show LTE technology capability to address new challenges regarding both mobile traffic growth, particularly critical in urban zones, and the current digital divide in fast broadband access in most rural zones.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, implementation and testing of non- commercial GaN HEMT in a simple buck converter for envelope amplifier in ET and EER transmission techn iques has been done. Comparing to the prototypes with commercially available EPC1014 and 1015 GaN HEMTs, experimentally demonstrated power supply provided better thermal management and increased the switching frequency up to 25MHz. 64QAM signal with 1MHz of large signal bandw idth and 10.5dB of Peak to Average Power Ratio was gener ated, using the switching frequency of 20MHz. The obtaine defficiency was 38% including the driving circuit an d the total losses breakdown showed that switching power losses in the HEMT are the dominant ones. In addition to this, some basic physical modeling has been done, in order to provide an insight on the correlation between the electrical characteristics of the GaN HEMT and physical design parameters. This is the first step in the optimization of the HEMT design for this particular application.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The evolution of the television market is led by 3DTV technology, and this tendency can accelerate during the next years according to expert forecasts. However, 3DTV delivery by broadcast networks is not currently developed enough, and acts as a bottleneck for the complete deployment of the technology. Thus, increasing interest is dedicated to ste-reo 3DTV formats compatible with current HDTV video equipment and infrastructure, as they may greatly encourage 3D acceptance. In this paper, different subsampling schemes for HDTV compatible transmission of both progressive and interlaced stereo 3DTV are studied and compared. The frequency characteristics and preserved frequency content of each scheme are analyzed, and a simple interpolation filter is specially designed. Finally, the advantages and disadvantages of the different schemes and filters are evaluated through quality testing on several progressive and interlaced video sequences.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Here, a novel and efficient moving object detection strategy by non-parametric modeling is presented. Whereas the foreground is modeled by combining color and spatial information, the background model is constructed exclusively with color information, thus resulting in a great reduction of the computational and memory requirements. The estimation of the background and foreground covariance matrices, allows us to obtain compact moving regions while the number of false detections is reduced. Additionally, the application of a tracking strategy provides a priori knowledge about the spatial position of the moving objects, which improves the performance of the Bayesian classifier

Relevância:

20.00% 20.00%

Publicador:

Resumo:

New actuation technology in functional or "smart" materials has opened new horizons in robotics actuation systems. Materials such as piezo-electric fiber composites, electro-active polymers and shape memory alloys (SMA) are being investigated as promising alternatives to standard servomotor technology [52]. This paper focuses on the use of SMAs for building muscle-like actuators. SMAs are extremely cheap, easily available commercially and have the advantage of working at low voltages. The use of SMA provides a very interesting alternative to the mechanisms used by conventional actuators. SMAs allow to drastically reduce the size, weight and complexity of robotic systems. In fact, their large force-weight ratio, large life cycles, negligible volume, sensing capability and noise-free operation make possible the use of this technology for building a new class of actuation devices. Nonetheless, high power consumption and low bandwidth limit this technology for certain kind of applications. This presents a challenge that must be addressed from both materials and control perspectives in order to overcome these drawbacks. Here, the latter is tackled. It has been demonstrated that suitable control strategies and proper mechanical arrangements can dramatically improve on SMA performance, mostly in terms of actuation speed and limit cycles.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El peso específico de las Comunicaciones Ópticas dentro del ámbito de la Ingeniería de Telecomunicación no cesa de crecer. Sus aplicaciones, inicialmente dedicadas a las grandes líneas que enlazan las centrales de conmutación, alcanzan en la actualidad, como se ha mencionado, hasta los mismos hogares. Los progresos en este campo, con una sucesión sin tregua, no sólo se destinan a incrementar la capacidad de transmisión de los sistemas, sino a ampliar la diversidad de los procesos que sobre las señales se efectúan en el dominio óptico. Este dinamismo demanda a los profesionales del sector una revisión y actualización de sus conocimientos que les permitan resolver con soltura las cuestiones de su actividad de ingeniería. Por otra parte, durante los últimos años la importancia de las Comunicaciones Ópticas también se ha reflejado en las diferentes titulaciones de Ingenierías de Telecomunicación, cuyos planes de estudio contemplan esta materia tanto en asignaturas troncales como optativas. A menudo, las fuentes de información disponibles abordan esta disciplina con una orientación principalmente teórica. Profesionales y estudiantes de Ingeniería, pues, frente a esta materia se encuentran unos temas que tratan fenómenos físicos complejos, abundantes en conceptos abstractos y con un florido aparato matemático, pero muchas veces carentes de una visión práctica, importantísima en ingeniería, y que es, en definitiva, lo que se exige a alumnos e ingenieros: saber resolver problemas y cuestiones relacionados con las Comunicaciones Ópticas. Los sistemas de comunicaciones ópticas, y en especial aquellos que utilizan la fibra óptica como medio para la transmisión de información, como se ha dicho, están alcanzando un desarrollo importante en el campo de las telecomunicaciones. Las bondades que ofrece la fibra, de sobra conocidos y mencionados en el apartado que antecede (gran ancho de banda, inmunidad total a las perturbaciones de origen electromagnético, así como la no producción de interferencias, baja atenuación, etc.), han hecho que, hoy en día, sea uno de los campos de las llamadas tecnologías de la información y la comunicación que presente mayor interés por parte de científicos, ingenieros, operadores de telecomunicaciones y, por supuesto, usuarios. Ante esta realidad, el objetivo y justificación de la realización de este proyecto, por tanto, no es otro que el de acercar esta tecnología al futuro ingeniero de telecomunicaciones, y/o a cualquier persona con un mínimo de interés en este tema, y mostrarle de una forma práctica y visual los diferentes fenómenos que tienen lugar en la transmisión de información por medio de fibra óptica, así como los diferentes bloques y dispositivos en que se divide dicha comunicación. Para conseguir tal objetivo, el proyecto fin de carrera aquí presentado tiene como misión el desarrollo de una interfaz gráfica de usuario (GUI, del inglés Graphic User Interface) que permita a aquel que la utilice configurar de manera sencilla cada uno de los bloques en que se compone un enlace punto a punto de fibra óptica. Cada bloque en que se divide este enlace estará compuesto por varias opciones, que al elegir y configurar como se quiera, hará variar el comportamiento del sistema y presentará al usuario los diferentes fenómenos presentes en un sistema de comunicaciones ópticas, como son el ruido, la dispersión, la atenuación, etc., para una mejor comprensión e interiorización de la teoría estudiada. Por tanto, la aplicación, implementada en MATLAB, fruto de la realización de este PFC pretende servir de complemento práctico para las asignaturas dedicadas al estudio de las comunicaciones ópticas a estudiantes en un entorno amigable e intuitivo. Optical Communications in the field of Telecommunications Engineering continues to grow. Its applications, initially dedicated to large central lines that link the switching currently achieved, as mentioned, to the same household nowadays. Progress in this field, with a relentless succession, not only destined to increase the transmission capacity of the systems, but to broaden the diversity of the processes that are performed on the signals in the optical domain. This demands to professionals reviewing and updating their skills to enable them resolve issues easily. Moreover, in recent years the importance of optical communications is also reflected in the different degrees of Telecommunications Engineering, whose curriculum contemplates this area. Often, the information sources available to tackle this discipline mainly theoretical orientation. Engineering professionals and students are faced this matter are few topics discussing complex physical phenomena, and abstract concepts abundant with a flowery mathematical apparatus, but often wotput a practical, important in engineering, and that is what is required of students and engineers: knowing how to solve problems and issues related to optical communications. Optical communications systems, particularly those using optical fiber as a medium for transmission of information, as stated, are reaching a significant development in the field of telecommunications. The advantages offered by the fiber, well known and referred to in the preceding paragraph (high bandwidth, immunity to electromagnetic disturbances of origin and production of non interference, low attenuation, etc..), have made today, is one of the fields of information and communication technology that this increased interest by scientists, engineers, telecommunications operators and, of course, users. Given this reality, the purpose and justification of this project is not other than to bring this technology to the future telecommunications engineer, and / or anyone with a passing interest in this subject, and showing of a practical and various visual phenomena occurring in the transmission of information by optical fiber, as well as different blocks and devices in which said communication is divided. To achieve that objective, the final project presented here has as its mission the development of a graphical user interface (GUI) that allows the user to configure each of the blocks in which divided a point-to-point optical fiber. Each block into which this link will consist of several options to choose and configure it as you like, this will change the behavior of the system and will present to the user with the different phenomena occurring in an optical communication system, such as noise, dispersion, attenuation, etc., for better understanding and internalization of the theory studied. Therefore, the application, implemented in MATLAB, the result of the completion of the thesis is intended to complement practical subjects for the study of optical communications students in a friendly and intuitive environment.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Energy management has always been recognized as a challenge in mobile systems, especially in modern OS-based mobile systems where multi-functioning are widely supported. Nowadays, it is common for a mobile system user to run multiple applications simultaneously while having a target battery lifetime in mind for a specific application. Traditional OS-level power management (PM) policies make their best effort to save energy under performance constraint, but fail to guarantee a target lifetime, leaving the painful trading off between the total performance of applications and the target lifetime to the user itself. This thesis provides a new way to deal with the problem. It is advocated that a strong energy-aware PM scheme should first guarantee a user-specified battery lifetime to a target application by restricting the average power of those less important applications, and in addition to that, maximize the total performance of applications without harming the lifetime guarantee. As a support, energy, instead of CPU or transmission bandwidth, should be globally managed as the first-class resource by the OS. As the first-stage work of a complete PM scheme, this thesis presents the energy-based fair queuing scheduling, a novel class of energy-aware scheduling algorithms which, in combination with a mechanism of battery discharge rate restricting, systematically manage energy as the first-class resource with the objective of guaranteeing a user-specified battery lifetime for a target application in OS-based mobile systems. Energy-based fair queuing is a cross-application of the traditional fair queuing in the energy management domain. It assigns a power share to each task, and manages energy by proportionally serving energy to tasks according to their assigned power shares. The proportional energy use establishes proportional share of the system power among tasks, which guarantees a minimum power for each task and thus, avoids energy starvation on any task. Energy-based fair queuing treats all tasks equally as one type and supports periodical time-sensitive tasks by allocating each of them a share of system power that is adequate to meet the highest energy demand in all periods. However, an overly conservative power share is usually required to guarantee the meeting of all time constraints. To provide more effective and flexible support for various types of time-sensitive tasks in general purpose operating systems, an extra real-time friendly mechanism is introduced to combine priority-based scheduling into the energy-based fair queuing. Since a method is available to control the maximum time one time-sensitive task can run with priority, the power control and time-constraint meeting can be flexibly traded off. A SystemC-based test-bench is designed to assess the algorithms. Simulation results show the success of the energy-based fair queuing in achieving proportional energy use, time-constraint meeting, and a proper trading off between them. La gestión de energía en los sistema móviles está considerada hoy en día como un reto fundamental, notándose, especialmente, en aquellos terminales que utilizando un sistema operativo implementan múltiples funciones. Es común en los sistemas móviles actuales ejecutar simultaneamente diferentes aplicaciones y tener, para una de ellas, un objetivo de tiempo de uso de la batería. Tradicionalmente, las políticas de gestión de consumo de potencia de los sistemas operativos hacen lo que está en sus manos para ahorrar energía y satisfacer sus requisitos de prestaciones, pero no son capaces de proporcionar un objetivo de tiempo de utilización del sistema, dejando al usuario la difícil tarea de buscar un compromiso entre prestaciones y tiempo de utilización del sistema. Esta tesis, como contribución, proporciona una nueva manera de afrontar el problema. En ella se establece que un esquema de gestión de consumo de energía debería, en primer lugar, garantizar, para una aplicación dada, un tiempo mínimo de utilización de la batería que estuviera especificado por el usuario, restringiendo la potencia media consumida por las aplicaciones que se puedan considerar menos importantes y, en segundo lugar, maximizar las prestaciones globales sin comprometer la garantía de utilización de la batería. Como soporte de lo anterior, la energía, en lugar del tiempo de CPU o el ancho de banda, debería gestionarse globalmente por el sistema operativo como recurso de primera clase. Como primera fase en el desarrollo completo de un esquema de gestión de consumo, esta tesis presenta un algoritmo de planificación de encolado equitativo (fair queueing) basado en el consumo de energía, es decir, una nueva clase de algoritmos de planificación que, en combinación con mecanismos que restrinjan la tasa de descarga de una batería, gestionen de forma sistemática la energía como recurso de primera clase, con el objetivo de garantizar, para una aplicación dada, un tiempo de uso de la batería, definido por el usuario, en sistemas móviles empotrados. El encolado equitativo de energía es una extensión al dominio de la energía del encolado equitativo tradicional. Esta clase de algoritmos asigna una reserva de potencia a cada tarea y gestiona la energía sirviéndola de manera proporcional a su reserva. Este uso proporcional de la energía garantiza que cada tarea reciba una porción de potencia y evita que haya tareas que se vean privadas de recibir energía por otras con un comportamiento más ambicioso. Esta clase de algoritmos trata a todas las tareas por igual y puede planificar tareas periódicas en tiempo real asignando a cada una de ellas una reserva de potencia que es adecuada para proporcionar la mayor de las cantidades de energía demandadas por período. Sin embargo, es posible demostrar que sólo se consigue cumplir con los requisitos impuestos por todos los plazos temporales con reservas de potencia extremadamente conservadoras. En esta tesis, para proporcionar un soporte más flexible y eficiente para diferentes tipos de tareas de tiempo real junto con el resto de tareas, se combina un mecanismo de planificación basado en prioridades con el encolado equitativo basado en energía. En esta clase de algoritmos, gracias al método introducido, que controla el tiempo que se ejecuta con prioridad una tarea de tiempo real, se puede establecer un compromiso entre el cumplimiento de los requisitos de tiempo real y el consumo de potencia. Para evaluar los algoritmos, se ha diseñado en SystemC un banco de pruebas. Los resultados muestran que el algoritmo de encolado equitativo basado en el consumo de energía consigue el balance entre el uso proporcional a la energía reservada y el cumplimiento de los requisitos de tiempo real.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There are many the requirements that modern power converters should fulfill. Most of the applications where these converters are used, demand smaller converters with high efficiency, improved power density and a fast dynamic response. For instance, loads like microprocessors demand aggressive current steps with very high slew rates (100A/mus and higher); besides, during these load steps, the supply voltage of the microprocessor should be kept within tight limits in order to ensure its correct performance. The accomplishment of these requirements is not an easy task; complex solutions like advanced topologies - such as multiphase converters- as well as advanced control strategies are often needed. Besides, it is also necessary to operate the converter at high switching frequencies and to use capacitors with high capacitance and low ESR. Improving the dynamic response of power converters does not rely only on the control strategy but also the power topology should be suited to enable a fast dynamic response. Moreover, in later years, a fast dynamic response does not only mean accomplishing fast load steps but output voltage steps are gaining importance as well. At least, two applications that require fast voltage changes can be named: Low power microprocessors. In these devices, the voltage supply is changed according to the workload and the operating frequency of the microprocessor is changed at the same time. An important reduction in voltage dependent losses can be achieved with such changes. This technique is known as Dynamic Voltage Scaling (DVS). Another application where important energy savings can be achieved by means of changing the supply voltage are Radio Frequency Power Amplifiers. For example, RF architectures based on ‘Envelope Tracking’ and ‘Envelope Elimination and Restoration’ techniques can take advantage of voltage supply modulation and accomplish important energy savings in the power amplifier. However, in order to achieve these efficiency improvements, a power converter with high efficiency and high enough bandwidth (hundreds of kHz or even tens of MHz) is necessary in order to ensure an adequate supply voltage. The main objective of this Thesis is to improve the dynamic response of DC-DC converters from the point of view of the power topology. And the term dynamic response refers both to the load steps and the voltage steps; it is also interesting to modulate the output voltage of the converter with a specific bandwidth. In order to accomplish this, the question of what is it that limits the dynamic response of power converters should be answered. Analyzing this question leads to the conclusion that the dynamic response is limited by the power topology and specifically, by the filter inductance of the converter which is found in series between the input and the output of the converter. The series inductance is the one that determines the gain of the converter and provides the regulation capability. Although the energy stored in the filter inductance enables the regulation and the capability of filtering the output voltage, it imposes a limitation which is the concern of this Thesis. The series inductance stores energy and prevents the current from changing in a fast way, limiting the slew rate of the current through this inductor. Different solutions are proposed in the literature in order to reduce the limit imposed by the filter inductor. Many publications proposing new topologies and improvements to known topologies can be found in the literature. Also, complex control strategies are proposed with the objective of improving the dynamic response in power converters. In the proposed topologies, the energy stored in the series inductor is reduced; examples of these topologies are Multiphase converters, Buck converter operating at very high frequency or adding a low impedance path in parallel with the series inductance. Control techniques proposed in the literature, focus on adjusting the output voltage as fast as allowed by the power stage; examples of these control techniques are: hysteresis control, V 2 control, and minimum time control. In some of the proposed topologies, a reduction in the value of the series inductance is achieved and with this, the energy stored in this magnetic element is reduced; less stored energy means a faster dynamic response. However, in some cases (as in the high frequency Buck converter), the dynamic response is improved at the cost of worsening the efficiency. In this Thesis, a drastic solution is proposed: to completely eliminate the series inductance of the converter. This is a more radical solution when compared to those proposed in the literature. If the series inductance is eliminated, the regulation capability of the converter is limited which can make it difficult to use the topology in one-converter solutions; however, this topology is suitable for power architectures where the energy conversion is done by more than one converter. When the series inductor is eliminated from the converter, the current slew rate is no longer limited and it can be said that the dynamic response of the converter is independent from the switching frequency. This is the main advantage of eliminating the series inductor. The main objective, is to propose an energy conversion strategy that is done without series inductance. Without series inductance, no energy is stored between the input and the output of the converter and the dynamic response would be instantaneous if all the devices were ideal. If the energy transfer from the input to the output of the converter is done instantaneously when a load step occurs, conceptually it would not be necessary to store energy at the output of the converter (no output capacitor COUT would be needed) and if the input source is ideal, the input capacitor CIN would not be necessary. This last feature (no CIN with ideal VIN) is common to all power converters. However, when the concept is actually implemented, parasitic inductances such as leakage inductance of the transformer and the parasitic inductance of the PCB, cannot be avoided because they are inherent to the implementation of the converter. These parasitic elements do not affect significantly to the proposed concept. In this Thesis, it is proposed to operate the converter without series inductance in order to improve the dynamic response of the converter; however, on the other side, the continuous regulation capability of the converter is lost. It is said continuous because, as it will be explained throughout the Thesis, it is indeed possible to achieve discrete regulation; a converter without filter inductance and without energy stored in the magnetic element, is capable to achieve a limited number of output voltages. The changes between these output voltage levels are achieved in a fast way. The proposed energy conversion strategy is implemented by means of a multiphase converter where the coupling of the phases is done by discrete two-winding transformers instead of coupledinductors since transformers are, ideally, no energy storing elements. This idea is the main contribution of this Thesis. The feasibility of this energy conversion strategy is first analyzed and then verified by simulation and by the implementation of experimental prototypes. Once the strategy is proved valid, different options to implement the magnetic structure are analyzed. Three different discrete transformer arrangements are studied and implemented. A converter based on this energy conversion strategy would be designed with a different approach than the one used to design classic converters since an additional design degree of freedom is available. The switching frequency can be chosen according to the design specifications without penalizing the dynamic response or the efficiency. Low operating frequencies can be chosen in order to favor the efficiency; on the other hand, high operating frequencies (MHz) can be chosen in order to favor the size of the converter. For this reason, a particular design procedure is proposed for the ‘inductorless’ conversion strategy. Finally, applications where the features of the proposed conversion strategy (high efficiency with fast dynamic response) are advantageus, are proposed. For example, in two-stage power architectures where a high efficiency converter is needed as the first stage and there is a second stage that provides the fine regulation. Another example are RF power amplifiers where the voltage is modulated following an envelope reference in order to save power; in this application, a high efficiency converter, capable of achieving fast voltage steps is required. The main contributions of this Thesis are the following: The proposal of a conversion strategy that is done, ideally, without storing energy in the magnetic element. The validation and the implementation of the proposed energy conversion strategy. The study of different magnetic structures based on discrete transformers for the implementation of the proposed energy conversion strategy. To elaborate and validate a design procedure. To identify and validate applications for the proposed energy conversion strategy. It is important to remark that this work is done in collaboration with Intel. The particular features of the proposed conversion strategy enable the possibility of solving the problems related to microprocessor powering in a different way. For example, the high efficiency achieved with the proposed conversion strategy enables it as a good candidate to be used for power conditioning, as a first stage in a two-stage power architecture for powering microprocessors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper discusses the use of sound waves to illustrate multipath radio propagation concepts. Specifically, a procedure is presented to measure the time-varying frequency response of the channel. This helps demonstrate how a propagation channel can be characterized in time and frequency, and provides visualizations of the concepts of coherence time and coherence bandwidth. The measurements are very simple to carry out, and the required equipment is easily available. The proposed method can be useful for wireless or mobile communication courses.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the last years, RF power amplifiers are taking advantage of the switched dc-dc converters to use them in several architectures that may improve the efficiency of the amplifier, keeping a good linearity. The use of linearization techniques such as Envelope Elimination and Restoration(EER) and Envelope Tracking (ET) requires a very fast dc-dc power converter to provide variable voltage supply to the power amplifier but theoretically the efficiency can be much higher than using the classical amplifiers belonging to classes A, B or AB. The purpose of this paper is to analyze the state of the art of the power converters used as envelope amplifiers in this application. The power topologies will be explored and several important parameters such as efficiency, bandwidth will be discussed.