812 resultados para Consumer electronics
Resumo:
This paper presents a novel background modeling system that uses a spatial grid of Support Vector Machines classifiers for segmenting moving objects, which is a key step in many video-based consumer applications. The system is able to adapt to a large range of dynamic background situations since no parametric model or statistical distribution are assumed. This is achieved by using a different classifier per image region that learns the specific appearance of that scene region and its variations (illumination changes, dynamic backgrounds, etc.). The proposed system has been tested with a recent public database, outperforming other state-of-the-art algorithms.
Quality-optimization algorithm based on stochastic dynamic programming for MPEG DASH video streaming
Resumo:
In contrast to traditional push-based protocols, adaptive streaming techniques like Dynamic Adaptive Streaming over HTTP (DASH) fix attention on the client, who dynamically requests different-quality portions of the content to cope with a limited and variable bandwidth but aiming at maximizing the quality perceived by the user. Since DASH adaptation logic at the client is not covered by the standard, we propose a solution based on Stochastic Dynamic Programming (SDP) techniques to find the optimal request policies that guarantee the users' Quality of Experience (QoE). Our algorithm is evaluated in a simulated streaming session and is compared with other adaptation approaches. The results show that our proposal outperforms them in terms of QoE, requesting higher qualities on average.
Resumo:
A frame-level distortion model based on perceptual features of the human visual system is proposed to improve the performance of unequal error protection strategies and provide better quality of experience to users in Side-by-Side 3D video delivery systems.
Resumo:
This paper gives an overview of three recent studies by the authors on the topic of 3D video Quality of Experience (QoE). Two of studies [1,2] investigated different psychological dimension that may be needed for describing 3D video QoE and the third the visibility and annoyance of crosstalk[3]. The results shows that the video quality scale could be sufficient for evaluating S3D video experience for coding and spatial resolution reduction distortions. It was also confirmed that with a more complex mixture of degradations more than one scale should be used to capture the QoE in these cases. The study found a linear relationship between the perceived crosstalk and the amount of crosstalk.
Resumo:
This paper presents a registration method for images with global illumination variations. The method is based on a joint iterative optimization (geometric and photometric) of the L1 norm of the intensity error. Two strategies are compared to directly find the appropriate intensity transformation within each iteration: histogram specification and the solution obtained by analyzing the necessary optimality conditions. Such strategies reduce the search space of the joint optimization to that of the geometric transformation between the images.
Resumo:
A novel and high-quality system for moving object detection in sequences recorded with moving cameras is proposed. This system is based on the collaboration between an automatic homography estimation module for image alignment, and a robust moving object detection using an efficient spatiotemporal nonparametric background modeling.
Resumo:
Video Quality Assessment needs to correspond to human perception. Pixel-based metrics (PSNR or MSE) fail in many circumstances for not taking into account the spatio-temporal property of human's visual perception. In this paper we propose a new pixel-weighted method to improve video quality metrics for artifacts evaluation. The method applies a psychovisual model based on motion, level of detail, pixel location and the appearance of human faces, which approximate the quality to the human eye's response. Subjective tests were developed to adjust the psychovisual model for demonstrating the noticeable improvement of an algorithm when weighting the pixels according to the factors analyzed instead of treating them equally. The analysis developed demonstrates the necessity of models adapted to the specific visualization of contents and the model presents an advance in quality to be applied over sequences when a determined artifact is analyzed.
Resumo:
El control, o cancelación activa de ruido, consiste en la atenuación del ruido presente en un entorno acústico mediante la emisión de una señal igual y en oposición de fase al ruido que se desea atenuar. La suma de ambas señales en el medio acústico produce una cancelación mutua, de forma que el nivel de ruido resultante es mucho menor al inicial. El funcionamiento de estos sistemas se basa en los principios de comportamiento de los fenómenos ondulatorios descubiertos por Augustin-Jean Fresnel, Christiaan Huygens y Thomas Young entre otros. Desde la década de 1930, se han desarrollado prototipos de sistemas de control activo de ruido, aunque estas primeras ideas eran irrealizables en la práctica o requerían de ajustes manuales cada poco tiempo que hacían inviable su uso. En la década de 1970, el investigador estadounidense Bernard Widrow desarrolla la teoría de procesado adaptativo de señales y el algoritmo de mínimos cuadrados LMS. De este modo, es posible implementar filtros digitales cuya respuesta se adapte de forma dinámica a las condiciones variables del entorno. Con la aparición de los procesadores digitales de señal en la década de 1980 y su evolución posterior, se abre la puerta para el desarrollo de sistemas de cancelación activa de ruido basados en procesado de señal digital adaptativo. Hoy en día, existen sistemas de control activo de ruido implementados en automóviles, aviones, auriculares o racks de equipamiento profesional. El control activo de ruido se basa en el algoritmo fxlms, una versión modificada del algoritmo LMS de filtrado adaptativo que permite compensar la respuesta acústica del entorno. De este modo, se puede filtrar una señal de referencia de ruido de forma dinámica para emitir la señal adecuada que produzca la cancelación. Como el espacio de cancelación acústica está limitado a unas dimensiones de la décima parte de la longitud de onda, sólo es viable la reducción de ruido en baja frecuencia. Generalmente se acepta que el límite está en torno a 500 Hz. En frecuencias medias y altas deben emplearse métodos pasivos de acondicionamiento y aislamiento, que ofrecen muy buenos resultados. Este proyecto tiene como objetivo el desarrollo de un sistema de cancelación activa de ruidos de carácter periódico, empleando para ello electrónica de consumo y un kit de desarrollo DSP basado en un procesador de muy bajo coste. Se han desarrollado una serie de módulos de código para el DSP escritos en lenguaje C, que realizan el procesado de señal adecuado a la referencia de ruido. Esta señal procesada, una vez emitida, produce la cancelación acústica. Empleando el código implementado, se han realizado pruebas que generan la señal de ruido que se desea eliminar dentro del propio DSP. Esta señal se emite mediante un altavoz que simula la fuente de ruido a cancelar, y mediante otro altavoz se emite una versión filtrada de la misma empleando el algoritmo fxlms. Se han realizado pruebas con distintas versiones del algoritmo, y se han obtenido atenuaciones de entre 20 y 35 dB medidas en márgenes de frecuencia estrechos alrededor de la frecuencia del generador, y de entre 8 y 15 dB medidas en banda ancha. ABSTRACT. Active noise control consists on attenuating the noise in an acoustic environment by emitting a signal equal but phase opposed to the undesired noise. The sum of both signals results in mutual cancellation, so that the residual noise is much lower than the original. The operation of these systems is based on the behavior principles of wave phenomena discovered by Augustin-Jean Fresnel, Christiaan Huygens and Thomas Young. Since the 1930’s, active noise control system prototypes have been developed, though these first ideas were practically unrealizable or required manual adjustments very often, therefore they were unusable. In the 1970’s, American researcher Bernard Widrow develops the adaptive signal processing theory and the Least Mean Squares algorithm (LMS). Thereby, implementing digital filters whose response adapts dynamically to the variable environment conditions, becomes possible. With the emergence of digital signal processors in the 1980’s and their later evolution, active noise cancellation systems based on adaptive signal processing are attained. Nowadays active noise control systems have been successfully implemented on automobiles, planes, headphones or racks for professional equipment. Active noise control is based on the fxlms algorithm, which is actually a modified version of the LMS adaptive filtering algorithm that allows compensation for the acoustic response of the environment. Therefore it is possible to dynamically filter a noise reference signal to obtain the appropriate cancelling signal. As the noise cancellation space is limited to approximately one tenth of the wavelength, noise attenuation is only viable for low frequencies. It is commonly accepted the limit of 500 Hz. For mid and high frequencies, conditioning and isolating passive techniques must be used, as they produce very good results. The objective of this project is to develop a noise cancellation system for periodic noise, by using consumer electronics and a DSP development kit based on a very-low-cost processor. Several C coded modules have been developed for the DSP, implementing the appropriate signal processing to the noise reference. This processed signal, once emitted, results in noise cancellation. The developed code has been tested by generating the undesired noise signal in the DSP. This signal is emitted through a speaker simulating the noise source to be removed, and another speaker emits an fxlms filtered version of the same signal. Several versions of the algorithm have been tested, obtaining attenuation levels around 20 – 35 dB measured in a tight bandwidth around the generator frequency, or around 8 – 15 dB measured in broadband.
Resumo:
A televisão nos dias atuais tem sofrido inúmeras inovações tecnológicas nos campos das transmissões multimídia, qualidade audio-visual e diversidade de funcionalidades. Entretanto, esta essencialmente mantêm sua característica de fornecer informações de forma quase que instantânea à população. O ambiente atual da televisão digital é caracterizado pela coexistência de inúmeros dispositivos capazes de oferecerem uma experiência televisa, associando-se computadores pessoais, smartphones, tablets e outros eletrônicos de consumo. Ainda, pode se incluir a este cenário a disponibilidade de inúmeras redes de transporte de dados tais como a radiodifusão, satélite, cabo e redes em banda larga. Este cenário diversificado, em termos de dispositivos e redes, é denominado de cenário de televisão digital híbrida, a qual destaca-se a interação do expectador com os diversos dispositivos. Estes cenários, por sua vez, motivam o desenvolvimento de tecnologias que permitem o aperfeiçoamento da pervasividade e dos meios pelos os quais os aplicativos possam ser suportados em diferentes plataformas. Este trabalho propõe ambientes interoperáveis envolvendo a televisão digital interativa e outros eletrônicos de consumo, aos quais foram realizados estudos e experimentos para se observar diferentes técnicas de sincronização e comunicação entre plataformas de interatividade para a televisão digital híbrida. Os resultados apontam para a possibilidade de cenários interoperáveis envolvendo o uso de marcadores e também recursos de redes e serviços TCP/IP, levando em consideração a eficiência e eficácia nos diferentes métodos. Conclui-se que os resultados odem motivar o desenvolvimento de cenários diferenciados envolvendo a televisão digital interativa e dispositivos de segunda tela, o que incrementa a interatividade e as formas de entretenimento.
Resumo:
O aumento no consumo mundial de novos aparelhos eletroeletrônicos aliado à redução no tempo de vida útil destes equipamentos tem como principal consequência ao meio ambiente a geração de resíduos. No Brasil, com a instituição da Política Nacional de Resíduos Sólidos, criou-se a obrigatoriedade legal da responsabilidade dos fabricantes pela logística reversa dos equipamentos eletroeletrônicos, incentivando pesquisas para o desenvolvimento dos métodos de reciclagem e tratamento dos materiais descartados. O processo de lixiviação foi avaliado como alternativa à etapa de separação magnética presente nas atuais rotas hidrometalúrgicas para recuperação de metais valiosos de placas de circuito impresso. Para avaliar a composição das placas, foi realizado ensaio de dissolução em água régia. As amostras foram moídas e submetidas a ensaios de lixiviação com ácido sulfúrico nas concentrações de 1 e 2mol/L, às temperaturas de 75ºC, 85ºC e 95ºC, durante 24 horas. Com ácido sulfúrico 2mol/L a 95ºC, o tempo necessário para se obter 100% de extração do ferro foi de 2 horas. Nestas condições, não foi detectada a presença de cobre dissolvido. A cinética da reação é controlada por reação química e obedece a equação .=1(1)3. A energia de ativação aparente do processo equivale a 90kJ/mol.
Resumo:
This paper examines the role of creative resources in the emergence of the Japanese video game industry. We argue that creative resources nurtured by popular cartoons and animation sector, combined with technological knowledge accumulated in the consumer electronics industry, facilitated the emergence of successful video game industry in Japan. First we trace the development of the industry from its origin to the rise of platform developers and software publishers. Then, knowledge and creative foundations that influenced the developmental trajectory of this industry are analyzed, with links to consumer electronics and in regards to cartoons and animation industry.
Resumo:
In this paper we explore the interrelationship between technological progress and the formation of industry-specific skills by analysing the evolution of the video-game industry in three countries: Japan, the United States, and the United Kingdom. We argue that the cross-sectoral transfer of skills occurs differently depending on national contexts, such as the social legitimacy and strength of preexisting industries, the socioeconomic status of entrepreneurs or pioneer firms in an emerging industry, and the sociocultural cohesiveness between the preexisting and emerging industries. Each country draws on a different set of creative resources, which results in a unique trajectory. Whereas Japan's video-game industry emerged out of corporate sponsorships in arcades, toys, and consumer electronics industries and drew skills from the comic book and animated-film sectors, the video-game industry in the United States evolved from arcades and personal computers. In the United Kingdom the video-game industry developed bottom-up, through a process of skills formation in the youth culture of 'bedroom coders' that nurtured self-taught programmers in their teens throughout the country.
Resumo:
Today’s business leaders must constantly review and develop their firm’s abilities to adapt to and benefit from external changes. Dynamic capabilities are the capacity of an organization to purposefully create, extend or modify its resource base. They enable it to exploit business, technological and market opportunities and adapt to market changes, an ability more often observed in highly dynamic industries, such as consumer electronics or telecommunications. Using the case study method, this article identifies dynamic capabilities in traditional, less dynamic industries when faced with a sudden drop of revenue. Four distinct routines emerge, namely structure and practices enduring time-sensitive strategic decision-making by the tice, and a culture encouraging learning and coevolving. Seemingly strategic paradox objectives encourage the management team to question the status quo and, when managed well, transform the tensions between old and new into an ability to advance superior ideas faster.
Resumo:
The intent of this work was to develop a mobile robotic platform that was controlled by a Palm Pilot PDA. Advances in consumer electronics are producing powerful yet small handheld devices. Some of these devices present quasi-PC capabilities for a fraction of the cost; furthermore, they are compact enough that they fit in all but the smallest of platforms. The platform prototype built for testing purposes has a differential-drive configuration to provide simple but agile movement control. The sensor package consisted of two infrared ranging sensors mounted on servomotors that provide a wide area of detection. Building such a platform involved selection of hardware, circuit integration and software development. The software suite selected to develop code for the Palm Pilot was CodeWarrior, a C compiler that can generate code in Palm-native PRC files.
Resumo:
A parallel pipelined array of cells suitable for realtime computation of histograms is proposed. The cell architecture builds on previous work to now allow operating on a stream of data at 1 pixel per clock cycle. This new cell is more suitable for interfacing to camera sensors or to microprocessors of 8-bit data buses which are common in consumer digital cameras. Arrays using the new proposed cells are obtained via C-slow retiming techniques and can be clocked at a 65% faster frequency than previous arrays. This achieves over 80% of the performance of two-pixel per clock cycle parallel pipelined arrays.