795 resultados para Annotated Video Clips
Resumo:
Our research investigates the impact that hearing has on the perception of digital video clips, with and without captions, by discussing how hearing loss, captions and deafness type affects user QoP (Quality of Perception). QoP encompasses not only a user's satisfaction with the quality of a multimedia presentation, but also their ability to analyse, synthesise and assimilate informational content of multimedia . Results show that hearing has a significant effect on participants’ ability to assimilate information, independent of video type and use of captions. It is shown that captions do not necessarily provide deaf users with a ‘greater level of information’ from video, but cause a change in user QoP, depending on deafness type, which provides a ‘greater level of context of the video’. It is also shown that post-lingual mild and moderately deaf participants predict less accurately their level of information assimilation than post-lingual profoundly deaf participants, despite residual hearing. A positive correlation was identified between level of enjoyment (LOE) and self-predicted level of information assimilation (PIA), independent of hearing level or hearing type. When this is considered in a QoP quality framework, it puts into question how the user perceives certain factors, such as ‘informative’ and ‘quality’.
Resumo:
We investigate the impact of captions on deaf and hearing perception of multimedia video clips. We measure perception using a parameter called Quality of Perception (QoP), which encompasses not only a user's satisfaction with multimedia clips, but also his/her ability to perceive, synthesise and analyse the informational content of such presentations. By studying perceptual diversity, it is our aim to identify trends that will help future implementation of adaptive multimedia technologies. Results show that although hearing level has a significant affect on information assimilation, the effect of captions is not significant on the objective level of information assimilated. Deaf participants predict that captions significantly improve their level of information assimilation, although no significant objective improvement was measured. The level of enjoyment is unaffected by a participant’s level of hearing or use of captions.
Resumo:
Maternal dissociative symptoms which can be comorbid with interpersonal violence-related post-traumatic stress disorder (IPV-PTSD) have been linked to decreased sensitivity and responsiveness to children's emotional communication. This study examined the influence of dissociation on neural activation independently of IPV-PTSD symptom severity when mothers watch video-stimuli of their children during stressful and non-stressful mother-child interactions. Based on previous observations in related fields, we hypothesized that more severe comorbid dissociation in IPV-PTSD would be associated with lower limbic system activation and greater neural activity in regions of the emotion regulation circuit such as the medial prefrontal cortex and dorsolateral prefrontal cortex (dlPFC). Twenty mothers (of children aged 12-42 months), with and without IPV-PTSD watched epochs showing their child during separation and play while undergoing functional magnetic resonance imaging (fMRI). Multiple regression indicated that when mothers diagnosed with IPV-PTSD watched their children during separation compared to play, dissociative symptom severity was indeed linked to lowered activation within the limbic system, while greater IPV-PTSD symptom severity was associated with heightened limbic activity. Concerning emotion regulation areas, there was activation associated to dissociation in the right dlPFC. Our results are likely a neural correlate of affected mothers' reduced capacity for sensitive responsiveness to their young child following exposure to interpersonal stress, situations that are common in day-to-day parenting.
Resumo:
Digital still cameras capable of filming short video clips are readily available, but the quality of these recordings for telemedicine has not been reported. We performed a blinded study using four commonly available digital cameras. A simulated patient with a hemiplegic gait pattern was filmed by the same videographer in an identical, brightly lit indoor setting. Six neurologists viewed the blinded video clips on their PC and comparisons were made between cameras, between video clips recorded with and without a tripod, and between video clips filmed on high- or low-quality settings. Use of a tripod had a smaller effect than expected, while images taken on a high-quality setting were strongly preferred to those taken on a low-quality setting. Although there was some variability in video quality between selected cameras, all were of sufficient quality to identify physical signs such as gait and tremor. Adequate-quality video clips of movement disorders can be produced with low-cost cameras and transmitted by email for teleneurology purposes.
Resumo:
The usage of digital content, such as video clips and images, has increased dramatically during the last decade. Local image features have been applied increasingly in various image and video retrieval applications. This thesis evaluates local features and applies them to image and video processing tasks. The results of the study show that 1) the performance of different local feature detector and descriptor methods vary significantly in object class matching, 2) local features can be applied in image alignment with superior results against the state-of-the-art, 3) the local feature based shot boundary detection method produces promising results, and 4) the local feature based hierarchical video summarization method shows promising new new research direction. In conclusion, this thesis presents the local features as a powerful tool in many applications and the imminent future work should concentrate on improving the quality of the local features.
Resumo:
Seminario realizado por cuatro profesoras de distintos centros educativos de Vizcaya para la elaboración de cinco 'projects', cada uno con ejercicios para diferentes niveles, que tienen como objetivo inmediato la práctica de estructuras del lenguaje en un marco comunicativo y como objetivo global la realización de una filmación sobre el tema básico del project. Los temas tratados son: el uso de la cámara de vídeo, las noticias, los anuncios (publicidad), video clips, cuentos y narraciones. Los resultados se valoran como muy positivos debido a la motivación de filmar el producto final.
Resumo:
Automatic indexing and retrieval of digital data poses major challenges. The main problem arises from the ever increasing mass of digital media and the lack of efficient methods for indexing and retrieval of such data based on the semantic content rather than keywords. To enable intelligent web interactions, or even web filtering, we need to be capable of interpreting the information base in an intelligent manner. For a number of years research has been ongoing in the field of ontological engineering with the aim of using ontologies to add such (meta) knowledge to information. In this paper, we describe the architecture of a system (Dynamic REtrieval Analysis and semantic metadata Management (DREAM)) designed to automatically and intelligently index huge repositories of special effects video clips, based on their semantic content, using a network of scalable ontologies to enable intelligent retrieval. The DREAM Demonstrator has been evaluated as deployed in the film post-production phase to support the process of storage, indexing and retrieval of large data sets of special effects video clips as an exemplar application domain. This paper provides its performance and usability results and highlights the scope for future enhancements of the DREAM architecture which has proven successful in its first and possibly most challenging proving ground, namely film production, where it is already in routine use within our test bed Partners' creative processes. (C) 2009 Published by Elsevier B.V.
Resumo:
One of critical challenges in automatic recognition of TV commercials is to generate a unique, robust and compact signature. Uniqueness indicates the ability to identify the similarity among the commercial video clips which may have slight content variation. Robustness means the ability to match commercial video clips containing the same content but probably with different digitalization/encoding, some noise data, and/or transmission and recording distortion. Efficiency is about the capability of effectively matching commercial video sequences with a low computation cost and storage overhead. In this paper, we present a binary signature based method, which meets all the three criteria above, by combining the techniques of ordinal and color measurements. Experimental results on a real large commercial video database show that our novel approach delivers a significantly better performance comparing to the existing methods.
Resumo:
Video streaming via Transmission Control Protocol (TCP) networks has become a popular and highly demanded service, but its quality assessment in both objective and subjective terms has not been properly addressed. In this paper, based on statistical analysis a full analytic model of a no-reference objective metric, namely pause intensity (PI), for video quality assessment is presented. The model characterizes the video playout buffer behavior in connection with the network performance (throughput) and the video playout rate. This allows for instant quality measurement and control without requiring a reference video. PI specifically addresses the need for assessing the quality issue in terms of the continuity in the playout of TCP streaming videos, which cannot be properly measured by other objective metrics such as peak signal-to-noise-ratio, structural similarity, and buffer underrun or pause frequency. The performance of the analytical model is rigidly verified by simulation results and subjective tests using a range of video clips. It is demonstrated that PI is closely correlated with viewers' opinion scores regardless of the vastly different composition of individual elements, such as pause duration and pause frequency which jointly constitute this new quality metric. It is also shown that the correlation performance of PI is consistent and content independent. © 2013 IEEE.
Resumo:
In this paper a full analytic model for pause intensity (PI), a no-reference metric for video quality assessment, is presented. The model is built upon the video play out buffer behavior at the client side and also encompasses the characteristics of a TCP network. Video streaming via TCP produces impairments in play continuity, which are not typically reflected in current objective metrics such as PSNR and SSIM. Recently the buffer under run frequency/probability has been used to characterize the buffer behavior and as a measurement for performance optimization. But we show, using subjective testing, that under run frequency cannot reflect the viewers' quality of experience for TCP based streaming. We also demonstrate that PI is a comprehensive metric made up of a combination of phenomena observed in the play out buffer. The analytical model in this work is verified with simulations carried out on ns-2, showing that the two results are closely matched. The effectiveness of the PI metric has also been proved by subjective testing on a range of video clips, where PI values exhibit a good correlation with the viewers' opinion scores. © 2012 IEEE.
Resumo:
In this work, we investigate a new objective measurement for assessing the video playback quality for services delivered in networks that use TCP as a transport layer protocol. We define the new metric as pause intensity to characterize the quality of playback in terms of its continuity since, in the case of TCP, data packets are protected from losses but not from delays. Using packet traces generated from real TCP connections in a lossy environment, we are able to simulate the playback of a video and monitor buffer behaviors in order to calculate pause intensity values. We also run subjective tests to verify the effectiveness of the metric introduced and show that the results of pause intensity and the subjective scores made over the same real video clips are closely correlated.
Resumo:
We demonstrate tomographic imaging of the refractive index of turbid media using bifocal optical coherence refractometry (BOCR). The technique, which is a variant of optical coherence tomography, is based on the measurement of the optical pathlength difference between two foci simultaneously present in a medium of interest. We describe a new method to axially shift the bifocal optical pathlength that avoids the need to physically relocate the objective lens or the sample during an axial scan, and present an experimental realization based on an adaptive liquid-crystal lens. We present experimental results, including video clips, which demonstrate refractive index tomography of a range of turbid liquid phantoms, as well as of human skin in vivo.