975 resultados para Visual word recognition
Resumo:
Hand gestures are a powerful way for human communication, with lots of potential applications in the area of human computer interaction. Vision-based hand gesture recognition techniques have many proven advantages compared with traditional devices, giving users a simpler and more natural way to communicate with electronic devices. This work proposes a generic system architecture based in computer vision and machine learning, able to be used with any interface for human-computer interaction. The proposed solution is mainly composed of three modules: a pre-processing and hand segmentation module, a static gesture interface module and a dynamic gesture interface module. The experiments showed that the core of visionbased interaction systems could be the same for all applications and thus facilitate the implementation. For hand posture recognition, a SVM (Support Vector Machine) model was trained and used, able to achieve a final accuracy of 99.4%. For dynamic gestures, an HMM (Hidden Markov Model) model was trained for each gesture that the system could recognize with a final average accuracy of 93.7%. The proposed solution as the advantage of being generic enough with the trained models able to work in real-time, allowing its application in a wide range of human-machine applications. To validate the proposed framework two applications were implemented. The first one is a real-time system able to interpret the Portuguese Sign Language. The second one is an online system able to help a robotic soccer game referee judge a game in real time.
Resumo:
Vision-based hand gesture recognition is an area of active current research in computer vision and machine learning. Being a natural way of human interaction, it is an area where many researchers are working on, with the goal of making human computer interaction (HCI) easier and natural, without the need for any extra devices. So, the primary goal of gesture recognition research is to create systems, which can identify specific human gestures and use them, for example, to convey information. For that, vision-based hand gesture interfaces require fast and extremely robust hand detection, and gesture recognition in real time. Hand gestures are a powerful human communication modality with lots of potential applications and in this context we have sign language recognition, the communication method of deaf people. Sign lan- guages are not standard and universal and the grammars differ from country to coun- try. In this paper, a real-time system able to interpret the Portuguese Sign Language is presented and described. Experiments showed that the system was able to reliably recognize the vowels in real-time, with an accuracy of 99.4% with one dataset of fea- tures and an accuracy of 99.6% with a second dataset of features. Although the im- plemented solution was only trained to recognize the vowels, it is easily extended to recognize the rest of the alphabet, being a solid foundation for the development of any vision-based sign language recognition user interface system.
Resumo:
In this paper, we present an integrated system for real-time automatic detection of human actions from video. The proposed approach uses the boundary of humans as the main feature for recognizing actions. Background subtraction is performed using Gaussian mixture model. Then, features are extracted from silhouettes and Vector Quantization is used to map features into symbols (bag of words approach). Finally, actions are detected using the Hidden Markov Model. The proposed system was validated using a newly collected real- world dataset. The obtained results show that the system is capable of achieving robust human detection, in both indoor and outdoor environments. Moreover, promising classification results were achieved when detecting two basic human actions: walking and sitting.
Resumo:
Biometric systems are increasingly being used as a means for authentication to provide system security in modern technologies. The performance of a biometric system depends on the accuracy, the processing speed, the template size, and the time necessary for enrollment. While much research has focused on the first three factors, enrollment time has not received as much attention. In this work, we present the findings of our research focused upon studying user’s behavior when enrolling in a biometric system. Specifically, we collected information about the user’s availability for enrollment in respect to the hand recognition systems (e.g., hand geometry, palm geometry or any other requiring positioning the hand on an optical scanner). A sample of 19 participants, chosen randomly apart their age, gender, profession and nationality, were used as test subjects in an experiment to study the patience of users enrolling in a biometric hand recognition system.
Resumo:
METHODS: Refractive lens exchange was performed with implantation of an AT Lisa 839M (trifocal) or 909MP (bifocal toric) IOL, the latter if corneal astigmatism was more than 0.75 diopter (D). The postoperative visual and refractive outcomes were evaluated. A prototype light-distortion analyzer was used to quantify the postoperative light-distortion indices. A control group of eyes in which a Tecnis ZCB00 1-piece monofocal IOL was implanted had the same examinations. RESULTS: A trifocal or bifocal toric IOL was implanted in 66 eyes. The control IOL was implanted in 18 eyes. All 3 groups obtained a significant improvement in uncorrected distance visual acuity (UDVA) (P < .001) and corrected distance visual acuity (CDVA) (P Z .001). The mean uncorrected near visual acuity (UNVA) was 0.123 logMAR with the trifocal IOL and 0.130 logMAR with the bifocal toric IOL. The residual refractive cylinder was less than 1.00 D in 86.7% of cases with the toric IOL. The mean light-distortion index was significantly higher in the multifocal IOL groups than in the monofocal group (P < .001), although no correlation was found between the light-distortion index and CDVA. CONCLUSIONS: The multifocal IOLs provided excellent UDVA and functional UNVA despite increased light-distortion indices. The light-distortion analyzer reliably quantified a subjective component of vision distinct from visual acuity; it may become a useful adjunct in the evaluation of visual quality obtained with multifocal IOLs.
Resumo:
Alzheimer's disease (AD) is commonly associated with marked memory deficits; however, nonamnestic variants have been consistently described as well. Posterior cortical atrophy (PCA) is a progressive degenerative condition in which posterior regions of the brain are predominantly affected, therefore resulting in a pattern of distinctive and marked visuospatial symptoms, such as apraxia, alexia, and spatial neglect. Despite the growing number of studies on cognitive and neural bases of the visual variant of AD, intervention studies remain relatively sparse. Current pharmacological treatments offer modest efficacy. Also, there is a scarcity of complementary nonpharmacological interventions with only two previous studies of PCA. Here we describe a highly educated 57-year-old patient diagnosed with a visual variant of AD who participated in a cognitive intervention program (comprising reality orientation, cognitive stimulation, and cognitive training exercises). Neuropsychological assessment was performed across moments (baseline, postintervention, follow-up) and consisted mainly of verbal and visual memory. Baseline neuropsychological assessment showed deficits in perceptive and visual-constructive abilities, learning and memory, and temporal orientation. After neuropsychological rehabilitation, we observed small improvements in the patient's cognitive functioning, namely in verbal memory, attention, and psychomotor abilities. This study shows evidence of small beneficial effects of cognitive intervention in PCA and is the first report of this approach with a highly educated patient in a moderate stage of the disease. Controlled studies are needed to assess the potential efficacy of cognition-focused approaches in these patients, and, if relevant, to grant their availability as a complementary therapy to pharmacological treatment and visual aids.
Resumo:
It is well known that color coding facilitates search and iden- tification in real-life tasks. The aim of this work was to compare reac- tion times for normal color and dichromatic observers in a visual search experiment. A unique distracter color was used to avoid abnormal color vision vulnerability to background complexity. Reaction times for nor- mal color observers and dichromats were estimated for 2◦ central vision at 48 directions around a white point in CIE L∗a∗b∗ color space for systematic examination on the mechanisms of dichromatic color percep- tion. The results show that mean search times for dichromats were twice larger compared to the normal color observers and for all directions. The difference between the copunctual confusion lines and the confusion direction measure experimentally was 5.5◦ for protanopes and 7.5◦ for deuteranopes.
Resumo:
OBJETIVO: Comparar o desempenho dos tipos de TDAH em tarefas de seletividade, sustentação e atenção alternada, considerando tempo médio de reação, número de erros por ação e número de erros por omissão em teste computadorizado de atenção visual (TAVIS-III). MÉTODOS: Cento e duas crianças e adolescentes de duas escolas particulares e uma escola pública da cidade do Rio de Janeiro, com idades entre 6 e 17 anos com diagnóstico de TDAH, segundo DSM-IV, foram submetidas ao TAVIS-III. A separação por grupos respeitou os tipos determinados por entrevista semi-estruturada (P-CHIPS), e os resultados foram comparados levando-se em consideração o fator grupo. RESULTADOS: O tipo combinado (C) revelou-se o mais comum (n = 65; 63,7%), seguido pelo predominantemente desatento (D) (n = 32; 31,4%). O tipo predominantemente hiperativo-impulsivo (HI) foi excluído das análises estatísticas em virtude da baixa freqüência. O desempenho do grupo C revelou-se inferior apenas em tarefa de atenção sustentada, no que tange ao número de erros por ação e tempo médio de reação (p < 0,05), ambos maiores quando comparados ao grupo D. CONCLUSÃO: Apenas em tarefa de atenção visual por tempo mais prolongado, o tipo combinado revelou diferenças significativas em relação ao grupo desatento.
Resumo:
Dissertação de mestrado em Engenharia e Gestão da Qualidade
Resumo:
OBJECTIVE: Compare pattern of exploratory eye movements during visual scanning of the Rorschach and TAT test cards in people with schizophrenia and controls. METHOD: 10 participants with schizophrenia and 10 controls matched by age, schooling and intellectual level participated in the study. Severity of symptoms was evaluated with the Positive and Negative Syndrome Scale. Test cards were divided into three groups: TAT cards with scenes content, TAT cards with interaction content (TAT-faces), and Rorschach cards with abstract images. Eye movements were analyzed for: total number, duration and location of fixation; and length of saccadic movements. RESULTS: Different pattern of eye movement was found, with schizophrenia participants showing lower number of fixations but longer fixation duration in Rorschach cards and TAT-faces. The biggest difference was observed in Rorschach, followed by TAT-faces and TAT-scene cards. CONCLUSIONS: Results suggest alteration in visual exploration mechanisms possibly related to integration of abstract visual information.
Resumo:
Dissertação de mestrado integrado em Engenharia e Gestão de Sistemas de Informação
Resumo:
Tese de Doutoramento em Engenharia de Eletrónica e de Computadores
Resumo:
Dissertação de mestrado Internacional em Sustentabilidade do Ambiente Construído
Resumo:
Relatório de estágio de mestrado em Ensino de Informática
Resumo:
Dissertação de mestrado em Media Interativos