1000 resultados para Recognition.
Resumo:
GPR40 was formerly an orphan G protein-coupled receptor whose endogenous ligands have recently been identified as free fatty acids (FFAs). The receptor, now named FFA receptor 1, has been implicated in the pathophysiology of type 2 diabetes and is a drug target because of its role in FFA-mediated enhancement of glucose-stimulated insulin release. Guided by molecular modeling, we investigated the molecular determinants contributing to binding of linoleic acid, a C18 polyunsaturated FFA, and GW9508, a synthetic small molecule agonist. Twelve residues within the putative GPR40-binding pocket including hydrophilic/positively charged, aromatic, and hydrophobic residues were identified and were subjected to site-directed mutagenesis. Our results suggest that linoleic acid and GW9508 are anchored on their carboxylate groups by Arg183, Asn244, and Arg258. Moreover, His86, Tyr91, and His137 may contribute to aromatic and/or hydrophobic interactions with GW9508 that are not present, or relatively weak, with linoleic acid. The anchor residues, as well as the residues Tyr12, Tyr91, His137, and Leu186, appear to be important for receptor activation also. Interestingly, His137 and particularly His86 may interact with GW9508 in a manner dependent on its protonation status. The greater number of putative interactions between GPR40 and GW9508 compared with linoleic acid may explain the higher potency of GW9508.
Resumo:
In this paper, we present a new approach to visual speech recognition which improves contextual modelling by combining Inter-Frame Dependent and Hidden Markov Models. This approach captures contextual information in visual speech that may be lost using a Hidden Markov Model alone. We apply contextual modelling to a large speaker independent isolated digit recognition task, and compare our approach to two commonly adopted feature based techniques for incorporating speech dynamics. Results are presented from baseline feature based systems and the combined modelling technique. We illustrate that both of these techniques achieve similar levels of performance when used independently. However significant improvements in performance can be achieved through a combination of the two. In particular we report an improvement in excess of 17% relative Word Error Rate in comparison to our best baseline system.
Resumo:
For many applications of emotion recognition, such as virtual agents, the system must select responses while the user is speaking. This requires reliable on-line recognition of the user’s affect. However most emotion recognition systems are based on turnwise processing. We present a novel approach to on-line emotion recognition from speech using Long Short-Term Memory Recurrent Neural Networks. Emotion is recognised frame-wise in a two-dimensional valence-activation continuum. In contrast to current state-of-the-art approaches, recognition is performed on low-level signal frames, similar to those used for speech recognition. No statistical functionals are applied to low-level feature contours. Framing at a higher level is therefore unnecessary and regression outputs can be produced in real-time for every low-level input frame. We also investigate the benefits of including linguistic features on the signal frame level obtained by a keyword spotter.
Resumo:
In this paper, a novel video-based multimodal biometric verification scheme using the subspace-based low-level feature fusion of face and speech is developed for specific speaker recognition for perceptual human--computer interaction (HCI). In the proposed scheme, human face is tracked and face pose is estimated to weight the detected facelike regions in successive frames, where ill-posed faces and false-positive detections are assigned with lower credit to enhance the accuracy. In the audio modality, mel-frequency cepstral coefficients are extracted for voice-based biometric verification. In the fusion step, features from both modalities are projected into nonlinear Laplacian Eigenmap subspace for multimodal speaker recognition and combined at low level. The proposed approach is tested on the video database of ten human subjects, and the results show that the proposed scheme can attain better accuracy in comparison with the conventional multimodal fusion using latent semantic analysis as well as the single-modality verifications. The experiment on MATLAB shows the potential of the proposed scheme to attain the real-time performance for perceptual HCI applications.
Resumo:
In this paper, a novel pattern recognition scheme, global harmonic subspace analysis (GHSA), is developed for face recognition. In the proposed scheme, global harmonic features are extracted at the semantic scale to capture the 2-D semantic spatial structures of a face image. Laplacian Eigenmap is applied to discriminate faces in their global harmonic subspace. Experimental results on the Yale and PIE face databases show that the proposed GHSA scheme achieves an improvement in face recognition accuracy when compared with conventional subspace approaches, and a further investigation shows that the proposed GHSA scheme has impressive robustness to noise.