961 resultados para Semantic priming
Resumo:
Objective
Pedestrian detection under video surveillance systems has always been a hot topic in computer vision research. These systems are widely used in train stations, airports, large commercial plazas, and other public places. However, pedestrian detection remains difficult because of complex backgrounds. Given its development in recent years, the visual attention mechanism has attracted increasing attention in object detection and tracking research, and previous studies have achieved substantial progress and breakthroughs. We propose a novel pedestrian detection method based on the semantic features under the visual attention mechanism.
Method
The proposed semantic feature-based visual attention model is a spatial-temporal model that consists of two parts: the static visual attention model and the motion visual attention model. The static visual attention model in the spatial domain is constructed by combining bottom-up with top-down attention guidance. Based on the characteristics of pedestrians, the bottom-up visual attention model of Itti is improved by intensifying the orientation vectors of elementary visual features to make the visual saliency map suitable for pedestrian detection. In terms of pedestrian attributes, skin color is selected as a semantic feature for pedestrian detection. The regional and Gaussian models are adopted to construct the skin color model. Skin feature-based visual attention guidance is then proposed to complete the top-down process. The bottom-up and top-down visual attentions are linearly combined using the proper weights obtained from experiments to construct the static visual attention model in the spatial domain. The spatial-temporal visual attention model is then constructed via the motion features in the temporal domain. Based on the static visual attention model in the spatial domain, the frame difference method is combined with optical flowing to detect motion vectors. Filtering is applied to process the field of motion vectors. The saliency of motion vectors can be evaluated via motion entropy to make the selected motion feature more suitable for the spatial-temporal visual attention model.
Result
Standard datasets and practical videos are selected for the experiments. The experiments are performed on a MATLAB R2012a platform. The experimental results show that our spatial-temporal visual attention model demonstrates favorable robustness under various scenes, including indoor train station surveillance videos and outdoor scenes with swaying leaves. Our proposed model outperforms the visual attention model of Itti, the graph-based visual saliency model, the phase spectrum of quaternion Fourier transform model, and the motion channel model of Liu in terms of pedestrian detection. The proposed model achieves a 93% accuracy rate on the test video.
Conclusion
This paper proposes a novel pedestrian method based on the visual attention mechanism. A spatial-temporal visual attention model that uses low-level and semantic features is proposed to calculate the saliency map. Based on this model, the pedestrian targets can be detected through focus of attention shifts. The experimental results verify the effectiveness of the proposed attention model for detecting pedestrians.
Resumo:
La littérature suggère que le sommeil paradoxal joue un rôle dans l'intégration associative de la mémoire émotionnelle. De plus, les rêves en sommeil paradoxal, en particulier leur nature bizarre et émotionnelle, semblent refléter cette fonction associative et émotionnelle du sommeil paradoxal. La conséquence des cauchemars fréquents sur ce processus est inconnue, bien que le réveil provoqué par un cauchemar semble interférer avec les fonctions du sommeil paradoxal. Le premier objectif de cette thèse était de reproduire conceptuellement des recherches antérieures démontrant que le sommeil paradoxal permet un accès hyper-associatif à la mémoire. L'utilisation d'une sieste diurne nous a permis d'évaluer les effets du sommeil paradoxal, comparativement au sommeil lent et à l’éveil, sur la performance des participants à une tâche sémantique mesurant « associational breadth » (AB). Les résultats ont montré que seuls les sujets réveillés en sommeil paradoxal ont répondu avec des associations atypiques, ce qui suggère que le sommeil paradoxal est spécifique dans sa capacité à intégrer les traces de la mémoire émotionnelle (article 1). En outre, les rapports de rêve en sommeil paradoxal étaient plus bizarres que ceux en sommeil lent, et plus intenses émotionnellement ; ces attributs semblent refléter la nature associative et émotionnelle du sommeil paradoxal (article 2). Le deuxième objectif de la thèse était de préciser si et comment le traitement de la mémoire émotionnelle en sommeil paradoxal est altéré dans le Trouble de cauchemars fréquents (NM). En utilisant le même protocole, nos résultats ont montré que les participants NM avaient des résultats plus élevés avant une sieste, ce qui correspond aux observations antérieures voulant que les personnes souffrant de cauchemars soient plus créatives. Après le sommeil paradoxal, les deux groupes, NM et CTL, ont montré des changements similaires dans leur accès associatif, avec des résultats AB-négatif plus bas et AB-positif plus grands. Une semaine plus tard, seul les participants NM a maintenu ce changement dans leur réseau sémantique (article 3). Ces résultats suggèrent qu’au fil du temps, les cauchemars peuvent interférer avec l'intégration de la mémoire émotionnelle pendant le sommeil paradoxal. En ce qui concerne l'imagerie, les participants NM avaient plus de bizarrerie et plus d’émotion positive, mais pas négative, dans leurs rêveries (article 4). Ces attributs intensifiés suggèrent à nouveau que les participants NM sont plus imaginatifs et créatifs à l’éveil. Dans l'ensemble, les résultats confirment le rôle du sommeil paradoxal dans l'intégration associative de la mémoire émotionnelle. Cependant, nos résultats concernant le Trouble de cauchemars ne sont pas entièrement en accord avec les théories suggérant que les cauchemars sont dysfonctionnels. Le groupe NM a montré plus d’associativité émotionnelle, de même que plus d'imagerie positive et bizarre à l’éveil. Nous proposons donc une nouvelle théorie de sensibilité environnementale associée au Trouble de cauchemar, suggérant qu'une sensibilité accrue à une gamme de contextes environnementaux sous-tendrait les symptômes uniques et la richesse imaginative observés chez les personnes souffrant de cauchemars fréquents. Bien que davantage de recherches doivent être faites, il est possible que ces personnes puissent bénéficier e milieux favorables, et qu’elles puissent avoir un avantage adaptatif à l'égard de l'expression créative, ce qui est particulièrement pertinent lorsque l'on considère leur pronostic et les différents types de traitements.
Resumo:
La presente ricerca tratta lo studio delle basi di conoscenza, volto a facilitare la raccolta, l'organizzazione e la distribuzione della conoscenza. La scelta dell’oggetto è dovuta all'importanza sempre maggiore acquisita da questo ambito di ricerca e all'innovazione che esso è in grado di apportare nel campo del Web semantico. Viene analizzata la base di conoscenza YAGO: se ne descrivono lo stato dell’arte, le applicazioni e i progetti per sviluppi futuri. Il lavoro è stato condotto esaminando le pubblicazioni relative al tema e rappresenta una risorsa in lingua italiana sull'argomento.
Resumo:
This study provides evidence for a Stroop-like interference effect in word recognition. Based on phonologic and semantic properties of simple words, participants who performed a same/different word-recognition task exhibited a significant response latency increase when word pairs (e.g., POLL, ROD) featured a comparison word (POLL) that was a homonym of a synonym (pole) of the target word (ROD). These results support a parallel-processing framework of lexical decision making, in which activation of the pathways to word recognition may occur at different levels automatically and in parallel. A subset of simple words that are also brand names was examined and exhibited this same interference. Implications for word recognition theory and practical implications for strategic marketing are discussed.
Resumo:
In this paper we use concepts from graph theory and cellular biology represented as ontologies, to carry out semantic mining tasks on signaling pathway networks. Specifically, the paper describes the semantic enrichment of signaling pathway networks. A cell signaling network describes the basic cellular activities and their interactions. The main contribution of this paper is in the signaling pathway research area, it proposes a new technique to analyze and understand how changes in these networks may affect the transmission and flow of information, which produce diseases such as cancer and diabetes. Our approach is based on three concepts from graph theory (modularity, clustering and centrality) frequently used on social networks analysis. Our approach consists into two phases: the first uses the graph theory concepts to determine the cellular groups in the network, which we will call them communities; the second uses ontologies for the semantic enrichment of the cellular communities. The measures used from the graph theory allow us to determine the set of cells that are close (for example, in a disease), and the main cells in each community. We analyze our approach in two cases: TGF-β and the Alzheimer Disease.
Resumo:
The number of connected devices collecting and distributing real-world information through various systems, is expected to soar in the coming years. As the number of such connected devices grows, it becomes increasingly difficult to store and share all these new sources of information. Several context representation schemes try to standardize this information, but none of them have been widely adopted. In previous work we addressed this challenge, however our solution had some drawbacks: poor semantic extraction and scalability. In this paper we discuss ways to efficiently deal with representation schemes' diversity and propose a novel d-dimension organization model. Our evaluation shows that d-dimension model improves scalability and semantic extraction.
Resumo:
In recent years the technological world has grown by incorporating billions of small sensing devices, collecting and sharing real-world information. As the number of such devices grows, it becomes increasingly difficult to manage all these new information sources. There is no uniform way to share, process and understand context information. In previous publications we discussed efficient ways to organize context information that is independent of structure and representation. However, our previous solution suffers from semantic sensitivity. In this paper we review semantic methods that can be used to minimize this issue, and propose an unsupervised semantic similarity solution that combines distributional profiles with public web services. Our solution was evaluated against Miller-Charles dataset, achieving a correlation of 0.6.
Resumo:
In this thesis, we propose to infer pixel-level labelling in video by utilising only object category information, exploiting the intrinsic structure of video data. Our motivation is the observation that image-level labels are much more easily to be acquired than pixel-level labels, and it is natural to find a link between the image level recognition and pixel level classification in video data, which would transfer learned recognition models from one domain to the other one. To this end, this thesis proposes two domain adaptation approaches to adapt the deep convolutional neural network (CNN) image recognition model trained from labelled image data to the target domain exploiting both semantic evidence learned from CNN, and the intrinsic structures of unlabelled video data. Our proposed approaches explicitly model and compensate for the domain adaptation from the source domain to the target domain which in turn underpins a robust semantic object segmentation method for natural videos. We demonstrate the superior performance of our methods by presenting extensive evaluations on challenging datasets comparing with the state-of-the-art methods.
Resumo:
The current study investigated the cognitive workload of sentence and clause wrap-up in younger and older readers. A large number of studies have demonstrated the presence of wrap-up effects, peaks in processing time at clause and sentence boundaries that some argue reflect attention to organizational and integrative semantic processes. However, the exact nature of these wrap-up effects is still not entirely clear, with some arguing that wrap-up is not related to processing difficulty, but rather is triggered by a low-level oculomotor response or the implicit monitoring of intonational contour. The notion that wrap-up effects are resource-demanding was directly tested by examining the degree to which sentence and clause wrap-up affects the parafoveal preview benefit. Older and younger adults read passages in which a target word N occurred in a sentence-internal, clause-final, or sentence-final position. A gaze-contingent boundary change paradigm was used in which, on some trials, a non-word preview of word N+1 was replaced by a target word once the eyes crossed an invisible boundary located between words N and N+1. All measures of reading time on word N were longer at clause and sentence boundaries than in the sentence-internal position. In the earliest measures of reading time, sentence and clause wrap-up showed evidence of reducing the magnitude of the preview benefit similarly for younger and older adults. However, this effect was moderated by age in gaze duration, such that older adults showed a complete reduction in the preview benefit in the sentence-final condition. Additionally, sentence and clause wrap-up were negatively associated with the preview benefit. Collectively, the findings from the current study suggest that wrap-up is cognitively demanding and may be less efficient with age, thus, resulting in a reduction of the parafoveal preview during normal reading.
Resumo:
Humans have a high ability to extract visual data information acquired by sight. Trought a learning process, which starts at birth and continues throughout life, image interpretation becomes almost instinctively. At a glance, one can easily describe a scene with reasonable precision, naming its main components. Usually, this is done by extracting low-level features such as edges, shapes and textures, and associanting them to high level meanings. In this way, a semantic description of the scene is done. An example of this, is the human capacity to recognize and describe other people physical and behavioral characteristics, or biometrics. Soft-biometrics also represents inherent characteristics of human body and behaviour, but do not allow unique person identification. Computer vision area aims to develop methods capable of performing visual interpretation with performance similar to humans. This thesis aims to propose computer vison methods which allows high level information extraction from images in the form of soft biometrics. This problem is approached in two ways, unsupervised and supervised learning methods. The first seeks to group images via an automatic feature extraction learning , using both convolution techniques, evolutionary computing and clustering. In this approach employed images contains faces and people. Second approach employs convolutional neural networks, which have the ability to operate on raw images, learning both feature extraction and classification processes. Here, images are classified according to gender and clothes, divided into upper and lower parts of human body. First approach, when tested with different image datasets obtained an accuracy of approximately 80% for faces and non-faces and 70% for people and non-person. The second tested using images and videos, obtained an accuracy of about 70% for gender, 80% to the upper clothes and 90% to lower clothes. The results of these case studies, show that proposed methods are promising, allowing the realization of automatic high level information image annotation. This opens possibilities for development of applications in diverse areas such as content-based image and video search and automatica video survaillance, reducing human effort in the task of manual annotation and monitoring.
Resumo:
[EN] The experiment discussed in this paper is a direct replication of Finkbeiner (2005) and an indirect replication of Jiang and Forster (2001) and Witzel and Forster (2012). The paper explores the use of episodic memory in L2 vocabulary processing. By administering an L1 episodic recognition task with L2 masked translation primes, reduced reaction times would suggest L2 vocabulary storage in episodic memory. The methodology follows Finkbeiner (2005), who argued that a blank screen introduced after the prime in Jiang Forster (2001) led to a ghosting effect, compromising the imperceptibility of the prime. The results here mostly corroborate Finkbeiner (2005) with no significant priming effects. While Finkbeiner discusses his findings in terms of the dissociability of episodic and semantic memory, and discounts Jiang and Forster’s (2001) results to participants’ strategic responding, I add a layer of analysis based on declarative and procedural constituents. From this perspective, Jiang and Forster (2001) and Witzel and Forster’s (2012) results can be seen as possible episodic memory activation, and Finkbeiner’s (2005) and my lack of priming effects might be due to the sole activation of procedural neural networks. Priming effects are found in concrete and abstract words but require verification through further experimentation.
Resumo:
[EN] The experiment discussed in this paper is a direct replication of Finkbeiner (2005) and an indirect replication of Jiang and Forster (2001) and Witzel and Forster(2012). The paper explores the use of episodic memory in L2 vocabulary processing. By administering an L1 episodic recognition task with L2 masked translation primes, reduced reaction times would suggest L2 vocabulary storage in episodic memory. The methodology follows Finkbeiner (2005) who argued that a blank screen introduced after the prime in Jiang Forster (2001) led to a ghosting effect, compromising the imperceptibility of the prime. The results here mostly corroborate Finkbeiner (2005) with no significant priming effects. While Finkbeiner discusses his findings in terms of the dissociability of episodic and semantic memory, and discounts Jiang and Forster’s (2001) results to participants’ strategic responding, I add a layer of analysis based on declarative and procedural constituents. From this perspective, Jiang and Forster (2001) and Witzel and Forster’s (2012) results can be seen as possible episodic memory activation, and Finkbeiner’s (2005) and my lack of priming effects might be due to the sole activation of procedural neural networks. Priming effects are found in concrete and abstract words but require verification through further experimentation.
Resumo:
220 p.
Resumo:
Nearest neighbour collaborative filtering (NNCF) algorithms are commonly used in multimedia recommender systems to suggest media items based on the ratings of users with similar preferences. However, the prediction accuracy of NNCF algorithms is affected by the reduced number of items – the subset of items co-rated by both users – typically used to determine the similarity between pairs of users. In this paper, we propose a different approach, which substantially enhances the accuracy of the neighbour selection process – a user-based CF (UbCF) with semantic neighbour discovery (SND). Our neighbour discovery methodology, which assesses pairs of users by taking into account all the items rated at least by one of the users instead of just the set of co-rated items, semantically enriches this enlarged set of items using linked data and, finally, applies the Collinearity and Proximity Similarity metric (CPS), which combines the cosine similarity with Chebyschev distance dissimilarity metric. We tested the proposed SND against the Pearson Correlation neighbour discovery algorithm off-line, using the HetRec data set, and the results show a clear improvement in terms of accuracy and execution time for the predicted recommendations.