953 resultados para object orientation processing
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
The integration of CMOS cameras with embedded processors and wireless communication devices has enabled the development of distributed wireless vision systems. Wireless Vision Sensor Networks (WVSNs), which consist of wirelessly connected embedded systems with vision and sensing capabilities, provide wide variety of application areas that have not been possible to realize with the wall-powered vision systems with wired links or scalar-data based wireless sensor networks. In this paper, the design of a middleware for a wireless vision sensor node is presented for the realization of WVSNs. The implemented wireless vision sensor node is tested through a simple vision application to study and analyze its capabilities, and determine the challenges in distributed vision applications through a wireless network of low-power embedded devices. The results of this paper highlight the practical concerns for the development of efficient image processing and communication solutions for WVSNs and emphasize the need for cross-layer solutions that unify these two so-far-independent research areas.
Resumo:
The web services (WS) technology provides a comprehensive solution for representing, discovering, and invoking services in a wide variety of environments, including Service Oriented Architectures (SOA) and grid computing systems. At the core of WS technology lie a number of XML-based standards, such as the Simple Object Access Protocol (SOAP), that have successfully ensured WS extensibility, transparency, and interoperability. Nonetheless, there is an increasing demand to enhance WS performance, which is severely impaired by XML's verbosity. SOAP communications produce considerable network traffic, making them unfit for distributed, loosely coupled, and heterogeneous computing environments such as the open Internet. Also, they introduce higher latency and processing delays than other technologies, like Java RMI and CORBA. WS research has recently focused on SOAP performance enhancement. Many approaches build on the observation that SOAP message exchange usually involves highly similar messages (those created by the same implementation usually have the same structure, and those sent from a server to multiple clients tend to show similarities in structure and content). Similarity evaluation and differential encoding have thus emerged as SOAP performance enhancement techniques. The main idea is to identify the common parts of SOAP messages, to be processed only once, avoiding a large amount of overhead. Other approaches investigate nontraditional processor architectures, including micro-and macrolevel parallel processing solutions, so as to further increase the processing rates of SOAP/XML software toolkits. This survey paper provides a concise, yet comprehensive review of the research efforts aimed at SOAP performance enhancement. A unified view of the problem is provided, covering almost every phase of SOAP processing, ranging over message parsing, serialization, deserialization, compression, multicasting, security evaluation, and data/instruction-level processing.
Resumo:
Metalinguistic skill is the ability to reflect upon language as an object of thought. Amongst metalinguistic skills, two seem to be associated with reading and spelling: morphological awareness and phonological awareness. Phonological awareness is the ability of reflecting upon the phonemes that compose words, and morphological awareness is the ability of reflecting upon the morphemes that compose the words. The latter seems to be particularly important for reading comprehension and contextual reading, as beyond phonological information, syntactic and semantic information are required. This study is set to investigate - with a longitudinal design - the relation between those abilities and contextual reading measured by the Cloze test. The first part of the study explores the relationship between morphological awareness tasks and Cloze scores through simple correlations and, in the second part, the specificity of such relationship was inquired using multiple regressions. The results give some support to the hypothesis that morphological awareness offers an independent contribution regarding phonological awareness to contextual reading in Brazilian Portuguese.
Resumo:
[EN]The human face provides useful information during interaction; therefore, any system integrating Vision- BasedHuman Computer Interaction requires fast and reliable face and facial feature detection. Different approaches have focused on this ability but only open source implementations have been extensively used by researchers. A good example is the Viola–Jones object detection framework that particularly in the context of facial processing has been frequently used.
Resumo:
The research activity carried out during the PhD course was focused on the development of mathematical models of some cognitive processes and their validation by means of data present in literature, with a double aim: i) to achieve a better interpretation and explanation of the great amount of data obtained on these processes from different methodologies (electrophysiological recordings on animals, neuropsychological, psychophysical and neuroimaging studies in humans), ii) to exploit model predictions and results to guide future research and experiments. In particular, the research activity has been focused on two different projects: 1) the first one concerns the development of neural oscillators networks, in order to investigate the mechanisms of synchronization of the neural oscillatory activity during cognitive processes, such as object recognition, memory, language, attention; 2) the second one concerns the mathematical modelling of multisensory integration processes (e.g. visual-acoustic), which occur in several cortical and subcortical regions (in particular in a subcortical structure named Superior Colliculus (SC)), and which are fundamental for orienting motor and attentive responses to external world stimuli. This activity has been realized in collaboration with the Center for Studies and Researches in Cognitive Neuroscience of the University of Bologna (in Cesena) and the Department of Neurobiology and Anatomy of the Wake Forest University School of Medicine (NC, USA). PART 1. Objects representation in a number of cognitive functions, like perception and recognition, foresees distribute processes in different cortical areas. One of the main neurophysiological question concerns how the correlation between these disparate areas is realized, in order to succeed in grouping together the characteristics of the same object (binding problem) and in maintaining segregated the properties belonging to different objects simultaneously present (segmentation problem). Different theories have been proposed to address these questions (Barlow, 1972). One of the most influential theory is the so called “assembly coding”, postulated by Singer (2003), according to which 1) an object is well described by a few fundamental properties, processing in different and distributed cortical areas; 2) the recognition of the object would be realized by means of the simultaneously activation of the cortical areas representing its different features; 3) groups of properties belonging to different objects would be kept separated in the time domain. In Chapter 1.1 and in Chapter 1.2 we present two neural network models for object recognition, based on the “assembly coding” hypothesis. These models are networks of Wilson-Cowan oscillators which exploit: i) two high-level “Gestalt Rules” (the similarity and previous knowledge rules), to realize the functional link between elements of different cortical areas representing properties of the same object (binding problem); 2) the synchronization of the neural oscillatory activity in the γ-band (30-100Hz), to segregate in time the representations of different objects simultaneously present (segmentation problem). These models are able to recognize and reconstruct multiple simultaneous external objects, even in difficult case (some wrong or lacking features, shared features, superimposed noise). In Chapter 1.3 the previous models are extended to realize a semantic memory, in which sensory-motor representations of objects are linked with words. To this aim, the network, previously developed, devoted to the representation of objects as a collection of sensory-motor features, is reciprocally linked with a second network devoted to the representation of words (lexical network) Synapses linking the two networks are trained via a time-dependent Hebbian rule, during a training period in which individual objects are presented together with the corresponding words. Simulation results demonstrate that, during the retrieval phase, the network can deal with the simultaneous presence of objects (from sensory-motor inputs) and words (from linguistic inputs), can correctly associate objects with words and segment objects even in the presence of incomplete information. Moreover, the network can realize some semantic links among words representing objects with some shared features. These results support the idea that semantic memory can be described as an integrated process, whose content is retrieved by the co-activation of different multimodal regions. In perspective, extended versions of this model may be used to test conceptual theories, and to provide a quantitative assessment of existing data (for instance concerning patients with neural deficits). PART 2. The ability of the brain to integrate information from different sensory channels is fundamental to perception of the external world (Stein et al, 1993). It is well documented that a number of extraprimary areas have neurons capable of such a task; one of the best known of these is the superior colliculus (SC). This midbrain structure receives auditory, visual and somatosensory inputs from different subcortical and cortical areas, and is involved in the control of orientation to external events (Wallace et al, 1993). SC neurons respond to each of these sensory inputs separately, but is also capable of integrating them (Stein et al, 1993) so that the response to the combined multisensory stimuli is greater than that to the individual component stimuli (enhancement). This enhancement is proportionately greater if the modality-specific paired stimuli are weaker (the principle of inverse effectiveness). Several studies have shown that the capability of SC neurons to engage in multisensory integration requires inputs from cortex; primarily the anterior ectosylvian sulcus (AES), but also the rostral lateral suprasylvian sulcus (rLS). If these cortical inputs are deactivated the response of SC neurons to cross-modal stimulation is no different from that evoked by the most effective of its individual component stimuli (Jiang et al 2001). This phenomenon can be better understood through mathematical models. The use of mathematical models and neural networks can place the mass of data that has been accumulated about this phenomenon and its underlying circuitry into a coherent theoretical structure. In Chapter 2.1 a simple neural network model of this structure is presented; this model is able to reproduce a large number of SC behaviours like multisensory enhancement, multisensory and unisensory depression, inverse effectiveness. In Chapter 2.2 this model was improved by incorporating more neurophysiological knowledge about the neural circuitry underlying SC multisensory integration, in order to suggest possible physiological mechanisms through which it is effected. This endeavour was realized in collaboration with Professor B.E. Stein and Doctor B. Rowland during the 6 months-period spent at the Department of Neurobiology and Anatomy of the Wake Forest University School of Medicine (NC, USA), within the Marco Polo Project. The model includes four distinct unisensory areas that are devoted to a topological representation of external stimuli. Two of them represent subregions of the AES (i.e., FAES, an auditory area, and AEV, a visual area) and send descending inputs to the ipsilateral SC; the other two represent subcortical areas (one auditory and one visual) projecting ascending inputs to the same SC. Different competitive mechanisms, realized by means of population of interneurons, are used in the model to reproduce the different behaviour of SC neurons in conditions of cortical activation and deactivation. The model, with a single set of parameters, is able to mimic the behaviour of SC multisensory neurons in response to very different stimulus conditions (multisensory enhancement, inverse effectiveness, within- and cross-modal suppression of spatially disparate stimuli), with cortex functional and cortex deactivated, and with a particular type of membrane receptors (NMDA receptors) active or inhibited. All these results agree with the data reported in Jiang et al. (2001) and in Binns and Salt (1996). The model suggests that non-linearities in neural responses and synaptic (excitatory and inhibitory) connections can explain the fundamental aspects of multisensory integration, and provides a biologically plausible hypothesis about the underlying circuitry.
Resumo:
In recent years, the use of Reverse Engineering systems has got a considerable interest for a wide number of applications. Therefore, many research activities are focused on accuracy and precision of the acquired data and post processing phase improvements. In this context, this PhD Thesis deals with the definition of two novel methods for data post processing and data fusion between physical and geometrical information. In particular a technique has been defined for error definition in 3D points’ coordinates acquired by an optical triangulation laser scanner, with the aim to identify adequate correction arrays to apply under different acquisition parameters and operative conditions. Systematic error in data acquired is thus compensated, in order to increase accuracy value. Moreover, the definition of a 3D thermogram is examined. Object geometrical information and its thermal properties, coming from a thermographic inspection, are combined in order to have a temperature value for each recognizable point. Data acquired by an optical triangulation laser scanner are also used to normalize temperature values and make thermal data independent from thermal-camera point of view.
Resumo:
Lesions to the primary geniculo-striate visual pathway cause blindness in the contralesional visual field. Nevertheless, previous studies have suggested that patients with visual field defects may still be able to implicitly process the affective valence of unseen emotional stimuli (affective blindsight) through alternative visual pathways bypassing the striate cortex. These alternative pathways may also allow exploitation of multisensory (audio-visual) integration mechanisms, such that auditory stimulation can enhance visual detection of stimuli which would otherwise be undetected when presented alone (crossmodal blindsight). The present dissertation investigated implicit emotional processing and multisensory integration when conscious visual processing is prevented by real or virtual lesions to the geniculo-striate pathway, in order to further clarify both the nature of these residual processes and the functional aspects of the underlying neural pathways. The present experimental evidence demonstrates that alternative subcortical visual pathways allow implicit processing of the emotional content of facial expressions in the absence of cortical processing. However, this residual ability is limited to fearful expressions. This finding suggests the existence of a subcortical system specialised in detecting danger signals based on coarse visual cues, therefore allowing the early recruitment of flight-or-fight behavioural responses even before conscious and detailed recognition of potential threats can take place. Moreover, the present dissertation extends the knowledge about crossmodal blindsight phenomena by showing that, unlike with visual detection, sound cannot crossmodally enhance visual orientation discrimination in the absence of functional striate cortex. This finding demonstrates, on the one hand, that the striate cortex plays a causative role in crossmodally enhancing visual orientation sensitivity and, on the other hand, that subcortical visual pathways bypassing the striate cortex, despite affording audio-visual integration processes leading to the improvement of simple visual abilities such as detection, cannot mediate multisensory enhancement of more complex visual functions, such as orientation discrimination.
Resumo:
Polymer-nanoparticle hybrids show synergistic effects, demonstrating both, the unique properties of nanosized structures and the good processability and functionalities of polymeric materials. This work shows the synthesis and application of block copolymers containing a soluble, functional block and a short anchor block, which efficiently binds to the surface of nanocrystals. We functionalized anisotropic, semiconducting nanoparticles, which can be dissolved in organic and polymeric matrices upon modification. The modified nanorods have the ability to form liquid crystalline phases, which behave similar to low molecular liquid crystals with a reversible clearing behaviour. These liquid crystalline phases could also be obtained in hole conducting matrices. For a macroscopic orientation of the nanorods, electric fields were applied and a switching (in analogy to known liquid crystals) to a homeotropic orientation was observed.rnBy introduction of dye molecules in the anchor block of a hole conducting block copolymer, all essential components of a solar cell can be combined in a single particle. Light absorption of the dye induces the injection of electrons into the particles, followed by a charging, that was monitored by a special AFM technique.rnLight emitting nanocrystals were functionalized analogously with a hole transporting polymer. The stability of the particles could be enhanced by the sterically stabilizing polymer corona and the particles showed improved properties in terms of processing. We applied these hybrid materials in light emitting devices, which showed better characteristics due to an improved hole injection and well dispersed emitting particles in the active device layer.rnThe work shows the broad spectrum of properties and applications based on the synergistic effects in hybrid and composite materials.
Resumo:
Zielgerichtete Orientierung ermöglicht es Lebewesen, überlebenswichtige Aufgaben, wie die Suche nach Ressourcen, Fortpflanzungspartnern und sicheren Plätzen zu bewältigen. Dafür ist es essentiell, die Umgebung sensorisch wahrzunehmen, frühere Erfahrungen zu speichern und wiederabzurufen und diese Informationen zu integrieren und in motorische Aktionen umzusetzen.rnWelche Neuronengruppen vermitteln zielgerichtete Orientierung im Gehirn einer Fliege? Welche sensorischen Informationen sind in einem gegebenen Kontext relevant und wie werden diese Informationen sowie gespeichertes Vorwissen in motorische Aktionen übersetzt? Wo findet im Gehirn der Übergang von der sensorischen Verarbeitung zur motorischen Kontrolle statt? rnDer Zentralkomplex, ein Verbund von vier Neuropilen des Zentralhirns von Drosophila melanogaster, fungiert als Übergang zwischen in den optischen Loben vorverarbeiteten visuellen Informationen und prämotorischem Ausgang. Die Neuropile sind die Protocerebralbrücke, der Fächerförmige Körper, der Ellipsoidkörper und die Noduli. rnIn der vorliegenden Arbeit konnte gezeigt werden, dass Fruchtfliegen ein räumliches Arbeitsgedächtnis besitzen. Dieses Gedächtnis kann aktuelle visuelle Information ersetzen, wenn die Sicht auf das Zielobjekt verloren geht. Dies erfordert die sensorische Wahrnehmung von Zielobjekten, die Speicherung der Position, die kontinuierliche Integration von Eigen-und Objektposition, sowie die Umsetzung der sensorischen Information in zielgerichtete Bewegung. Durch konditionale Expression von Tetanus Toxin mittels des GAL4/UAS/GAL80ts Systems konnte gezeigt werden, dass die Ringneurone, welche in den Ellipsoidkörper projizieren, für das Orientierungsgedächtnis notwendig sind. Außerdem konnte gezeigt werden, dass Fliegen, denen die ribosomale Serinkinase S6KII fehlt, die Richtung verlieren, sobald keine Objekte mehr sichtbar sind und, dass die partielle Rettung dieser Kinase ausschließlich in den Ringneuronenklassen R3 und R4d hinreichend ist, um das Gedächtnis wieder herzustellen. Bei dieser Gedächtnisleistung scheint es sich um eine idiothetische Form der Orientierung zu handeln. rn Während das räumliche Arbeitsgedächtnis nach Verschwinden von Objekten relevant ist, wurde in der vorliegende Arbeit auch die Vermittlung zielgerichteter Bewegung auf sichtbare Objekte untersucht. Dabei wurde die zentrale Frage bearbeitet, welche Neuronengruppen visuelle Orientierung vermitteln. Anhand von Gehirnstrukturmutanten konnte gezeigt werden, dass eine intakte Protocerebralbrücke notwendig ist, um Laufgeschwindigkeit, Laufaktivität und Zielgenauigkeit bei der Ansteuerung visueller Stimuli korrekt zu vermitteln. Dabei scheint das Horizontale Fasersystem, welches von der Protocerebralbrücke über den Fächerförmigen Körper auf den Zentralkomplex assoziierte Neuropile, die Ventralkörper, projiziert, notwendig für die lokomotorische Kontrolle und die zielgenaue Bewegung zu sein. Letzeres konnte zum einen durch Blockade der synaptischen Transmission anhand konditionaler Tetanus Toxin Expression mittels des GAL4/UAS/GAL80ts Systems im Horizontalen Fasersystem gezeigt werden;. zum anderen auch durch partielle Rettung der in den Strukturmutanten betroffenen Gene. rn Den aktuellen Ergebnissen und früheren Studien folgend, ergibt sich dabei ein Modell, wie zielgerichtete Bewegung auf visuelle Stimuli neuronal vermittelt werden könnte. Nach diesem Modell bildet die Protocerebralbrücke die Azimuthpositionen von Objekten ab und das Horizontale Fasersystem vermittelt die entsprechende lokomotorische Wo-Information für zielgerichtete Bewegungen. Die Eigenposition in Relation zum Zielobjekt wird über die Ringneurone und den Ellipsoidkörper vermittelt. Wenn das Objekt aus der Sicht verschwindet, kann die Relativposition ideothetisch ermittelt werden und integriert werden mit Vorinformation über das Zielobjekt, die im Fächerförmigen Körper abgelegt ist (Was-Information). Die resultierenden Informationen könnten dann über das Horizontale Fasersystem in den Ventralkörpern auf absteigende Neurone gelangen und in den Thorax zu den motorischen Zentren weitergeleitet werden.rn
Resumo:
In recent years, Deep Learning techniques have shown to perform well on a large variety of problems both in Computer Vision and Natural Language Processing, reaching and often surpassing the state of the art on many tasks. The rise of deep learning is also revolutionizing the entire field of Machine Learning and Pattern Recognition pushing forward the concepts of automatic feature extraction and unsupervised learning in general. However, despite the strong success both in science and business, deep learning has its own limitations. It is often questioned if such techniques are only some kind of brute-force statistical approaches and if they can only work in the context of High Performance Computing with tons of data. Another important question is whether they are really biologically inspired, as claimed in certain cases, and if they can scale well in terms of "intelligence". The dissertation is focused on trying to answer these key questions in the context of Computer Vision and, in particular, Object Recognition, a task that has been heavily revolutionized by recent advances in the field. Practically speaking, these answers are based on an exhaustive comparison between two, very different, deep learning techniques on the aforementioned task: Convolutional Neural Network (CNN) and Hierarchical Temporal memory (HTM). They stand for two different approaches and points of view within the big hat of deep learning and are the best choices to understand and point out strengths and weaknesses of each of them. CNN is considered one of the most classic and powerful supervised methods used today in machine learning and pattern recognition, especially in object recognition. CNNs are well received and accepted by the scientific community and are already deployed in large corporation like Google and Facebook for solving face recognition and image auto-tagging problems. HTM, on the other hand, is known as a new emerging paradigm and a new meanly-unsupervised method, that is more biologically inspired. It tries to gain more insights from the computational neuroscience community in order to incorporate concepts like time, context and attention during the learning process which are typical of the human brain. In the end, the thesis is supposed to prove that in certain cases, with a lower quantity of data, HTM can outperform CNN.
Resumo:
In the field of computer assisted orthopedic surgery (CAOS) the anterior pelvic plane (APP) is a common concept to determine the pelvic orientation by digitizing distinct pelvic landmarks. As percutaneous palpation is - especially for obese patients - known to be error-prone, B-mode ultrasound (US) imaging could provide an alternative means. Several concepts of using ultrasound imaging to determine the APP landmarks have been introduced. In this paper we present a novel technique, which uses local patch statistical shape models (SSMs) and a hierarchical speed of sound compensation strategy for an accurate determination of the APP. These patches are independently matched and instantiated with respect to associated point clouds derived from the acquired ultrasound images. Potential inaccuracies due to the assumption of a constant speed of sound are compensated by an extended reconstruction scheme. We validated our method with in-vitro studies using a plastic bone covered with a soft-tissue simulation phantom and with a preliminary cadaver trial.
Resumo:
Capuchin monkeys, Cebus sp., utilize a wide array of gestural displays in the wild, including facial displays such as lip-smacking and bare-teeth displays. In captivity, they have been shown to respond to the head orientation of humans, show sensitivity to human attentional states, as well as follow human gazes behind barriers. In this study, I investigated whether tufted capuchin monkeys (Cebus apella) would attend to and utilize the gestural cues of a conspecific to obtain a hidden reward. Two capuchins faced each other in separate compartments of an apparatus with an open field in between. The open field contained two cups with holes on one side such that only one monkey, a so-called cuing monkey, could see the reward inside one of the cups. I then moved the cups toward the other signal-receiving monkey and assessed whether it would utilize untrained cues provided by the cuing monkey to select the cup containing the reward. Two of four female capuchin monkeys learned to select the cup containing the reward significantly more often than chance. Neither of these two monkeys performed over chance spontaneously, however, and the other two monkeys never performed above chance despite many blocks of trials. Successful choices by two monkeys to obtain hidden rewards provided experimental evidence that capuchin monkeys attend to and utilize the gestural cues of conspecifics.
Resumo:
The processing of orientations is at the core of our visual experience. Orientation selectivity in human visual cortex has been inferred from psychophysical experiments and more recently demonstrated with functional magnetic resonance imaging (fMRI). One method to identify orientation-selective responses is fMRI adaptation, in which two stimuli—either with the same or with different orientations—are presented successively. A region containing orientation-selective neurons should demonstrate an adapted response to the “same orientation” condition in contrast to the “different orientation” condition. So far, human primary visual cortex (V1) showed orientation-selective fMRI adaptation only in experimental designs using prolonged pre-adaptation periods (∼40 s) in combination with top-up stimuli that are thought to maintain the adapted level. This finding has led to the notion that orientation-selective short-term adaptation in V1 (but not V2 or V3) cannot be demonstrated using fMRI. The present study aimed at re-evaluating this question by testing three differently timed adaptation designs. With the use of a more sensitive analysis technique, we show robust orientation-selective fMRI adaptation in V1 evoked by a short-term adaptation design.