8 resultados para sound source segregation

em QUB Research Portal - Research Directory and Institutional Repository for Queen's University Belfast


Relevância:

90.00% 90.00%

Publicador:

Resumo:

To intercept a moving object, one needs to be in the right place at the right time. In order to do this, it is necessary to pick up and use perceptual information that specifies the time to arrival of an object at an interception point. In the present study, we examined the ability to intercept a laterally moving virtual sound object by controlling the displacement of a sliding handle and tested whether and how the interaural time difference (ITD) could be the main source of perceptual information for successfully intercepting the virtual object. The results revealed that in order to accomplish the task, one might need to vary the duration of the movement, control the hand velocity and time to reach the peak velocity (speed coupling), while the adjustment of movement initiation did not facilitate performance. Furthermore, the overall performance was more successful when subjects employed a time-to-contact (tau) coupling strategy. This result shows that prospective information is available in sound for guiding goal-directed actions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Experience continuously imprints on the brain at all stages of life. The traces it leaves behind can produce perceptual learning [1], which drives adaptive behavior to previously encountered stimuli. Recently, it has been shown that even random noise, a type of sound devoid of acoustic structure, can trigger fast and robust perceptual learning after repeated exposure [2]. Here, by combining psychophysics, electroencephalography (EEG), and modeling, we show that the perceptual learning of noise is associated with evoked potentials, without any salient physical discontinuity or obvious acoustic landmark in the sound. Rather, the potentials appeared whenever a memory trace was observed behaviorally. Such memory-evoked potentials were characterized by early latencies and auditory topographies, consistent with a sensory origin. Furthermore, they were generated even on conditions of diverted attention. The EEG waveforms could be modeled as standard evoked responses to auditory events (N1-P2) [3], triggered by idiosyncratic perceptual features acquired through learning. Thus, we argue that the learning of noise is accompanied by the rapid formation of sharp neural selectivity to arbitrary and complex acoustic patterns, within sensory regions. Such a mechanism bridges the gap between the short-term and longer-term plasticity observed in the learning of noise [2, 4-6]. It could also be key to the processing of natural sounds within auditory cortices [7], suggesting that the neural code for sound source identification will be shaped by experience as well as by acoustics.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Situational awareness is achieved naturally by the human senses of sight and hearing in combination. Automatic scene understanding aims at replicating this human ability using microphones and cameras in cooperation. In this paper, audio and video signals are fused and integrated at different levels of semantic abstractions. We detect and track a speaker who is relatively unconstrained, i.e., free to move indoors within an area larger than the comparable reported work, which is usually limited to round table meetings. The system is relatively simple: consisting of just 4 microphone pairs and a single camera. Results show that the overall multimodal tracker is more reliable than single modality systems, tolerating large occlusions and cross-talk. System evaluation is performed on both single and multi-modality tracking. The performance improvement given by the audio–video integration and fusion is quantified in terms of tracking precision and accuracy as well as speaker diarisation error rate and precision–recall (recognition). Improvements vs. the closest works are evaluated: 56% sound source localisation computational cost over an audio only system, 8% speaker diarisation error rate over an audio only speaker recognition unit and 36% on the precision–recall metric over an audio–video dominant speaker recognition method.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ambisonics and higher order ambisonics (HOA) technologies aim at reproducing sound field either synthesised or previously recorded with dedicated microphones. Based on a spherical harmonic decomposition, the sound field is more precisely described when higher-order components are used. The presented study evaluated the perceptual and objective localisation accuracy of the sound field encoded with four microphones of order one to four and decoded over a ring of loudspeakers. A perceptual test showed an improvement of the localisation with higher order ambisonic microphones. Reproduced localisation indices were estimated for the four microphones and the respective synthetic systems of order one to four. The perceptual and objective analysis revealed the same conclusions. The localisation accuracy depends on the ambisonic order as well as the source incidence. Furthermore, impairments linked to the microphones were highlighted.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Finite Difference Time Domain (FDTD) method is becoming increasingly popular for room acoustics simulation. Yet, the literature on grid excitation methods is relatively sparse, and source functions are traditionally implemented in a hard or additive form
using arbitrarily-shaped functions which do not necessarily obey the physical laws of sound generation. In this paper we formulate
a source function based on a small pulsating sphere model. A physically plausible method to inject a source signal into the grid
is derived from first principles, resulting in a source with a near-flat spectrum that does not scatter incoming waves. In the final
discrete-time formulation, the source signal is the result of passing a Gaussian pulse through a digital filter simulating the dynamics of the pulsating sphere, hence facilitating a physically correct means to design source functions that generate a prescribed sound field.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper examines the impact of territoriality on young people’s everyday experiences in Northern Ireland’s segregated communities. It shows how urban encounters are reproduced through negotiating differences and the ways in which living in divided communities escalates moods of social inequality and spatial imbalances. The empirical study undertaken in the city of Derry shows how individuals and community groups position and identify themselves under the impact of social segregation. Building on Gordon Allport’s (1954) theories of contact, I explain how people in Derry have established their own sense of belonging, of who they are, based on their group memberships which eventually became an important source of pride and selfesteem. They also presented their own intertextual references as a cause of routine survival and belonging, allowing them to be more constructive about their future. Under deeply rooted segregation in Northern Ireland, the young generations are sought to provide lasting change to foster peace and integration between the two communities.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Sounds offer a rich source of information about events taking place in our physical and social environment. However, outside the domains of speech and music, little is known about whether humans can recognize and act upon the intentions of another agent’s actions detected through auditory information alone. In this study we assessed whether intention can be inferred from the sound an action makes, and in turn, whether this information can be used to prospectively guide movement. In two experiments experienced and novice basketball players had to virtually intercept an attacker by listening to audio recordings of that player’s movements. In the first experiment participants had to move a slider, while in the second one their body, to block the perceived passage of the attacker as they would in a real basketball game. Combinations of deceptive and non-deceptive movements were used to see if novice and/or experienced listeners could perceive the attacker’s intentions through sound alone. We showed that basketball players were able to more accurately predict final running direction compared to non-players, particularly in the second experiment when the interceptive action was more basketball specific. We suggest that athletes present better action anticipation by being able to pick up and use the relevant kinematic features of deceptive movement from event-related sounds alone. This result suggests that action intention can be perceived through the sound a movement makes and that the ability to determine another person’s action intention from the information conveyed through sound is honed through practice.