7 resultados para INTELLIGIBILITY
em CentAUR: Central Archive University of Reading - UK
Resumo:
When speech is in competition with interfering sources in rooms, monaural indicators of intelligibility fail to take account of the listener’s abilities to separate target speech from interfering sounds using the binaural system. In order to incorporate these segregation abilities and their susceptibility to reverberation, Lavandier and Culling [J. Acoust. Soc. Am. 127, 387–399 (2010)] proposed a model which combines effects of better-ear listening and binaural unmasking. A computationally efficient version of this model is evaluated here under more realistic conditions that include head shadow, multiple stationary noise sources, and real-room acoustics. Three experiments are presented in which speech reception thresholds were measured in the presence of one to three interferers using real-room listening over headphones, simulated by convolving anechoic stimuli with binaural room impulse-responses measured with dummy-head transducers in five rooms. Without fitting any parameter of the model, there was close correspondence between measured and predicted differences in threshold across all tested conditions. The model’s components of better-ear listening and binaural unmasking were validated both in isolation and in combination. The computational efficiency of this prediction method allows the generation of complex “intelligibility maps” from room designs. © 2012 Acoustical Society of America
Resumo:
Sirens used by police, fire and paramedic vehicles have been designed so that they can be heard over large distances, but unfortunately the siren noise enters the vehicle and corrupts intelligibility of voice communications from the emergency vehicle to the control room. Often the siren needs to be turned off to enable the control room to hear what is being said. This paper discusses a siren noise filter system that is capable of removing the siren noise picked up by the two-way radio microphone inside the vehicle. The removal of the siren noise improves the response time for emergency vehicles and thus save lives. To date, the system has been trialed within a fire tender in a non-emergency situation, with good results.
Resumo:
Emergency vehicles use high-amplitude sirens to warn pedestrians and other road users of their presence. Unfortunately, the siren noise enters the vehicle and corrupts the intelligibility of two-way radio voice com-munications from the emergency vehicle to a control room. Often the siren has to be turned off to enable the control room to hear what is being said which subsequently endangers people's lives. A digital signal processing (DSP) based system for the cancellation of siren noise embedded within speech is presented. The system has been tested with the least mean square (LMS), normalised least mean square (NLMS) and affine projection algorithm (APA) using recordings from three common types of sirens (two-tone, wail and yelp) from actual test vehicles. It was found that the APA with a projection order of 2 gives comparably improved cancellation over the LMS and NLMS with only a moderate increase in algorithm complexity and code size. Therefore, this siren noise cancellation system using the APA offers an improvement in cancellation achieved by previous systems. The removal of the siren noise improves the response time for the emergency vehicle and thus the system can contribute to saving lives. The system also allows voice communication to take place even when the siren is on and as such the vehicle offers less risk of danger when moving at high speeds in heavy traffic.
Resumo:
Sirens’ used by police, fire and paramedic vehicles generate noise that propagates inside the vehicle cab that subsequently corrupts intelligibility of voice communications from the emergency vehicle to the control room. It is even common for the siren to be turned off to enable the control room to hear what is being said. Both fixed filter and adaptive filter systems have previously been developed to help cancel the transmission of the siren noise over the radio. Previous cancellation systems have only concentrated on the traditional 2-tone, wail and yelp sirens. This paper discusses an improvement to a previous adaptive filter system and presents the cancellation results to three new types of sirens; being chirp pulsar and localiser. A siren noise filter system has the capability to improve the response time for an emergency vehicle and thus help save lives. To date, this system has been tested using live recordings taken from a nonemergency situation with good results.
Resumo:
Background Cluttering is a fluency disorder characterised by overly rapid or jerky speech patterns that compromise intelligibility. The neural correlates of cluttering are unknown but theoretical accounts implicate the basal ganglia and medial prefrontal cortex. Dysfunction in these brain areas would be consistent with difficulties in selection and control of speech motor programs that are characteristic of speech disfluencies in cluttering. There is a surprising lack of investigation into this disorder using modern imaging techniques. Here, we used functional MRI to investigate the neural correlates of cluttering. Method We scanned 17 adults who clutter and 17 normally fluent control speakers matched for age and sex. Brain activity was recorded using sparse-sampling functional MRI while participants viewed scenes and either (i) produced overt speech describing the scene or (ii) read out loud a sentence provided that described the scene. Speech was recorded and analysed off line. Differences in brain activity for each condition compared to a silent resting baseline and between conditions were analysed for each group separately (cluster-forming threshold Z > 3.1, extent p < 0.05, corrected) and then these differences were further compared between the two groups (voxel threshold p < 0.01, extent > 30 voxels, uncorrected). Results In both conditions, the patterns of activation in adults who clutter and control speakers were strikingly similar, particularly at the cortical level. Direct group comparisons revealed greater activity in adults who clutter compared to control speakers in the lateral premotor cortex bilaterally and, as predicted, on the medial surface (pre-supplementary motor area). Subcortically, adults who clutter showed greater activity than control speakers in the basal ganglia. Specifically, the caudate nucleus and putamen were overactive in adults who clutter for the comparison of picture description with sentence reading. In addition, adults who clutter had reduced activity relative to control speakers in the lateral anterior cerebellum bilaterally. Eleven of the 17 adults who clutter also stuttered. This comorbid diagnosis of stuttering was found to contribute to the abnormal overactivity seen in the group of adults who clutter in the right ventral premotor cortex and right anterior cingulate cortex. In the remaining areas of abnormal activity seen in adults who clutter compared to controls, the subgroup who clutter and stutter did not differ from the subgroup who clutter but do not stutter. Conclusions Our findings were in good agreement with theoretical predictions regarding the neural correlates of cluttering. We found evidence for abnormal function in the basal ganglia and their cortical output target, the medial prefrontal cortex. The findings are discussed in relation to models of cluttering that point to problems with motor control of speech.
Resumo:
What this paper adds? What is already known on the subject? Multi-sensory treatment approaches have been shown to impact outcome measures positively, such as accuracy of speech movement patterns and speech intelligibility in adults with motor speech disorders, as well as in children with apraxia of speech, autism and cerebral palsy. However, there has been no empirical study using multi-sensory treatment for children with speech sound disorders (SSDs) who demonstrate motor control issues in the jaw and orofacial structures (e.g. jaw sliding, jaw over extension, inadequate lip rounding/retraction and decreased integration of speech movements). What this paper adds? Findings from this study indicate that, for speech production disorders where both the planning and production of spatiotemporal parameters of movement sequences for speech are disrupted, multi-sensory treatment programmes that integrate auditory, visual and tactile–kinesthetic information improve auditory and visual accuracy of speech production. The training (practised in treatment) and test words (not practised in treatment) both demonstrated positive change in most participants, indicating generalization of target features to untrained words. It is inferred that treatment that focuses on integrating multi-sensory information and normalizing parameters of speech movements is an effective method for treating children with SSDs who demonstrate speech motor control issues.
Resumo:
Films that feature high-speed diegetic motion, and present those high speeds through fast mobile framing and fast cutting, are frequently charged with generating a sensory overload which empties out meaning or any sense of spatial orientation. Inherent in this discourse is a privileging of optical-spatial intelligibility that suppresses consideration of the ways cinema can represent diegetic velocity, and the spectator’s sensory experience of the same. This paper will instead highlight the centrality of the evocation of a trajectory for movement for the spectator’s experience of diegetic speed, an evocation that does not depend on optical-spatial legibility for its affective force.