57 resultados para loudness
Resumo:
The deliberate addition of Gaussian noise to cochlear implant signals has previously been proposed to enhance the time coding of signals by the cochlear nerve. Potentially, the addition of an inaudible level of noise could also have secondary benefits: it could lower the threshold to the information-bearing signal, and by desynchronization of nerve discharges, it could increase the level at which the information-bearing signal becomes uncomfortable. Both these effects would lead to an increased dynamic range, which might be expected to enhance speech comprehension and make the choice of cochlear implant compression parameters less critical (as with a wider dynamic range, small changes in the parameters would have less effect on loudness). The hypothesized secondary effects were investigated with eight users of the Clarion cochlear implant; the stimulation was analogue and monopolar. For presentations in noise, noise at 95% of the threshold level was applied simultaneously and independently to all the electrodes. The noise was found in two-alternative forced-choice (2AFC) experiments to decrease the threshold to sinusoidal stimuli (100 Hz, 1 kHz, 5 kHz) by about 2.0 dB and increase the dynamic range by 0.7 dB. Furthermore, in 2AFC loudness balance experiments, noise was found to decrease the loudness of moderate to intense stimuli. This suggests that loudness is partially coded by the degree of phase-locking of cochlear nerve fibers. The overall gain in dynamic range was modest, and more complex noise strategies, for example, using inhibition between the noise sources, may be required to get a clinically useful benefit. © 2006 Association for Research in Otolaryngology.
Resumo:
26 tinnitus patients received either electromyogram (EMG) biofeedback with counterdemand instructions, EMG biofeedback with neutral demand instructions, or no treatment. Assessment was conducted on self-report measures of the distress associated with tinnitus, the loudness, annoyance and awareness of tinnitus, sleep-onset difficulties, depression, and anxiety. Audiological assessment of tinnitus was also conducted and EMG levels were measured (the latter only in the 2 treatment groups). No significant treatment effects were found on any of the measures. There was a significant decrease in the ratings of tinnitus awareness over the assessment occasions, but the degree of change was equivalent for treated and untreated groups. Results do not support the assertion that EMG biofeedback is an effective treatment for tinnitus.
Resumo:
When communicating emotion in music, composers and performers encode their expressive intentions through the control of basic musical features such as: pitch, loudness, timbre, mode, and articulation. The extent to which emotion can be controlled through the systematic manipulation of these features has not been fully examined. In this paper we present CMERS, a Computational Music Emotion Rule System for the control of perceived musical emotion that modifies features at the levels of score and performance in real-time. CMERS performance was evaluated in two rounds of perceptual testing. In experiment I, 20 participants continuously rated the perceived emotion of 15 music samples generated by CMERS. Three music works, each with five emotional variations were used (normal, happy, sad, angry, and tender). The intended emotion by CMERS was correctly identified 78% of the time, with significant shifts in valence and arousal also recorded, regardless of the works’ original emotion.
Resumo:
The early detection of hearing deficits is important to a child's development. However, examining small children with behavioural methods is often difficult. Research with ERPs (event-related potentials), recorded with EEG (electroencephalography), does not require attention or action from the child. Especially in children's ERP research, it is essential that the duration of a recording session is not too long. A new, faster optimum paradigm has been developed to record MMN (mismatch negativity), where ERPs to several sound features can be recorded in one recording session. This substantially shortens the time required for the experiment. So far, the new paradigm has been used in adult and school-aged children research. This study examines if MMN, LDN (late discriminative negativity) and P3a components can be recorded in two-year-olds with the new paradigm. The standard stimulus (p=0.50) was an 80 dB harmonic tone consisting of three harmonic frequencies (500 Hz, 1000 Hz and 1500 Hz) with a duration of 200 ms. The loudness deviants (p=0.067) were at a level of +6 dB or -6 dB compared to the standards. The frequency deviants (p=0.112) had a fundamental frequency of 550 or 454.4 Hz (small deviation), 625 or 400 Hz (medium deviation) or 750 or 333.3 Hz (large deviation). The duration deviants (p=0.112) had a duration of 175 ms (small deviation), 150 ms (medium deviation) or 100 ms (large deviation). The direction deviants (p=0.067) were presented from the left or right loudspeaker only. The gap deviant (p=0.067) included a 5-ms silent gap in the middle of the sound. Altogether 17 children participated in the experiment, of whom the data of 12 children was used in the analysis. ERP components were observed for all deviant types. The MMN was significant for duration and gap deviants. The LDN was significant for the large duration deviant and all other deviant types. No significant P3a was observed. These results indicate that the optimum paradigm can be used with two-year-olds. With this paradigm, data on several sound features can be recorded in a shorter time than with the previous paradigms used in ERP research.
Resumo:
Real-Time services are traditionally supported on circuit switched network. However, there is a need to port these services on packet switched network. Architecture for audio conferencing application over the Internet in the light of ITU-T H.323 recommendations is considered. In a conference, considering packets only from a set of selected clients can reduce speech quality degradation because mixing packets from all clients can lead to lack of speech clarity. A distributed algorithm and architecture for selecting clients for mixing is suggested here based on a new quantifier of the voice activity called “Loudness Number” (LN). The proposed system distributes the computation load and reduces the load on client terminals. The highlights of this architecture are scalability, bandwidth saving and speech quality enhancement. Client selection for playing out tries to mimic a physical conference where the most vocal participants attract more attention. The contributions of the paper are expected to aid H.323 recommendations implementations for Multipoint Processors (MP). A working prototype based on the proposed architecture is already functional.
Resumo:
The change in acoustic characteristics in personal computers to console gaming and home entertainment systems with the change in the Graphics Processing Unit (GPU), is presented. The tests are carried out using identical configurations of the software and system hardware. The prime components of the hardware used in the project are central processing unit, motherboard, hard disc drive, memory, power supply, optical drive, and additional cooling system. The results from the measurements taken for each GPU tested are analyzed and compared. The test results are obtained using a photo tachometer and reflective tape adhered to one particular fan blade. The test shows that loudness is a psychoacoustic metric developed by Zwicker and Fastal that aims to quantify how loud a sound is perceived as compared to a standard sound. The acoustic experiment reveals that the inherent noise generation mechanism increases with the increase of the complexity of the cooling solution.
Resumo:
This paper discusses the development of a computationally efficient numerical method for predicting the acoustics of rattle events upfront in the design cycle. The method combines Finite Elements, Boundary Elements and SEA and enables the loudness of a large number of rattle events to be efficiently predicted across a broad frequency range. A low frequency random vibro-acoustic model is used in conjunction with various closed form analytical expressions in order to quickly predict impact probabilities and locations. An existing method has been extended to estimate the statistics of the contact forces across a broad frequency range. Finally, broadband acoustic radiation is predicted using standard low, mid and high frequency vibro-acoustic methods and used to estimate impact loudness. The approach is discussed and a number of validation examples are presented.
Resumo:
As a species of internal representation, how is mental imagery organized in the brain? There are two issues related to this question: the time course and the nature of mental imagery. On the nature of mental imagery, today's imagery debate is influenced by two opposing theories: (1) Pylyshyn’s propositional theory and (2) Kosslyn’s depictive representation theory. Behavioural studies indicated that imagery encodes properties of the physical world, such as the spacial and size information of the visual world. Neuroimaging and neuropsychological data indicated that sensory cortex; especially the primary sensory cortex, is involved in imagery. In visual modality, neuroimaging data further indicated that during visual imagery, spatial information is mapped in the primary visual, providing strong evidences for depictive theory. In the auditory modality, behavioural studies also indicated that auditory imagery represents loudness and pitch of sound; this kind of neuroimaging evidence, however, is absent. The aim of the present study was to investigate the time course of auditory imagery processing, and to provide the neuroimaging evidence that imaginal auditory representations encode loudness and pitch information, using the ERP method and a cue-imagery (S1)-S2 paradigm. The results revealed that imagery effects started with an enhancement of the P2, probably indexing the top-down allocation of attention to the imagery task; and continued into a more positive-going late positive potentials (LPC), probably reflecting the formation of auditory imagery. The amplitude of this LPC was inversely related to the pitch of the imagined sound, but directly related to the loudness of the imagined sound, which were consistent with auditory perception related N1 component, providing evidences that auditory imagery encodes pitch and loudness information. When the S2 showed difference in pitch of loudness from the previously imagined S1, the behavioral performance were significantly worse and accordingly a conflict related N2 was elicited; and the high conflict elicited greater N2 amplitude than low conflict condition, providing further evidences that imagery is analog of perception and can encode pitch and loudness information. The present study suggests that imagery starts with an mechanism of top-down allocation of attention to the imagery task; and continuing into the step of imagery formation during which the physical features of the imagined stimulus can be encoded, providing supports to Kosslyn’s depictive representation theory.
Resumo:
Objective. To compare the voice performance of children involved in street labor with regular children using perceptual-auditory and acoustic analyses.Methods. A controlled cross-sectional study was carried out on 7- to 10-year-old children of both genders. Children from both groups lived with their families and attended school regularly; however, child labor was evident in one group and not the other. A total of 200 potentially eligible street children, assisted by the Child Labor Elimination Programme (PETI), and 400 regular children were interviewed. Those with any vocal discomfort (106, 53% and 90, 22.5%) had their voices assessed for resonance, pitch, loudness, speech rate, maximum phonation time, and other acoustic measurements.Results. A total of 106 street children (study group [SG]) and 90 regular children (control group [CG]) were evaluated. the SG group demonstrated higher oral and nasal resonance, reduced loudness, a lower pitch, and a slower speech rate than the CG. the maximum phonation time, fundamental frequency, and upper harmonics were higher in the SG than the CG. Jitter and shimmer were higher in the CG than the SG.Conclusion. Using perceptual-auditory and acoustic analyses, we determined that there were differences in voice performance between the two groups, with street children having better quality perceptual and acoustic vocal parameters than regular children. We believe that this is due to the procedures and activities performed by the Child Labor Elimination Program (PETI), which helps children to cope with their living conditions.
Resumo:
Karaoke singing is a popular form of entertainment in several parts of the world. Since this genre of performance attracts amateurs, the singing often has artifacts related to scale, tempo, and synchrony. We have developed an approach to correct these artifacts using cross-modal multimedia streams information. We first perform adaptive sampling on the user's rendition and then use the original singer's rendition as well as the video caption highlighting information in order to correct the pitch, tempo and the loudness. A method of analogies has been employed to perform this correction. The basic idea is to manipulate the user's rendition in a manner to make it as similar as possible to the original singing. A pre-processing step of noise removal due to feedback and huffing also helps improve the quality of the user's audio. The results are described in the paper which shows the effectiveness of this multimedia approach.
Resumo:
This thesis explores the possibilities of spatial hearing in relation to sound perception, and presents three acousmatic compositions based on a musical aesthetic that emphasizes this relation in musical discourse. The first important characteristic of these compositions is the exclusive use of sine waves and other time invariant sound signals. Even though these types of sound signals present no variations in time, it is possible to perceive pitch, loudness, and tone color variations as soon as they move in space due to acoustic processes involved in spatial hearing. To emphasize the perception of such variations, this thesis proposes to divide a tone in multiple sound units and spread them in space using several loudspeakers arranged around the listener. In addition to the perception of sound attribute variations, it is also possible to create rhythm and texture variations that depend on how sound units are arranged in space. This strategy permits to overcome the so called "sound surrogacy" implicit in acousmatic music, as it is possible to establish cause-effect relations between sound movement and the perception of sound attribute, rhythm, and texture variations. Another important consequence of using sound fragmentation together with sound spatialization is the possibility to produce diffuse sound fields independently from the levels of reverberation of the room, and to create sound spaces with a certain spatial depth without using any kind of artificial sound delay or reverberation.
Resumo:
Portugal has strong musical traditions, which have been perpetrated by decades through folkloristic activities. In folk groups from Alto Minho (north of Portugal), folk singing is mostly performed by cantadeiras, amateur female solo singers who learn this style orally. Their vocal characteristics are distinctive when compared with other regions of the country; however, deep understanding of these vocal practices is still missing. The present work aims at studying Alto Minho cantadeira’s vocal performance in a multidimensional perspective, envisioning social, cultural and physiological understanding of this musical style. Thus, qualitative and quantitative data analyses were carried out, to: (i) describe current performance practices, (ii) explore existent perceptions about most relevant voice features in this region, (iii) investigate physiological and acoustic properties of this style, and (iv) compare this style of singing with other non-classical singing styles of other countries. Dataset gathered involved: 78 groups whose members were telephone interviewed, 13 directors who were asked to fill in a questionnaire on performance practices, 1 cantadeira in a pilot study, 16 cantadeiras in preliminary voice recordings, 77 folk group members in listening tests, and 10 cantadeiras in multichannel recordings, including audio, ELG, air flow and intra-oral pressure signals. Data were analysed through thematic content analysis, descriptive and inferential statistics, hierarchical principal components, and multivariate linear regression models. Most representative voices have a high pitched and loud voice, with a bright timbre, predominance of chest register without excessive effort, and good text intelligibility with regional accent. High representativeness levels were obtained by few cantadeiras; these sing with high levels of subglottal pressure and vocal fold contact quotient, predominance of high spectrum energy and vocal loudness, corroborating indications of prevalence of pressed phonation. These vocal characteristics resemble belting in musical theatre and share similarities with country (USA) and ojikanje (Croatia) singing. Strategies that may contribute to the preservation of this type of singing and the vocal health of current cantadeiras are discussed, pointing at the direction of continuous education among folk groups, following practices that are already adopted elsewhere in Europe.
Resumo:
Le diagnostic de l’acouphène repose sur le rapport verbal du patient. Cependant, les paramètres psychoacoustiques, tels que la hauteur et la sonie de l’acouphène, sont essentiels pour l’évaluation et pour discriminer les fausses plaintes. Quantifier le percept de l’acouphène reste un objectif de recherche important. Nous avons cherché à: (1) évaluer la précision d'une nouvelle méthode d'évaluation des acouphènes du type « likeness rating » avec une méthode de présentation continue de la hauteur, en considérant la formation musicale, et (2) vérifier si les mesures psychoacoustiques de l’acouphène ont la sensibilité et la spécificité pour détecter les personnes simulant un acouphène. Nous avons recruté des musiciens et des non-musiciens souffrant d'acouphènes et des simulateurs sans acouphènes. La plupart d’entre eux ont été testés une seconde fois quelques semaines plus tard. La hauteur de l’acouphène a d’abord été évaluée en utilisant la méthode « likness rating ». Des sons purs ont été présentés de manière aléatoire de 0.25 kHz à 16 kHz et les participants devaient coter la ressemblance de chaque son par rapport à leur acouphène, et en ajuster son volume de 0 à 100 dB SPL. La hauteur a ensuite été évaluée par une méthode, où les participants devaient apparier la hauteur de leur acouphène en déplaçant leur doigt sur une bande tactile générant des sons purs en continu de 0.5 kHz à 20 kHz par pas de 1 Hz. Les capacités à apparier des sons externes ont été évaluées en utilisant cet appareil. La hauteur prédominante de l’acouphène était similaire entre les deux méthodes pour les musiciens et les non-musiciens, bien que les musiciens montraient de meilleures capacités d’appariement face à des sons externes. Les simulateurs ont coté la sonie bien au-dessus que les autres groupes avec un grand degré de spécificité (94,4%), et ont été inconsistants dans l’appariement de cette sonie (pas de la hauteur) d’une session à une autre. Les données de la seconde session montrent des réponses similaires pour l’appariement de la hauteur pour les deux méthodes ainsi que pour tous nos participants. En conclusion, la hauteur et la sonie correspondent au percept de l’acouphène et doivent en plus être utilisées avec des échelles visuelles analogiques, qui sont plus corrélées avec la gêne et la détresse. Enfin, l’appariement de la sonie est sensible et spécifique à déterminer la présence de l’acouphène.
Resumo:
Prosody is an important feature of language, comprising intonation, loudness, and tempo. Emotional prosodic processing forms an integral part of our social interactions. The main aim of this study was to use bold contrast fMRI to clarify the normal functional neuroanatomy of emotional prosody, in passive and active contexts. Subjects performed six separate scanning studies, within which two different conditions were contrasted: (1) "pure" emotional prosody versus rest; (2) congruent emotional prosody versus 'neutral' sentences; (3) congruent emotional prosody versus rest; (4) incongruent emotional prosody versus rest; (5) congruent versus incongruent emotional prosody; and (6) an active experiment in which subjects were instructed to either attend to the emotion conveyed by semantic content or that conveyed by tone of voice. Data resulting from these contrasts were analysed using SPM99. Passive listening to emotional prosody consistently activated the lateral temporal lobe (superior and/or middle temporal gyri). This temporal lobe response was relatively right-lateralised with or without semantic information. Both the separate and direct comparisons of congruent and incongruent emotional prosody revealed that subjects used fewer brain regions to process incongruent emotional prosody than congruent. The neural response to attention to semantics, was left lateralised, and recruited an extensive network not activated by attention to emotional prosody. Attention to emotional prosody modulated the response to speech, and induced right-lateralised activity, including the middle temporal gyrus. In confirming the results of lesion and neuropsychological studies, the current study emphasises the importance of the right hemisphere in the processing of emotional prosody, specifically the lateral temporal lobes. (C) 2003 Elsevier Science Ltd. All rights reserved.
Resumo:
While the beneficial effect of levodopa on traditional motor control tasks have been well documented over the decades. its effect on speech motor control has rarely been objectively examined and the existing literature remains inconclusive. This paper aims to examine the effect of levodopa on speech in patients with Parkinson's disease. It was hypothesized that levodopa would improve preparatory motor set related activity and alleviate hypophonia. Patients fasted and abstained from levodopa overnight. Motor examination and speech testing was performed the following day, pre-levodopa during their "off' state, then at hourly intervals post-medication to obtain the best "on" state. All speech stimuli showed a consistent tendency for increased loudness and faster rate during the "on" state, but this was accompanied by a greater extent of intensity decay. Pitch and articulation remained unchanged. Levodopa effectively upscaled the overall gain setting of vocal amplitude and tempo, similar to its well-known effect on limb movement. However, unlike limb movement, this effect on the final acoustic product of speech may or may not be advantageous, depending on the existing speech profile of individual patients. (C) 2007 Movement Disorder Society.