57 resultados para loudness
Resumo:
Listeners can attend to one of several simultaneous messages by tracking one speaker’s voice characteristics. Using differences in the location of sounds in a room, we ask how well cues arising from spatial position compete with these characteristics. Listeners decided which of two simultaneous target words belonged in an attended “context” phrase when it was played simultaneously with a different “distracter” context. Talker difference was in competition with position difference, so the response indicates which cue‐type the listener was tracking. Spatial position was found to override talker difference in dichotic conditions when the talkers are similar (male). The salience of cues associated with differences in sounds, bearings decreased with distance between listener and sources. These cues are more effective binaurally. However, there appear to be other cues that increase in salience with distance between sounds. This increase is more prominent in diotic conditions, indicating that these cues are largely monaural. Distances between spectra calculated using a gammatone filterbank (with ERB‐spaced CFs) of the room’s impulse responses at different locations were computed, and comparison with listeners’ responses suggested some slight monaural loudness cues, but also monaural “timbre” cues arising from the temporal‐ and spectral‐envelope differences in the speech from different locations.
Resumo:
Parkinson's disease (PD) is a degenerative illness whose cardinal symptoms include rigidity, tremor, and slowness of movement. In addition to its widely recognized effects PD can have a profound effect on speech and voice.The speech symptoms most commonly demonstrated by patients with PD are reduced vocal loudness, monopitch, disruptions of voice quality, and abnormally fast rate of speech. This cluster of speech symptoms is often termed Hypokinetic Dysarthria.The disease can be difficult to diagnose accurately, especially in its early stages, due to this reason, automatic techniques based on Artificial Intelligence should increase the diagnosing accuracy and to help the doctors make better decisions. The aim of the thesis work is to predict the PD based on the audio files collected from various patients.Audio files are preprocessed in order to attain the features.The preprocessed data contains 23 attributes and 195 instances. On an average there are six voice recordings per person, By using data compression technique such as Discrete Cosine Transform (DCT) number of instances can be minimized, after data compression, attribute selection is done using several WEKA build in methods such as ChiSquared, GainRatio, Infogain after identifying the important attributes, we evaluate attributes one by one by using stepwise regression.Based on the selected attributes we process in WEKA by using cost sensitive classifier with various algorithms like MultiPass LVQ, Logistic Model Tree(LMT), K-Star.The classified results shows on an average 80%.By using this features 95% approximate classification of PD is acheived.This shows that using the audio dataset, PD could be predicted with a higher level of accuracy.
Resumo:
The aim of this study was to perform voice evaluation in teachers with and without vocal symptoms, identifying etiologic factors of dysphonia, voice symptoms, vocal qualities, and laryngeal lesions. Eighty teachers were divided into two groups: GI (without or sporadic symptoms, 40) and GII (with frequent vocal symptoms, 40). They answered a specific questionnaire, and were subject to a perceptual vocal assessment (maximum phonation time, glottal attack, resonance, coordination of breathing and voicing, pitch, and loudness), GIRBAS scale, and to videolaryngoscopy. Females were predominant in both groups, and the age range was from 36 to 50 years. Elementary teachers predominated, working in classes with 31-40 students. Voice symptoms and alterations in the perceptual vocal analysis and in the GIRBAS scale were more frequent in GII In 46 teachers (GI-16; GII-30), videolaryngoscopy exams were abnormal with the vocal nodules being the most frequent lesions. These results indicate that a teacher's voice is compromised, and requires more attention including control of environmental factors and associated diseases, preventive vocal hygiene, periodic laryngeal examinations, and access to adequate specialist treatment.
Resumo:
This study investigates the possible differences between actors' and nonactors' vocal projection strategies using acoustic and perceptual analyses. A total of 11 male actors and 10 male nonactors volunteered as subjects, reading an extended text sample in habitual, moderate, and loud levels. The samples were analyzed for sound pressure level (SPL), alpha ratio (difference between the average SPL of the 1-5 kHz region and the average SPL of the 50 Hz-1 kHz region), fundamental frequency (F0), and long-term average spectrum (LTAS). Through LTAS, the mean frequency of the first formant (171) range, the mean frequency of the actor's formant, the level differences between the F1 frequency region and the F0 region (L1-L0), and the level differences between the strongest peak at 0-1 kHz and that at 3-4 kHz were measured. Eight voice specialists evaluated perceptually the degree of projection, loudness, and tension in the samples. The actors had a greater alpha ratio, stronger level of the actor's formant range, and a higher degree of perceived projection and loudness in all loudness levels. SPL, however, did not differ significantly between the actors and nonactors, and no differences were found in the mean formant frequencies ranges. The alpha ratio and the relative level of the actor's formant range seemed to be related to the degree of perceived loudness. From the physiological point of view, a more favorable glottal setting' providing a higher glottal closing speed, may be characteristic of these actors' projected voices. So, the projected voices, in this group of actors, were more related to the glottic source than to the resonance of the vocal tract.
Electroglottographic analysis of actresses and nonactresses' voices in different levels of intensity
Resumo:
Background: Previous studies with long-term average spectrum (LTAS) showed the importance of the glottal source for understanding the projected voices of actresses. In this study, electroglottographic (EGG) analysis was used to investigate the contribution of the glottal source to the projected voice, comparing actresses and nonactresses' voices, in different levels of intensity. Method: Thirty actresses and 30 nonactresses sustained vowels in habitual, moderate, and loud intensity levels. The EGG variables were contact quotient (CQ), closing quotient (QCQ), and opening quotient (QOQ). Other variables were sound pressure level (SPL) and fundamental frequency (F0). A KayPENTAX EGG was used. Variables were inputted in a general linear model. Results/Discussion: Actresses showed significantly higher values for SPL, in all levels, and both groups increased SPL significantly while changing from habitual to moderate and further to loud. There were no significant differences between groups for EGG quotients. There were significant differences between the levels only for F0 and CQ for both groups. Conclusion: SPL was significantly higher among actresses in all intensity levels, but in the EGG analysis, no differences were found. This apparently weak contribution of the glottal source in the supposedly projected voices of actresses, contrary to previous LTAS studies, might be because of a higher subglottal pressure or perhaps greater vocal tract contribution in SPL. Results from the present study suggest that trained subjects did not produce a significant higher SPL than untrained individuals by increasing the cost in terms of higher vocal fold collision and hence more impact stress. Future researches should explore the difference between trained and nontrained voices by aerodynamic measurements to evaluate the relationship between physiologic findings and the acoustic and EGG data. Moreover, further studies should consider both types of vocal tasks, sustained vowel and running speech, for both EGG and LTAS analysis. © 2013 The Voice Foundation.
Resumo:
A disfonia espasmódica é um distúrbio vocal grave, caracterizada por espasmos musculares laríngeos durante a fonação, produzindo voz tensa e estrangulada. Atualmente, acredita-se que tenha origem neurológica, mas há evidências que associam suas causas com alterações psicogênicas. O presente estudo visa descrever dois casos diagnosticados como disfonia espasmódica, com manifestações fonoaudiológicas e psicológicas distintas. O estudo descritivo foi realizado por meio de análise dos prontuários, incluindo avaliação otorrinolaringológica, neurológica, avaliação vocal perceptivo-auditiva, análise acústica e Protocolo de Qualidade de Vida e Voz (QVV). Caso 1: sexo feminino, 65 anos, discreta coaptação de bandas ventriculares, fenda fusiforme e intenso tremor laríngeo. Apresentou quadro de alucinação, diagnosticada, posteriormente, como Esquizofrenia. Faz uso de Risperidona e Pondera. Apresentou loudness rebaixado, pitch agravado, rouquidão, hipernasalidade leve, tempo máximo de fonação reduzido, incoordenação pneumofonoarticulátoria e G², R², Bº, Aº, S³. Caso 2: sexo feminino, 68 anos, alteração das pregas vocais e intenso estrangulamento de bandas ventriculares. Sofreu três acidentes vasculares encefálicos (AVE) e, após o primeiro, iniciou-se a rouquidão. Realiza atendimento psicológico, ainda sem diagnóstico fechado. Apresentou loudness rebaixado, pitch agudizado, diplofonia durante a conversação, tempo máximo de fonação reduzido, incoordenação pneumofonoarticulatória e G³, R², Bº, Aº, S³. Os dados demonstraram distintos graus de severidades nas manifestações otorrinolaringológicas e fonoaudiológicas, além de diferentes envolvimentos psicogênicos. O grau de severidade ficou implícito no impacto da voz na qualidade de vida das pacientes estudadas.
Resumo:
Purpose. The present study aimed to compare actors/actresses's voices and vocally trained subjects through aerodynamic and electroglottographic (EGG) analyses. We hypothesized that glottal and breathing functions would reflect technical and physiological differences between vocally trained and untrained subjects.Methods. Forty participants with normal voices participated in this study (20 professional theater actors and 20 untrained participants). In each group, 10 male and 10 female subjects were assessed. All participants underwent aerodynamic and EGG assessment of voice. From the Phonatory Aerodynamic System, three protocols were used: comfortable sustained phonation with EGG, voice efficiency with EGG, and running speech. Contact quotient was calculated from EGG. All phonatory tasks were produced at three different loudness levels. Mean sound pressure level and fundamental frequency were also assessed. Univariate, multivariate, and correlation statistical analyses were performed.Results. Main differences between vocally trained and untrained participants were found in the following variables: mean sound pressure level, phonatory airflow, subglottic pressure, inspiratory airflow duration, inspiratory airflow, and inspiratory volume. These variables were greater for trained participants. Mean pitch was found to be lower for trained voices.Conclusions. The glottal source seemed to have a weak contribution when differentiating the training status in speaking voice. More prominent changes between vocally trained and untrained participants are demonstrated in respiratory-related variables. These findings may be related to better management of breathing function (better breath support).
Resumo:
The present study aimed to compare elderly and young female voices in habitual and high intensity. The effect of increased intensity on the acoustic and perceptual parameters was assessed. Sound pressure level, fundamental frequency, jitter, shimmer, and harmonic to noise ratio were obtained at habitual and high intensity voice in a group of 30 elderly women and 30 young women. Perceptual assessment was also performed. Both groups demonstrated an increase in sound pressure level and fundamental frequency from habitual voice to high intensity voice. No differences were found between groups in any acoustic variables on samples recorded with habitual intensity level. No significant differences between groups were found in habitual intensity level for pitch, hoarseness, roughness, and breathiness. Asthenia and instability obtained significant higher values in elderly than young participants, whereas, the elderly demonstrated lower values for perceived tension and loudness than young subjects. Acoustic and perceptual measures do not demonstrate evident differences between elderly and young speakers in habitual intensity level. The parameters analyzed may lack the sensitivity necessary to detect differences in subjects with normal voices. Phonation with high intensity highlights differences between groups, especially in perceptual parameters. Therefore, high intensity should be included to compare elderly and young voice.
Resumo:
Objective: to compare acoustic and perceptual parameters regarding the voice of cochlear implanted children, with normal hearing children. Method: this is a cross-sectional, quantitative and qualitative study. Methods: Thirty six cochlear implanted children aged between 3y and 3 m to 5y and 9 m and 25 children with normal hearing, aged between 3y and 11 m and 6y and 6 m, participated in this study. The recordings and the acoustics analysis of the sustained vowel/a/and spontaneous speech were performed using the PRAAT program. The parameters analyzed for the sustained vowel were the mean of the fundamental frequency, jitter, shimmer and harmonic-to-noise ratio (HNR). For the spontaneous speech, the minimum and maximum frequencies and the number of semitones were extracted. The perceptual analysis of the speech material was analyzed using visual-analogical scales of 100 points, composing the aspects related to the overall severity of the vocal deviation, roughness, breathiness, strain, pitch, loudness and resonance deviation, and instability. This last parameter was only analyzed for the sustained vowel. Results: The results demonstrated that the majority of the vocal parameters analyzed in the samples of the implanted children disclosed values similar to those obtained by the group of children with normal hearing. Conclusion: implanted children who participate in a (re) habilitation and follow-up program, can present vocal characteristics similar to those vocal characteristics of children with normal hearing. (C) 2012 Elsevier Ireland Ltd. All rights reserved.
The gaseous environment of radio galaxies: a new perspective from high-resolution x-ray spectroscopy
Resumo:
It is known that massive black holes have a profound effect on the evolution of galaxies, and possibly on their formation by regulating the amount of gas available for the star formation. However, how black hole and galaxies communicate is still an open problem, depending on how much of the energy released interacts with the circumnuclear matter. In the last years, most studies of feedback have primarily focused on AGN jet/cavity systems in the most massive galaxy clusters. This thesis intends to investigate the feedback phenomena in radio--loud AGNs from a different perspective studying isolated radio galaxies, through high-resolution spectroscopy. In particular one NLRG and three BLRG are studied, searching for warm gas, both in emission and absorption, in the soft X-ray band. I show that the soft spectrum of 3C33 originates from gas photoionized by the central engine. I found for the first time WA in 3C382 and 3C390.3. I show that the observed warm emitter/absorbers is not uniform and probably located in the NLR. The detected WA is slow implying a mass outflow rate and kinetic luminosity always well below 1% the L(acc) as well as the P(jet). Finally the radio--loud properties are compared with those of type 1 RQ AGNs. A positive correlation is found between the mass outflow rate/kinetic luminosity, and the radio loudness. This seems to suggest that the presence of a radio source (the jet?) affects the distribution of the absorbing gas. Alternatively, if the gas distribution is similar in Seyferts and radio galaxies, the M(out) vs rl relation could simply indicate a major ejection of matter in the form of wind in powerful radio AGNs.
Resumo:
In this prospective multicenter study, tinnitus loudness and tinnitus-related distress were investigated in 174 cochlear implant (CI) candidates who underwent CI surgery at a Swiss cochlear implant center. All subjects participated in two session, one preoperatively and one 6 months after device activation. In both sessions, tinnitus loudness was assessed using a visual analogue scale and tinnitus distress using a standardized tinnitus questionnaire. The data were compared with unaided pre- and postoperative pure tone thresholds, and postoperative speech reception scores. 71.8% of the subjects reported tinnitus preoperatively. Six months after CI surgery 20.0% of these reported abolition of their tinnitus, 51.2% a subjective improvement, 21.6% no change and 7.2% a deterioration. Of the 49 (28.2%) subjects with no tinnitus preoperatively, 5 developed tinnitus 6 months after CI. These 5 had poorer speech understanding after CI surgery with their device than the group who remained tinnitus free. We found no correlation between tinnitus improvement, age, duration of tinnitus, or change in unaided hearing thresholds between the two sessions.
Resumo:
From the customer satisfaction point of view, sound quality of any product has become one of the important factors these days. The primary objective of this research is to determine factors which affect the acceptability of impulse noise. Though the analysis is based on a sample impulse sound file of a Commercial printer, the results can be applied to other similar impulsive noise. It is assumed that impulsive noise can be tuned to meet the accepTable criteria. Thus it is necessary to find the most significant factors which can be controlled physically. This analysis is based on a single impulse. A sample impulsive sound file is tweaked for different amplitudes, background noise, attack time, release time and the spectral content. A two level factorial design of experiments (DOE) is applied to study the significant effects and interactions. For each impulse file modified as per the DOE, the magnitude of perceived annoyance is calculated from the objective metric developed recently at Michigan Technological University. This metric is based on psychoacoustic criteria such as loudness, sharpness, roughness and loudness based impulsiveness. Software called ‘Artemis V11.2’ developed by HEAD Acoustics is used to calculate these psychoacoustic terms. As a result of two level factorial analyses, a new objective model of perceived annoyance is developed in terms of above mentioned physical parameters such as amplitudes, background noise, impulse attack time, impulse release time and the spectral content. Also the effects of the significant individual factors as well as two level interactions are also studied. The results show that all the mentioned five factors affect annoyance level of an impulsive sound significantly. Thus annoyance level can be reduced under the criteria by optimizing the levels. Also, an additional analysis is done to study the effect of these five significant parameters on the individual psychoacoustic metrics.
Resumo:
The present study was designed to elucidate sex-related differences in two basic auditory and one basic visual aspect of sensory functioning, namely sensory discrimination of pitch, loudness, and brightness. Although these three aspects of sensory functioning are of vital importance in everyday life, little is known about whether men and women differ from each other in these sensory functions. Participants were 100 male and 100 female volunteers ranging in age from 18 to 30 years. Since sensory sensitivity may be positively related to individual levels of intelligence and musical experience, measures of psychometric intelligence and musical background were also obtained. Reliably better performance for men compared to women was found for pitch and loudness, but not for brightness discrimination. Furthermore, performance on loudness discrimination was positively related to psychometric intelligence, while pitch discrimination was positively related to both psychometric intelligence and levels of musical training. Additional regression analyses revealed that each of three predictor variables (sex, psychometric intelligence, and musical training) accounted for a statistically significant portion of unique variance in pitch discrimination. With regard to loudness discrimination, regression analysis yielded a statistically significant portion of unique variance for sex as a predictor variable, whereas psychometric intelligence just failed to reach statistical significance. The potential influence of sex hormones on sex-related differences in sensory functions is discussed.
Resumo:
OBJECTIVES To establish whether complex signal processing is beneficial for users of bone anchored hearing aids. METHODS Review and analysis of two studies from our own group, each comparing a speech processor with basic digital signal processing (either Baha Divino or Baha Intenso) and a processor with complex digital signal processing (either Baha BP100 or Baha BP110 power). The main differences between basic and complex signal processing are the number of audiologist accessible frequency channels and the availability and complexity of the directional multi-microphone noise reduction and loudness compression systems. RESULTS Both studies show a small, statistically non-significant improvement of speech understanding in quiet with the complex digital signal processing. The average improvement for speech in noise is +0.9 dB, if speech and noise are emitted both from the front of the listener. If noise is emitted from the rear and speech from the front of the listener, the advantage of the devices with complex digital signal processing as opposed to those with basic signal processing increases, on average, to +3.2 dB (range +2.3 … +5.1 dB, p ≤ 0.0032). DISCUSSION Complex digital signal processing does indeed improve speech understanding, especially in noise coming from the rear. This finding has been supported by another study, which has been published recently by a different research group. CONCLUSIONS When compared to basic digital signal processing, complex digital signal processing can increase speech understanding of users of bone anchored hearing aids. The benefit is most significant for speech understanding in noise.