988 resultados para Speech Rate
Resumo:
OBJETIVO: analisar os fatores gênero, idade, tipo de surgimento da gagueira, tempo de duração e tipologia das disfluências, fatores estressantes físicos e emocionais, e fatores comunicativos e qualitativos associados em crianças disfluentes sem recorrência familial do distúrbio. MÉTODO: participaram 43 crianças com alto risco para a gagueira de ambos os gêneros. A coleta de dados foi realizada por meio do Protocolo de Risco para a Gagueira do Desenvolvimento - PRGD. RESULTADOS: a razão masculino/feminino foi de 3,3:1. A única diferença estatisticamente significante dos fatores de risco analisados nos gêneros masculino e feminino foi a maior ocorrência de fatores comunicativos associados no gênero masculino (p=0,003). Houve uma semelhança dos achados entre os meninos e as meninas: quanto ao tempo de duração das disfluências a maioria apresentou mais de 12 meses de duração, a tipologia gaga foi a mais freqüente, a presença de fatores estressantes emocionais ocorreu na maior parte das crianças, e finalmente os fatores qualitativos associados, como taxa de elocução aumentada, tensão visível e incoordenação pneumo-fono-articulatória estiveram presentes em grande parte da amostra. CONCLUSÃO: os resultados desta investigação permitiram concluir que nos casos de crianças com alto risco para a gagueira isolada ocorreu a interação de inúmeros fatores, sugerindo que o distúrbio é multifatorial. Também foi possível concluir que a interação de alguns fatores como gênero masculino, tipologia gaga manifestada por mais de 12 meses, com início persistente, na presença de fatores qualitativos e comunicativos associados pode representar risco maior para o desenvolvimento da gagueira persistente.
Resumo:
Pós-graduação em Fonoaudiologia - FFC
Resumo:
PURPOSE: to evaluate and compare the fluency between the familial and the sporadic persistent developmental stuttering, characterizing the typology and the frequency of the disfluencies, the speech rate and the severity of the stuttering. METHOD: 40 participants aged from 6 to 42 years old, divided in two groups with twenty participants in each one: Familial Persistent Developmental Stuttering and Sporadic Persistent Developmental Stuttering. The procedures used were: clinical and familial history, assessment of fluency and Stuttering Severity Instrument. RESULTS: there were no statistically significant differences between the groups regarding the frequency of stuttering like disfluencies, the flow of syllables and words per minute and the severity of stuttering. It was noted a tendency of the group with familial stuttering to show a bigger variability of the severity of stuttering, going from mild to very severe, whereas in the group with sporadic stuttering, the severity varied from mild to severe. CONCLUSION: this study represents the first effort to the characterization of the speech fluency profile of the subgroups of people who stutter, namely familial persistent developmental stuttering and sporadic persistent developmental stuttering. It is possible to conclude that the speech fluency profile of people who stutter, independently of the familial history, is similar. It is noteworthy that the occurrence of some stuttering-like disfluencies, monosyllabic word repetition, block and intrusion were different between groups.
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
One of the most controversial issues in the history of Phonetics is the discussion about the typology of speech rhythm. Out of the Greek and Latin tradition on poetry versification, the notion of rhythm has been misunderstood as speech rate. In the early years of the 20th century, a dichotomy merged classifying the speech rhythm into stress-timed and syllable-timed languages, inspired by the old theory of poetic versification. Following the same old pattern, later on, a third type of language were proposed: the moraic languages, initially attributed only to Japanese. With the facilities to carry on acoustic research, in the second half of the 20th century, the typology of language rhythm came to a dead end. Different types of language were set out. This paper discusses these ideas, showing a great misunderstanding among researchers in relation to the characterization of a syllable-timed language. The notion of mora is revisited and its role in the study of speech rhythm is better defined.
Resumo:
Two physiological assessments, electromagnetic articulography (EMA) and electropalatography (EPG), were used simultaneously to investigate the articulatory dynamics in an 18-year-old male with dysarthria 9 years following traumatic brain injury (TBI). Eight words consisting of /t, s, integral, k/ in word initial and word final positions were produced up to 10 times. A nonneurologically impaired male served as a control subject. Six parameters were analyzed using EMA: velocity, acceleration, deceleration, distance, duration, and motion path of tongue movements. Using EPG, the pattern and amount of tongue-to-palate contact and the duration of the closure/constriction phase of each consonant produced were assessed. Timing disturbances in the TBI speaker's speech were highlighted in perceptual assessments in the form of prolonged phonemes and a reduced speech rate. EMA analysis revealed that the approach and release phase durations of the consonant productions were within normal limits. Kinematic strategies such as decreased velocity and decreased distances traveled by the tongue, however, may have counterbalanced each other to produce these appropriate results. EPG examination revealed significantly longer closure/constriction phase periods, which may have contributed to the prolonged phonemes and reduced speech rate observed. The implications of these findings for the development of treatment programs for dysarthria subsequent to TBI will be highlighted.
Resumo:
This study examined the role of global processing speed in mediating age increases in auditory memory span in 5- to 13-year-olds. Children were tested on measures of memory span, processing speed, single-word speech rate, phonological sensitivity, and vocabulary. Structural equation modeling supported a model in which age-associated increases in processing speed predicted the availability of long-term memory phonological representations for redintegration processes. The availability of long-term phonological representations, in turn, explained variance in memory span. Maximum speech rate did not predict independent variance in memory span. (c) 2005 Elsevier Inc. All rights reserved.
Resumo:
Over recent years, evidence has been accumulating in favour of the importance of long-term information as a variable which can affect the success of short-term recall. Lexicality, word frequency, imagery and meaning have all been shown to augment short term recall performance. Two competing theories as to the causes of this long-term memory influence are outlined and tested in this thesis. The first approach is the order-encoding account, which ascribes the effect to the usage of resources at encoding, hypothesising that word lists which require less effort to process will benefit from increased levels of order encoding, in turn enhancing recall success. The alternative view, trace redintegration theory, suggests that order is automatically encoded phonologically, and that long-term information can only influence the interpretation of the resultant memory trace. The free recall experiments reported here attempted to determine the importance of order encoding as a facilitatory framework and to determine the locus of the effects of long-term information in free recall. Experiments 1 and 2 examined the effects of word frequency and semantic categorisation over a filled delay, and experiments 3 and 4 did the same for immediate recall. Free recall was improved by both long-term factors tested. Order information was not used over a short filled delay, but was evident in immediate recall. Furthermore, it was found that both long-term factors increased the amount of order information retained. Experiment 5 induced an order encoding effect over a filled delay, leaving a picture of short-term processes which are closely associated with long-term processes, and which fit conceptions of short-term memory being part of language processes rather better than either the encoding or the retrieval-based models. Experiments 6 and 7 aimed to determine to what extent phonological processes were responsible for the pattern of results observed. Articulatory suppression affected the encoding of order information where speech rate had no direct influence, suggesting that it is ease of lexical access which is the most important factor in the influence of long-term memory on immediate recall tasks. The evidence presented in this thesis does not offer complete support for either the retrieval-based account or the order encoding account of long-term influence. Instead, the evidence sits best with models that are based upon language-processing. The path urged for future research is to find ways in which this diffuse model can be better specified, and which can take account of the versatility of the human brain.
Resumo:
Common approaches to IP-traffic modelling have featured the use of stochastic models, based on the Markov property, which can be classified into black box and white box models based on the approach used for modelling traffic. White box models, are simple to understand, transparent and have a physical meaning attributed to each of the associated parameters. To exploit this key advantage, this thesis explores the use of simple classic continuous-time Markov models based on a white box approach, to model, not only the network traffic statistics but also the source behaviour with respect to the network and application. The thesis is divided into two parts: The first part focuses on the use of simple Markov and Semi-Markov traffic models, starting from the simplest two-state model moving upwards to n-state models with Poisson and non-Poisson statistics. The thesis then introduces the convenient to use, mathematically derived, Gaussian Markov models which are used to model the measured network IP traffic statistics. As one of the most significant contributions, the thesis establishes the significance of the second-order density statistics as it reveals that, in contrast to first-order density, they carry much more unique information on traffic sources and behaviour. The thesis then exploits the use of Gaussian Markov models to model these unique features and finally shows how the use of simple classic Markov models coupled with use of second-order density statistics provides an excellent tool for capturing maximum traffic detail, which in itself is the essence of good traffic modelling. The second part of the thesis, studies the ON-OFF characteristics of VoIP traffic with reference to accurate measurements of the ON and OFF periods, made from a large multi-lingual database of over 100 hours worth of VoIP call recordings. The impact of the language, prosodic structure and speech rate of the speaker on the statistics of the ON-OFF periods is analysed and relevant conclusions are presented. Finally, an ON-OFF VoIP source model with log-normal transitions is contributed as an ideal candidate to model VoIP traffic and the results of this model are compared with those of previously published work.
Resumo:
Although persuasion often occurs via oral communication, it remains a comparatively understudied area. This research tested the hypothesis that changes in three properties of voice influence perceptions of speaker confidence, which in turn differentially affects attitudes according to different underlying psychological processes that the Elaboration Likelihood Model (ELM, Petty & Cacioppo, 1984), suggests should emerge under different levels of thought. Experiment 1 was a 2 (Elaboration: high vs. low) x 2 (Vocal speed: increased speed vs. decreased speed) x 2 (Vocal intonation: falling intonation vs. rising intonation) between participants factorial design. Vocal speed and vocal intonation influenced perceptions of speaker confidence as predicted. In line with the ELM, under high elaboration, confidence biased thought favorability, which in turn influenced attitudes. Under low elaboration, confidence did not bias thoughts but rather directly influenced attitudes as a peripheral cue. Experiment 2 used a similar design as Experiment 1 but focused on vocal pitch. Results confirmed pitch influenced perceptions of confidence as predicted. Importantly, we also replicated the bias and cue processes found in Experiment 1. Experiment 3 investigated the process by which a broader spectrum of speech rate influenced persuasion under moderate elaboration. In a 2 (Argument quality: strong vs. weak) x 4 (Vocal speed: extremely slow vs. moderately slow vs. moderately fast vs. extremely fast) between participants factorial design, results confirmed the hypothesized non-linear relationship between speech rate and perceptions of confidence. In line with the ELM, speech rate influenced persuasion based on the amount of processing. Experiment 4 investigated the effects of a broader spectrum of vocal intonation on persuasion under moderate elaboration and used a similar design as Experiment 3. Results indicated a partial success of our vocal intonation manipulation. No evidence was found to support the hypothesized mechanism. These studies show that changes in several different properties of voice can influence the extent to which others perceive them as confident. Importantly, evidence suggests different vocal properties influence persuasion by the same bias and cue processes under high and low thought. Evidence also suggests that under moderate thought, speech rate influences persuasion based on the amount of processing.
Resumo:
In the present study, Korean-English bilingual (KEB) and Korean monolingual (KM) children, between the ages of 8 and 13 years, and KEB adults, ages 18 and older, were examined with one speech perception task, called the Nonsense Syllable Confusion Matrix (NSCM) task (Allen, 2005), and two production tasks, called the Nonsense Syllable Imitation Task (NSIT) and the Nonword Repetition Task (NRT; Dollaghan & Campbell, 1998). The present study examined (a) which English sounds on the NSCM task were identified less well, presumably due to interference from Korean phonology, in bilinguals learning English as a second language (L2) and in monolinguals learning English as a foreign language (FL); (b) which English phonemes on the NSIT were more challenging for bilinguals and monolinguals to produce; (c) whether perception on the NSCM task is related to production on the NSIT, or phonological awareness, as measured by the NRT; and (d) whether perception and production differ in three age-language status groups (i.e., KEB children, KEB adults, and KM children) and in three proficiency subgroups of KEB children (i.e., English-dominant, ED; balanced, BAL; and Korean-dominant, KD). In order to determine English proficiency in each group, language samples were extensively and rigorously analyzed, using software, called Systematic Analysis of Language Transcripts (SALT). Length of samples in complete and intelligible utterances, number of different and total words (NDW and NTW, respectively), speech rate in words per minute (WPM), and number of grammatical errors, mazes, and abandoned utterances were measured and compared among the three initial groups and the three proficiency subgroups. Results of the language sample analysis (LSA) showed significant group differences only between the KEBs and the KM children, but not between the KEB children and adults. Nonetheless, compared to normative means (from a sample length- and age-matched database provided by SALT), the KEB adult group and the KD subgroup produced English at significantly slower speech rates than expected for monolingual, English-speaking counterparts. Two existing models of bilingual speech perception and production—the Speech Learning Model or SLM (Flege, 1987, 1992) and the Perceptual Assimilation Model or PAM (Best, McRoberts, & Sithole, 1988; Best, McRoberts, & Goodell, 2001)—were considered to see if they could account for the perceptual and production patterns evident in the present study. The selected English sounds for stimuli in the NSCM task and the NSIT were 10 consonants, /p, b, k, g, f, θ, s, z, ʧ, ʤ/, and 3 vowels /I, ɛ, æ/, which were used to create 30 nonsense syllables in a consonant-vowel structure. Based on phonetic or phonemic differences between the two languages, English sounds were categorized either as familiar sounds—namely, English sounds that are similar, but not identical, to L1 Korean, including /p, k, s, ʧ, ɛ/—or unfamiliar sounds—namely, English sounds that are new to L1, including /b, g, f, θ, z, ʤ, I, æ/. The results of the NSCM task showed that (a) consonants were perceived correctly more often than vowels, (b) familiar sounds were perceived correctly more often than unfamiliar ones, and (c) familiar consonants were perceived correctly more often than unfamiliar ones across the three age-language status groups and across the three proficiency subgroups; and (d) the KEB children perceived correctly more often than the KEB adults, the KEB children and adults perceived correctly more often than the KM children, and the ED and BAL subgroups perceived correctly more often than the KD subgroup. The results of the NSIT showed (a) consonants were produced more accurately than vowels, and (b) familiar sounds were produced more accurately than unfamiliar ones, across the three age-language status groups. Also, (c) familiar consonants were produced more accurately than unfamiliar ones in the KEB and KM child groups, and (d) unfamiliar vowels were produced more accurately than a familiar one in the KEB child group, but the reverse was true in the KEB adult and KM child groups. The KEB children produced sounds correctly significantly more often than the KM children and the KEB adults, though the percent correct differences were smaller than for perception. Production differences were not found among the three proficiency subgroups. Perception on the NSCM task was compared to production on the NSIT and NRT. Weak positive correlations were found between perception and production (NSIT) for unfamiliar consonants and sounds, whereas a weak negative correlation was found for unfamiliar vowels. Several correlations were significant for perceptual performance on the NSCM task and overall production performance on the NRT: for unfamiliar consonants, unfamiliar vowels, unfamiliar sounds, consonants, vowels, and overall performance on the NSCM task. Nonetheless, no significant correlation was found between production on the NSIT and NRT. Evidently these are two very different production tasks, where immediate imitation of single syllables on the NSIT results in high performance for all groups. Findings of the present study suggest that (a) perception and production of L2 consonants differ from those of vowels; (b) perception and production of L2 sounds involve an interaction of sound type and familiarity; (c) a weak relation exists between perception and production performance for unfamiliar sounds; and (d) L2 experience generally predicts perceptual and production performance. The present study yields several conclusions. The first is that familiarity of sounds is an important influence on L2 learning, as claimed by both SLM and PAM. In the present study, familiar sounds were perceived and produced correctly more often than unfamiliar ones in most cases, in keeping with PAM, though experienced L2 learners (i.e., the KEB children) produced unfamiliar vowels better than familiar ones, in keeping with SLM. Nonetheless, the second conclusion is that neither SLM nor PAM consistently and thoroughly explains the results of the present study. This is because both theories assume that the influence of L1 on the perception of L2 consonants and vowels works in the same way as for production of them. The third and fourth conclusions are two proposed arguments: that perception and production of consonants are different than for vowels, and that sound type interacts with familiarity and L2 experience. These two arguments can best explain the current findings. These findings may help us to develop educational curricula for bilingual individuals listening to and articulating English. Further, the extensive analysis of spontaneous speech in the present study should contribute to the specification of parameters for normal language development and function in Korean-English bilingual children and adults.
Resumo:
This paper considers the high-rate performance of source coding for noisy discrete symmetric channels with random index assignment (IA). Accurate analytical models are developed to characterize the expected distortion performance of vector quantization (VQ) for a large class of distortion measures. It is shown that when the point density is continuous, the distortion can be approximated as the sum of the source quantization distortion and the channel-error induced distortion. Expressions are also derived for the continuous point density that minimizes the expected distortion. Next, for the case of mean squared error distortion, a more accurate analytical model for the distortion is derived by allowing the point density to have a singular component. The extent of the singularity is also characterized. These results provide analytical models for the expected distortion performance of both conventional VQ as well as for channel-optimized VQ. As a practical example, compression of the linear predictive coding parameters in the wideband speech spectrum is considered, with the log spectral distortion as performance metric. The theory is able to correctly predict the channel error rate that is permissible for operation at a particular level of distortion.