911 resultados para Consonance dissonance sounds
Resumo:
We compared magnetoencephalographic responses for natural vowels and for sounds consisting of two pure tones that represent the two lowest formant frequencies of these vowels. Our aim was to determine whether spectral changes in successive stimuli are detected differently for speech and nonspeech sounds. The stimuli were presented in four blocks applying an oddball paradigm (20% deviants, 80% standards): (i) /α/ tokens as deviants vs. /i/ tokens as standards; (ii) /e/ vs. /i/; (iii) complex tones representing /α/ formants vs. /i/ formants; and (iv) complex tones representing /e/ formants vs. /i/ formants. Mismatch fields (MMFs) were calculated by subtracting the source waveform produced by standards from that produced by deviants. As expected, MMF amplitudes for the complex tones reflected acoustic deviation: the amplitudes were stronger for the complex tones representing /α/ than /e/ formants, i.e., when the spectral difference between standards and deviants was larger. In contrast, MMF amplitudes for the vowels were similar despite their different spectral composition, whereas the MMF onset time was longer for /e/ than for /α/. Thus the degree of spectral difference between standards and deviants was reflected by the MMF amplitude for the nonspeech sounds and by the MMF latency for the vowels.
Resumo:
Neuronal models predict that retrieval of specific event information reactivates brain regions that were active during encoding of this information. Consistent with this prediction, this positron-emission tomography study showed that remembering that visual words had been paired with sounds at encoding activated some of the auditory brain regions that were engaged during encoding. After word-sound encoding, activation of auditory brain regions was also observed during visual word recognition when there was no demand to retrieve auditory information. Collectively, these observations suggest that information about the auditory components of multisensory event information is stored in auditory responsive cortex and reactivated at retrieval, in keeping with classical ideas about “redintegration,” that is, the power of part of an encoded stimulus complex to evoke the whole experience.
Resumo:
A fundamental question in human memory is how the brain represents sensory-specific information during the process of retrieval. One hypothesis is that regions of sensory cortex are reactivated during retrieval of sensory-specific information (1). Here we report findings from a study in which subjects learned a set of picture and sound items and were then given a recall test during which they vividly remembered the items while imaged by using event-related functional MRI. Regions of visual and auditory cortex were activated differentially during retrieval of pictures and sounds, respectively. Furthermore, the regions activated during the recall test comprised a subset of those activated during a separate perception task in which subjects actually viewed pictures and heard sounds. Regions activated during the recall test were found to be represented more in late than in early visual and auditory cortex. Therefore, results indicate that retrieval of vivid visual and auditory information can be associated with a reactivation of some of the same sensory regions that were activated during perception of those items.
Resumo:
Owls and other animals, including humans, use the difference in arrival time of sounds between the ears to determine the direction of a sound source in the horizontal plane. When an interaural time difference (ITD) is conveyed by a narrowband signal such as a tone, human beings may fail to derive the direction represented by that ITD. This is because they cannot distinguish the true ITD contained in the signal from its phase equivalents that are ITD ± nT, where T is the period of the stimulus tone and n is an integer. This uncertainty is called phase-ambiguity. All ITD-sensitive neurons in birds and mammals respond to an ITD and its phase equivalents when the ITD is contained in narrowband signals. It is not known, however, if these animals show phase-ambiguity in the localization of narrowband signals. The present work shows that barn owls (Tyto alba) experience phase-ambiguity in the localization of tones delivered by earphones. We used sound-induced head-turning responses to measure the sound-source directions perceived by two owls. In both owls, head-turning angles varied as a sinusoidal function of ITD. One owl always pointed to the direction represented by the smaller of the two ITDs, whereas a second owl always chose the direction represented by the larger ITD (i.e., ITD − T).
Resumo:
The barn owl (Tyto alba) uses interaural time difference (ITD) cues to localize sounds in the horizontal plane. Low-order binaural auditory neurons with sharp frequency tuning act as narrow-band coincidence detectors; such neurons respond equally well to sounds with a particular ITD and its phase equivalents and are said to be phase ambiguous. Higher-order neurons with broad frequency tuning are unambiguously selective for single ITDs in response to broad-band sounds and show little or no response to phase equivalents. Selectivity for single ITDs is thought to arise from the convergence of parallel, narrow-band frequency channels that originate in the cochlea. ITD tuning to variable bandwidth stimuli was measured in higher-order neurons of the owl’s inferior colliculus to examine the rules that govern the relationship between frequency channel convergence and the resolution of phase ambiguity. Ambiguity decreased as stimulus bandwidth increased, reaching a minimum at 2–3 kHz. Two independent mechanisms appear to contribute to the elimination of ambiguity: one suppressive and one facilitative. The integration of information carried by parallel, distributed processing channels is a common theme of sensory processing that spans both modality and species boundaries. The principles underlying the resolution of phase ambiguity and frequency channel convergence in the owl may have implications for other sensory systems, such as electrolocation in electric fish and the computation of binocular disparity in the avian and mammalian visual systems.
Resumo:
Two and a half millennia ago Pythagoras initiated the scientific study of the pitch of sounds; yet our understanding of the mechanisms of pitch perception remains incomplete. Physical models of pitch perception try to explain from elementary principles why certain physical characteristics of the stimulus lead to particular pitch sensations. There are two broad categories of pitch-perception models: place or spectral models consider that pitch is mainly related to the Fourier spectrum of the stimulus, whereas for periodicity or temporal models its characteristics in the time domain are more important. Current models from either class are usually computationally intensive, implementing a series of steps more or less supported by auditory physiology. However, the brain has to analyze and react in real time to an enormous amount of information from the ear and other senses. How is all this information efficiently represented and processed in the nervous system? A proposal of nonlinear and complex systems research is that dynamical attractors may form the basis of neural information processing. Because the auditory system is a complex and highly nonlinear dynamical system, it is natural to suppose that dynamical attractors may carry perceptual and functional meaning. Here we show that this idea, scarcely developed in current pitch models, can be successfully applied to pitch perception.
Resumo:
Hearing underlies our ability to locate sound sources in the environment, our appreciation of music, and our ability to communicate. Participants in the National Academy of Sciences colloquium on Auditory Neuroscience: Development, Transduction, and Integration presented research results bearing on four key issues in auditory research. How does the complex inner ear develop? How does the cochlea transduce sounds into electrical signals? How does the brain's ability to compute the location of a sound source develop? How does the forebrain analyze complex sounds, particularly species-specific communications? This article provides an introduction to the papers stemming from the meeting.
Resumo:
The anatomical and biophysical specializations of octopus cells allow them to detect the coincident firing of groups of auditory nerve fibers and to convey the precise timing of that coincidence to their targets. Octopus cells occupy a sharply defined region of the most caudal and dorsal part of the mammalian ventral cochlear nucleus. The dendrites of octopus cells cross the bundle of auditory nerve fibers just proximal to where the fibers leave the ventral and enter the dorsal cochlear nucleus, each octopus cell spanning about one-third of the tonotopic array. Octopus cells are excited by auditory nerve fibers through the activation of rapid, calcium-permeable, α-amino-3-hydroxy-5-methyl-4-isoxazole-propionate receptors. Synaptic responses are shaped by the unusual biophysical characteristics of octopus cells. Octopus cells have very low input resistances (about 7 MΩ), and short time constants (about 200 μsec) as a consequence of the activation at rest of a hyperpolarization-activated mixed-cation conductance and a low-threshold, depolarization-activated potassium conductance. The low input resistance causes rapid synaptic currents to generate rapid and small synaptic potentials. Summation of small synaptic potentials from many fibers is required to bring an octopus cell to threshold. Not only does the low input resistance make individual excitatory postsynaptic potentials brief so that they must be generated within 1 msec to sum but also the voltage-sensitive conductances of octopus cells prevent firing if the activation of auditory nerve inputs is not sufficiently synchronous and depolarization is not sufficiently rapid. In vivo in cats, octopus cells can fire rapidly and respond with exceptionally well-timed action potentials to periodic, broadband sounds such as clicks. Thus both the anatomical specializations and the biophysical specializations make octopus cells detectors of the coincident firing of their auditory nerve fiber inputs.
Resumo:
The functional specialization and hierarchical organization of multiple areas in rhesus monkey auditory cortex were examined with various types of complex sounds. Neurons in the lateral belt areas of the superior temporal gyrus were tuned to the best center frequency and bandwidth of band-passed noise bursts. They were also selective for the rate and direction of linear frequency modulated sweeps. Many neurons showed a preference for a limited number of species-specific vocalizations (“monkey calls”). These response selectivities can be explained by nonlinear spectral and temporal integration mechanisms. In a separate series of experiments, monkey calls were presented at different spatial locations, and the tuning of lateral belt neurons to monkey calls and spatial location was determined. Of the three belt areas the anterolateral area shows the highest degree of specificity for monkey calls, whereas neurons in the caudolateral area display the greatest spatial selectivity. We conclude that the cortical auditory system of primates is divided into at least two processing streams, a spatial stream that originates in the caudal part of the superior temporal gyrus and projects to the parietal cortex, and a pattern or object stream originating in the more anterior portions of the lateral belt. A similar division of labor can be seen in human auditory cortex by using functional neuroimaging.
Resumo:
Sound localization relies on the neural processing of monaural and binaural spatial cues that arise from the way sounds interact with the head and external ears. Neurophysiological studies of animals raised with abnormal sensory inputs show that the map of auditory space in the superior colliculus is shaped during development by both auditory and visual experience. An example of this plasticity is provided by monaural occlusion during infancy, which leads to compensatory changes in auditory spatial tuning that tend to preserve the alignment between the neural representations of visual and auditory space. Adaptive changes also take place in sound localization behavior, as demonstrated by the fact that ferrets raised and tested with one ear plugged learn to localize as accurately as control animals. In both cases, these adjustments may involve greater use of monaural spectral cues provided by the other ear. Although plasticity in the auditory space map seems to be restricted to development, adult ferrets show some recovery of sound localization behavior after long-term monaural occlusion. The capacity for behavioral adaptation is, however, task dependent, because auditory spatial acuity and binaural unmasking (a measure of the spatial contribution to the “cocktail party effect”) are permanently impaired by chronically plugging one ear, both in infancy but especially in adulthood. Experience-induced plasticity allows the neural circuitry underlying sound localization to be customized to individual characteristics, such as the size and shape of the head and ears, and to compensate for natural conductive hearing losses, including those associated with middle ear disease in infancy.
Resumo:
In the past decade, tremendous advances in the state of the art of automatic speech recognition by machine have taken place. A reduction in the word error rate by more than a factor of 5 and an increase in recognition speeds by several orders of magnitude (brought about by a combination of faster recognition search algorithms and more powerful computers), have combined to make high-accuracy, speaker-independent, continuous speech recognition for large vocabularies possible in real time, on off-the-shelf workstations, without the aid of special hardware. These advances promise to make speech recognition technology readily available to the general public. This paper focuses on the speech recognition advances made through better speech modeling techniques, chiefly through more accurate mathematical modeling of speech sounds.
Resumo:
INTRODUÇÃO: Durante procedimentos endovasculares, estão presentes os riscos relacionados ao uso dos contrastes iodados, tais como a nefropatia por contraste, uma vez que é fundamental o emprego de um meio de contraste para a obtenção das imagens radiológicas vasculares. A injeção intravascular de gás CO2 purificado é reconhecidamente uma alternativa relativamente mais segura ao contraste iodado, contudo, seu manuseio artesanal pode também trazer dificuldades técnicas e riscos aos pacientes. Para contemplar estas questões, foi desenvolvido o protótipo de um injetor intravascular de CO2 medicinal, microprocessado, dedicado à obtenção de imagens angiográficas. OBJETIVOS: Realizar os testes de viabilidade técnica inicial do protótipo em modelo in vivo. MÉTODOS: Realizar a angioplastia da artéria renal esquerda de 10 porcos, divididos em 2 grupos: Grupo 1 (n=5) injeção de contraste iodado, Grupo 2 (n=5) injeção de CO2 através do protótipo. Monitorização clínica e laboratorial dos animais no pré, intra e pós-operatório, com exames de função renal na véspera e 48h após os procedimentos e 3 gasometrias arteriais seriadas no intra-operatório. Observação clínica foi mantida por 48h no pós- operatório. RESULTADOS: Os procedimentos de angioplastia com CO2 foram realizados com sucesso técnico de 100%, sem necessidade de complementação com injeção de contraste iodado no Grupo 2. Não foram identificadas falhas no protótipo em funcionamento. Não foram identificadas alterações clínicas ou radiológicas sugestivas de contaminação por ar ambiente do sistema de CO2 e nem alterações laboratoriais nos animais. As imagens angiográficas obtidas no Grupo 2 foram consideradas, numa avaliação subjetiva, relativamente inferiores às imagens obtidas no Grupo 1. DISCUSSÃO: A qualidade inferior de imagem no Grupo 2 pode ser atribuída ao equipamento de fluoroscopia utilizado, com software desatualizado em relação aos equipamentos atuais, que incluem pré-configurações para angiografia com CO2; no entanto, ainda assim todos os procedimentos propostos no Grupo 2 foram realizados com sucesso técnico, o que nos leva a classificar as imagens deste grupo 2 como satisfatórias. O manuseio do aparelho mostrou-se ágil e eficiente, com a programação dos parâmetros sendo realizada com facilidade através do visor \"touch screen\", comparativamente superior ao método artesanal de injeção de CO2 com seringas em selo d\'água. CONCLUSÕES: O protótipo do aparelho injetor intravascular de CO2 funcionou de forma adequada durante os testes e as imagens obtidas permitiram a compleição com sucesso dos procedimentos. Portanto, os resultados positivos obtidos sugerem que o equipamento é tecnicamente viável
Resumo:
INTRODUÇÃO: Os marcadores clínicos de desenvolvimento possibilitam aos profissionais se familiarizarem com a sequência do desenvolvimento das habilidades auditivas e de linguagem e sinalizarem para a família quando há algum padrão desviante do esperado para o desenvolvimento da criança. O objetivo da presente pesquisa foi determinar os marcadores clínicos de desenvolvimento das habilidades auditivas e de linguagem falada, a partir da análise dos primeiros cinco anos de uso do IC de crianças implantadas antes dos 36 meses; e investigar a influência da idade de implantação no desenvolvimento das habilidades citadas. MÉTODOS: Estudo longitudinal retrospectivo realizado na Seção de Implante Coclear - Centro de Pesquisas Audiológicas (CPA-HRAC/USP). Fizeram parte da amostra 230 crianças que, para análise comparativa, foram dividas em três grupos: operadas e ativadas antes dos 18 meses, entre 19 e 24 meses e entre 25 e 36 meses de idade. Os procedimentos analisados foram: a Infant-Toddler: Meaningful Auditory Integration Scale (IT-MAIS), a Meaningful Use of Speech Scale (MUSS) e as Categorias de Audição e de Linguagem. Os dados coletados foram analisados por meio das estatísticas descritiva e indutiva. RESULTADOS: Durante os primeiros cinco anos de uso do IC foram analisados nove retornos das crianças ao Centro. A partir da análise da mediana, até os 30 ± 3 meses de uso do dispositivo eletrônico grande parte da amostra atingiu 100% na IT-MAIS, quando as habilidades de atenção e de atribuição dos significados aos sons já estavam superadas. Até os 68 ± 6 meses a maioria das crianças alcançou a porcentagem máxima na MUSS e a pontuação máxima nas Categorias de Audição e de Linguagem, ou seja, as crianças já utilizavam a fala espontânea e as estratégias de comunicação em sua rotina, bem como apresentavam as habilidades de reconhecimento auditivo em conjunto aberto e a fluência da linguagem oral, respectivamente. Quando comparados os desempenhos dos grupos, nas avaliações auditivas não houve um padrão de significância estatística e nas avaliações da linguagem os resultados foram significativamente melhores para as crianças implantadas após os 18 meses nos primeiros retornos. Houve fortes correlações entre os resultados das Escalas e Categorias. CONCLUSÕES: As crianças da amostra desenvolveram progressivamente as habilidades auditivas e de linguagem falada ao longo dos primeiros cinco anos de uso do IC. Foi possível determinar os marcadores clínicos de desenvolvimento para as Escalas e Categorias estudadas. A partir deles os profissionais que acompanham a criança no processo de habilitação auditiva, poderão nortear a família, bem como os demais profissionais que atuam com a criança, quanto aos resultados esperados na IT-MAIS, na MUSS e nas Categorias de Audição e Linguagem. Também, foi possível identificar que, mesmo havendo uma restrição quanto as possíveis variáveis que podem interferir na determinação dos marcadores clínicos, houve pacientes com resultados desviantes, sugerindo a importância da definição dos marcadores para, juntamente com a família, o profissional discutir e encontrar outras variáveis que possam influenciar no baixo desempenho da criança. A implantação dentro do período sensível do desenvolvimento pode explicar comportamento auditivo dos grupos quando comparados. Já, quando analisada a linguagem falada, acredita-se que houve a influência de outras variáveis no processo de habilitação auditiva e não apenas a implantação durante o período crítico
Resumo:
This dissertation project explored professionalism and the performance of identities by examining Taiwanese commercial airline pilots' discursive practices in everyday life. The intentions for this project were to not only expand current knowledge of organizational communication from a critical rhetorical perspective, but to further explore the under-appreciated concept of professionalism of organizational members. Theoretically, I traced theoretical analysis in the sociology of professions and further investigated scholarship from identity research in organizational communication studies. This research agenda helped to advance communication-based understandings of the meanings and practices of professional identity as a complement to the sociological conception. I further merged a performance paradigm and critical rhetorical perspective to examine the discursive practices of organizational members and to challenge the bias of traditional textual approaches. Methodologically, I conducted ethnographic interviews with Taiwanese commercial airline pilots in order to understand how they construct their personal, social, and professional identities. Five narrative themes were identified and demonstrated in this project: (1) It takes a lot to become a commercial airline pilot, (2) Being a professional commercial airline pilot is to build up sufficient knowledge, beyond average skill, and correct attitude, (3) Pilots' resistance and dissent toward company management, (4) Popular (re)presentation influences professionalism, (5) Power and fear affect professionalism. Pilots' personal narratives were presented in performative writing and in poetic transcription to make word alive with sounds featuring their meanings. Their personal storytelling created a dialogic space to not only let pilots' voice to be heard but also revealed how identities are created within and against a larger organizational identity. Overall, this project demonstrated the interdisciplinary examination of the meanings, functions, and consequence of discursive practices in everyday professional life. It also critiqued relationships between power, domination, and resistance while reintroducing the roles of the body and materiality in the domain of professionalism, and provides ethical readings of larger and complex organizational cultures. Applying communication-oriented analysis to study professionalism indeed challenged the long time neglected phenomena regarding the power of the symbolic in sociological approaches and raised the awareness of structural, material, and bodily condition of work.
Resumo:
Objetivos: estabelecer amostras de referência constituídas por gravações julgadas com consenso como representativas da presença ou ausência da oclusiva glotal (OG) e comparar julgamentos perceptivo-auditivos da presença e ausência da OG com e sem o uso de amostras de referência. Metodologia: o estudo foi dividido em duas etapas. Durante a ETAPA 1, 480 frases referentes aos sons oclusivos e fricativos produzidas por falantes com história de fissura labiopalatina foram julgadas por três fonoaudiólogas experientes quanto à identificação da OG. As frases foram julgadas individualmente e aquelas que não apresentaram consenso inicial foram julgadas novamente de maneira simultânea. As amostras julgadas com consenso com relação à presença ou ausência da OG durante produção das seis consoantes-alvo oclusivas e seis fricativas foram selecionadas para estabelecer um Banco de Amostras Representativas da OG. A ETAPA 2 consistiu na seleção de 48 amostras de referência referentes aos 12 sons de interesse e 120 amostras experimentais e, o julgamento dessas amostras experimentais por três grupos de juízes, cada grupo com três juízes com experiências distintas com relação ao julgamento de fala na fissura de palato. Os juízes julgaram as amostras experimentais duas vezes, primeiro sem acesso às referências e, após uma semana, com acesso às referências. Resultados: os julgamentos realizados na ETAPA 1 evidenciaram consenso com relação a OG em 352 amostras, sendo 120 frases com produção adequada para os sons de interesse e 232 representativas do uso da OG. Essas 352 amostras constituíram o Banco de amostras Representativas da OG. Os resultados da ETAPA 2 indicaram que ao comparar a média do valor de Kappa obtida para os 12 sons de interesse em cada um dos grupos nos julgamentos sem e com acesso às amostras de referência a concordância para o grupo 1 (G1) passou de regular (K=0,35) para moderada (K=0,55), para o grupo 2 (G2) passou de moderada (K=0,44) para substancial (K=0,76) e para o grupo 3 (G3) passou de substancial (K=0,72) para quase perfeita (K=0,83). Observou-se que as melhores concordâncias ocorreram para o grupo dos fonoaudiólogos experientes (G3), seguido dos fonoaudiólogos recém-formados (G2), com as piores observadas para o grupo de alunos de graduação (G1). Conclusão: um Banco de Amostras de Referência Representativas da OG foi estabelecido e os julgamentos perceptivo-auditivos de juízes com uso das amostras de referência foram obtidos com concordância inter-juízes e porcentagem de acertos melhor do que os julgamentos sem acesso às referências. Os resultados sugerem a importância do uso de amostras de referência para minimizar a subjetividade da avaliação perceptivo auditiva da fala.