917 resultados para Auxiliares de Audição
Resumo:
Intelligent Transport Systems (ITS) have the potential to substantially reduce the number of crashes caused by human errors at railway levels crossings. However, such systems could overwhelm drivers, generate different types of driver errors and have negative effects on safety at level crossing. The literature shows an increasing interest for new ITS for increasing driver situational awareness at level crossings, as well as evaluations of such new systems on compliance. To our knowledge, the potential negative effects of such technologies have not been comprehensively evaluated yet. This study aimed at assessing the effect of different ITS interventions, designed to enhance driver behaviour at railway crossings, on driver’s cognitive loads. Fifty eight participants took part in a driving simulator study in which three ITS devices were tested: an in-vehicle visual ITS, an in-vehicle audio ITS, and an on-road valet system. Driver cognitive load was objectively and subjectively assessed for each ITS intervention. Objective data were collected from a heart rate monitor and an eye tracker, while subjective data was collected with the NASA-TLX questionnaire. Overall, results indicated that the three trialled technologies did not result in significant changes in cognitive load while approaching crossings.
Resumo:
Background Haemodialysis (HD) nursing is characterised by frequent, intense interactions with patients over long periods of time resulting in a unique nurse-patient relationship. Due to the life-limiting nature of end-stage renal failure, nurses are likely to have repeated exposures to the death of patients with whom they have formed relationships. Repeated exposure to patient death translates into frequent grief experiences. There is scant literature on the psychological impact of patient death for nurses working in the HD setting. Aims To explore HD nurses experiences of patient death and coping mechanisms used. Methods A sequential mixed method study investigating job satisfaction, stress and burnout found that HD nurses had high levels of stress and burnout. These results were explored in more detail during 8 semi-structured interviews with HD nurses. Interviews were audio-recorded, transcribed verbatim and subjected to thematic analysis. Results Three themes were identified that highlight the stress experienced by nurses when a haemodialysis patient dies. The first theme, “quazi-family” describes the close relationship which forms between nurses and patients. The “complicated grief” theme outlines the impact of death on HD nurses, and the final theme, “remembrance” explains some of the coping mechanisms used in the grieving process. Conclusion Nurses develop individual coping mechanisms to accommodate the grief and loss experienced when a “close” patient dies. The grieving process caused by the death of patient’s needs to be recognised by nurses and nurse managers as causing psychological stress and strain.
Resumo:
Spoken term detection (STD) is the task of looking up a spoken term in a large volume of speech segments. In order to provide fast search, speech segments are first indexed into an intermediate representation using speech recognition engines which provide multiple hypotheses for each speech segment. Approximate matching techniques are usually applied at the search stage to compensate the poor performance of automatic speech recognition engines during indexing. Recently, using visual information in addition to audio information has been shown to improve phone recognition performance, particularly in noisy environments. In this paper, we will make use of visual information in the form of lip movements of the speaker in indexing stage and will investigate its effect on STD performance. Particularly, we will investigate if gains in phone recognition accuracy will carry through the approximate matching stage to provide similar gains in the final audio-visual STD system over a traditional audio only approach. We will also investigate the effect of using visual information on STD performance in different noise environments.
Resumo:
Bioacoustic monitoring has become a significant research topic for species diversity conservation. Due to the development of sensing techniques, acoustic sensors are widely deployed in the field to record animal sounds over a large spatial and temporal scale. With large volumes of collected audio data, it is essential to develop semi-automatic or automatic techniques to analyse the data. This can help ecologists make decisions on how to protect and promote the species diversity. This paper presents generic features to characterize a range of bird species for vocalisation retrieval. In the implementation, audio recordings are first converted to spectrograms using short-time Fourier transform, then a ridge detection method is applied to the spectrogram for detecting points of interest. Based on the detected points, a new region representation are explored for describing various bird vocalisations and a local descriptor including temporal entropy, frequency bin entropy and histogram of counts of four ridge directions is calculated for each sub-region. To speed up the retrieval process, indexing is carried out and the retrieved results are ranked according to similarity scores. The experiment results show that our proposed feature set can achieve 0.71 in term of retrieval success rate which outperforms spectral ridge features alone (0.55) and Mel frequency cepstral coefficients (0.36).
Resumo:
Acoustic recordings play an increasingly important role in monitoring terrestrial environments. However, due to rapid advances in technology, ecologists are accumulating more audio than they can listen to. Our approach to this big-data challenge is to visualize the content of long-duration audio recordings by calculating acoustic indices. These are statistics which describe the temporal-spectral distribution of acoustic energy and reflect content of ecological interest. We combine spectral indices to produce false-color spectrogram images. These not only reveal acoustic content but also facilitate navigation. An additional analytic challenge is to find appropriate descriptors to summarize the content of 24-hour recordings, so that it becomes possible to monitor long-term changes in the acoustic environment at a single location and to compare the acoustic environments of different locations. We describe a 24-hour ‘acoustic-fingerprint’ which shows some preliminary promise.
Resumo:
Unified communications as a service (UCaaS) can be regarded as a cost-effective model for on-demand delivery of unified communications services in the cloud. However, addressing security concerns has been seen as the biggest challenge to the adoption of IT services in the cloud. This study set up a cloud system via VMware suite to emulate hosting unified communications (UC), the integration of two or more real time communication systems, services in the cloud in a laboratory environment. An Internet Protocol Security (IPSec) gateway was also set up to support network-level security for UCaaS against possible security exposures. This study was aimed at analysis of an implementation of UCaaS over IPSec and evaluation of the latency of encrypted UC traffic while protecting that traffic. Our test results show no latency while IPSec is implemented with a G.711 audio codec. However, the performance of the G.722 audio codec with an IPSec implementation affects the overall performance of the UC server. These results give technical advice and guidance to those involved in security controls in UC security on premises as well as in the cloud.
Resumo:
Engaging middle-school students in science continues to be a challenge in Australian schools. One initiative that has been tried in the senior years but is a more recent development in the middle years is the context-based approach. In this ethnographic study, we researched the teaching and learning transactions that occurred in one 9th grade science class studying a context-based Environmental Science unit that included visits to the local creek for 11 weeks. Data were derived from field notes, audio and video recorded conversations, interviews, student journals and classroom documents with a particular focus on two selected groups of students. This paper presents two assertions that highlight pedagogical approaches that contributed to learning. Firstly, spontaneous teaching episodes created opportunities for in-the-moment questioning by the teacher that led to students’ awareness of environmental issues and the scientific method; secondly, group work using flip cameras afforded opportunities for students to connect the science concepts with the context. Furthermore, students reported positively about the unit and expressed their appreciation for the opportunity to visit the creek frequently. This findings from this study should encourage teachers to take students into the real-world field for valuable teaching and learning experiences that are not available in the formal classroom.
Resumo:
These are turbulent times for audio- visual production companies. Radical changes, both inside and outside the organizations, reach across national markets and different genres. For instance, production methods are changing; the demand from audiences and advertisers is changing; power relations between the actors involved in the value chain are changing; and increasing concentration makes the market even more competitive for small independent players. From a perspective of the structure–conduct– performance paradigm (Ramstad, 1997) it is reasonable to expect that these changes on a structural level of the industry will cause the production companies to adapt their strategic behaviour. The current challenges for media companies are a combination of rising complexity and uncertainty in the market (Picard, 2004). The increasing complexity can for instance be observed in the growing number of market segments and in the continuing trend towards cross- media strategies where media companies operate in multiple markets and on multiple platforms...
Resumo:
Avian species richness surveys, which measure the total number of unique avian species, can be conducted via remote acoustic sensors. An immense quantity of data can be collected, which, although rich in useful information, places a great workload on the scientists who manually inspect the audio. To deal with this big data problem, we calculated acoustic indices from audio data at a one-minute resolution and used them to classify one-minute recordings into five classes. By filtering out the non-avian minutes, we can reduce the amount of data by about 50% and improve the efficiency of determining avian species richness. The experimental results show that, given 60 one-minute samples, our approach enables to direct ecologists to find about 10% more avian species.
Resumo:
Frog species have been declining worldwide at unprecedented rates in the past decades. There are many reasons for this decline including pollution, habitat loss, and invasive species [1]. To preserve, protect, and restore frog biodiversity, it is important to monitor and assess frog species. In this paper, a novel method using image processing techniques for analyzing Australian frog vocalisations is proposed. An FFT is applied to audio data to produce a spectrogram. Then, acoustic events are detected and isolated into corresponding segments through image processing techniques applied to the spectrogram. For each segment, spectral peak tracks are extracted with selected seeds and a region growing technique is utilised to obtain the contour of each frog vocalisation. Based on spectral peak tracks and the contour of each frog vocalisation, six feature sets are extracted. Principal component analysis reduces each feature set down to six principal components which are tested for classification performance with a k-nearest neighbor classifier. This experiment tests the proposed method of classification on fourteen frog species which are geographically well distributed throughout Queensland, Australia. The experimental results show that the best average classification accuracy for the fourteen frog species can be up to 87%.
Resumo:
Acoustic classification of anurans (frogs) has received increasing attention for its promising application in biological and environment studies. In this study, a novel feature extraction method for frog call classification is presented based on the analysis of spectrograms. The frog calls are first automatically segmented into syllables. Then, spectral peak tracks are extracted to separate desired signal (frog calls) from background noise. The spectral peak tracks are used to extract various syllable features, including: syllable duration, dominant frequency, oscillation rate, frequency modulation, and energy modulation. Finally, a k-nearest neighbor classifier is used for classifying frog calls based on the results of principal component analysis. The experiment results show that syllable features can achieve an average classification accuracy of 90.5% which outperforms Mel-frequency cepstral coefficients features (79.0%).
Resumo:
Frogs have received increasing attention due to their effectiveness for indicating the environment change. Therefore, it is important to monitor and assess frogs. With the development of sensor techniques, large volumes of audio data (including frog calls) have been collected and need to be analysed. After transforming the audio data into its spectrogram representation using short-time Fourier transform, the visual inspection of this representation motivates us to use image processing techniques for analysing audio data. Applying acoustic event detection (AED) method to spectrograms, acoustic events are firstly detected from which ridges are extracted. Three feature sets, Mel-frequency cepstral coefficients (MFCCs), AED feature set and ridge feature set, are then used for frog call classification with a support vector machine classifier. Fifteen frog species widely spread in Queensland, Australia, are selected to evaluate the proposed method. The experimental results show that ridge feature set can achieve an average classification accuracy of 74.73% which outperforms the MFCCs (38.99%) and AED feature set (67.78%).
Resumo:
Monitoring gas purity is an important aspect of gas recovery stations where air is usually one of the major impurities. Purity monitors of Katherometric type ate commercially available for this purpose. Alternatively, we discuss here a helium gas purity monitor based on acoustic resonance of a cavity at audio frequencies. It measures the purity by monitoring the resonant frequency of a cylindrical cavity filled with the gas under test and excited by conventional telephone transducers fixed at the ends. The use of the latter simplifies the design considerably. The paper discusses the details of the resonant cavity and the electronic circuit along with temperature compensation. The unit has been calibrated with helium gas of known purities. The unit has a response time of the order of 10 minutes and measures the gas purity to an accuracy of 0.02%. The unit has been installed in our helium recovery system and is found to perform satisfactorily.
Resumo:
This research has made contributions to the area of spoken term detection (STD), defined as the process of finding all occurrences of a specified search term in a large collection of speech segments. The use of visual information in the form of lip movements of the speaker in addition to audio and the use of topic of the speech segments, and the expected frequency of words in the target speech domain, are proposed. By using these complementary information, improvement in the performance of STD has been achieved which enables efficient search of key words in large collection of multimedia documents.
Resumo:
Bioacoustic data can be used for monitoring animal species diversity. The deployment of acoustic sensors enables acoustic monitoring at large temporal and spatial scales. We describe a content-based birdcall retrieval algorithm for the exploration of large data bases of acoustic recordings. In the algorithm, an event-based searching scheme and compact features are developed. In detail, ridge events are detected from audio files using event detection on spectral ridges. Then event alignment is used to search through audio files to locate candidate instances. A similarity measure is then applied to dimension-reduced spectral ridge feature vectors. The event-based searching method processes a smaller list of instances for faster retrieval. The experimental results demonstrate that our features achieve better success rate than existing methods and the feature dimension is greatly reduced.