2 resultados para acoustic speech recognition system
em Glasgow Theses Service
Resumo:
This thesis describes two separate projects. The first is a theoretical and experimental investigation of surface acoustic wave streaming in microfluidics. The second is the development of a novel acoustic glucose sensor. A separate abstract is given for each here. Optimization of acoustic streaming in microfluidic channels by SAWs Surface Acoustic Waves, (SAWs) actuated on flat piezoelectric substrates constitute a convenient and versatile tool for microfluidic manipulation due to the easy and versatile interfacing with microfluidic droplets and channels. The acoustic streaming effect can be exploited to drive fast streaming and pumping of fluids in microchannels and droplets (Shilton et al. 2014; Schmid et al. 2011), as well as size dependant sorting of particles in centrifugal flows and vortices (Franke et al. 2009; Rogers et al. 2010). Although the theory describing acoustic streaming by SAWs is well understood, very little attention has been paid to the optimisation of SAW streaming by the correct selection of frequency. In this thesis a finite element simulation of the fluid streaming in a microfluidic chamber due to a SAW beam was constructed and verified against micro-PIV measurements of the fluid flow in a fabricated device. It was found that there is an optimum frequency that generates the fastest streaming dependent on the height and width of the chamber. It is hoped this will serve as a design tool for those who want to optimally match SAW frequency with a particular microfluidic design. An acoustic glucose sensor Diabetes mellitus is a disease characterised by an inability to properly regulate blood glucose levels. In order to keep glucose levels under control some diabetics require regular injections of insulin. Continuous monitoring of glucose has been demonstrated to improve the management of diabetes (Zick et al. 2007; Heinemann & DeVries 2014), however there is a low patient uptake of continuous glucose monitoring systems due to the invasive nature of the current technology (Ramchandani et al. 2011). In this thesis a novel way of monitoring glucose levels is proposed which would use ultrasonic waves to ‘read’ a subcutaneous glucose sensitive-implant, which is only minimally invasive. The implant is an acoustic analogy of a Bragg stack with a ‘defect’ layer that acts as the sensing layer. A numerical study was performed on how the physical changes in the sensing layer can be deduced by monitoring the reflection amplitude spectrum of ultrasonic waves reflected from the implant. Coupled modes between the skin and the sensing layer were found to be a potential source of error and drift in the measurement. It was found that by increasing the number of layers in the stack that this could be minimized. A laboratory proof of concept system was developed using a glucose sensitive hydrogel as the sensing layer. It was possible to monitor the changing thickness and speed of sound of the hydrogel due to physiological relevant changes in glucose concentration.
Resumo:
The production and perception of music is a multimodal activity involving auditory, visual and conceptual processing, integrating these with prior knowledge and environmental experience. Musicians utilise expressive physical nuances to highlight salient features of the score. The question arises within the literature as to whether performers’ non-technical, non-sound-producing movements may be communicatively meaningful and convey important structural information to audience members and co-performers. In the light of previous performance research (Vines et al., 2006, Wanderley, 2002, Davidson, 1993), and considering findings within co-speech gestural research and auditory and audio-visual neuroscience, this thesis examines the nature of those movements not directly necessary for the production of sound, and their particular influence on audience perception. Within the current research 3D performance analysis is conducted using the Vicon 12- camera system and Nexus data-processing software. Performance gestures are identified as repeated patterns of motion relating to music structure, which not only express phrasing and structural hierarchy but are consistently and accurately interpreted as such by a perceiving audience. Gestural characteristics are analysed across performers and performance style using two Chopin preludes selected for their diverse yet comparable structures (Opus 28:7 and 6). Effects on perceptual judgements of presentation modes (visual-only, auditory-only, audiovisual, full- and point-light) and viewing conditions are explored. This thesis argues that while performance style is highly idiosyncratic, piano performers reliably generate structural gestures through repeated patterns of upper-body movement. The shapes and locations of phrasing motions are identified particular to the sample of performers investigated. Findings demonstrate that despite the personalised nature of the gestures, performers use increased velocity of movements to emphasise musical structure and that observers accurately and consistently locate phrasing junctures where these patterns and variation in motion magnitude, shape and velocity occur. By viewing performance motions in polar (spherical) rather than cartesian coordinate space it is possible to get mathematically closer to the movement generated by each of the nine performers, revealing distinct patterns of motion relating to phrasing structures, regardless of intended performance style. These patterns are highly individualised both to each performer and performed piece. Instantaneous velocity analysis indicates a right-directed bias of performance motion variation at salient structural features within individual performances. Perceptual analyses demonstrate that audience members are able to accurately and effectively detect phrasing structure from performance motion alone. This ability persists even for degraded point-light performances, where all extraneous environmental information has been removed. The relative contributions of audio, visual and audiovisual judgements demonstrate that the visual component of a performance does positively impact on the over- all accuracy of phrasing judgements, indicating that receivers are most effective in their recognition of structural segmentations when they can both see and hear a performance. Observers appear to make use of a rapid online judgement heuristics, adjusting response processes quickly to adapt and perform accurately across multiple modes of presentation and performance style. In line with existent theories within the literature, it is proposed that this processing ability may be related to cognitive and perceptual interpretation of syntax within gestural communication during social interaction and speech. Findings of this research may have future impact on performance pedagogy, computational analysis and performance research, as well as potentially influencing future investigations of the cognitive aspects of musical and gestural understanding.