6 resultados para Speech in Noise
Resumo:
We wished to replicate evidence that an experimental paradigm of speech illusions is associated with psychotic experiences. Fifty-four patients with a first episode of psychosis (FEP) and 150 healthy subjects were examined in an experimental paradigm assessing the presence of speech illusion in neutral white noise. Socio-demographic, cognitive function and family history data were collected. The Positive and Negative Syndrome Scale (PANSS) was administered in the patient group and the Structured Interview for Schizotypy-Revised (SIS-R), and the Community Assessment of Psychic Experiences (CAPE) in the control group. Patients had a much higher rate of speech illusions (33.3% versus 8.7%, ORadjusted: 5.1, 95% CI: 2.3-11.5), which was only partly explained by differences in IQ (ORadjusted: 3.4, 95% CI: 1.4-8.3). Differences were particularly marked for signals in random noise that were perceived as affectively salient (ORadjusted: 9.7, 95% CI: 1.8-53.9). Speech illusion tended to be associated with positive symptoms in patients (ORadjusted: 3.3, 95% CI: 0.9-11.6), particularly affectively salient illusions (ORadjusted: 8.3, 95% CI: 0.7-100.3). In controls, speech illusions were not associated with positive schizotypy (ORadjusted: 1.1, 95% CI: 0.3-3.4) or self-reported psychotic experiences (ORadjusted: 1.4, 95% CI: 0.4-4.6). Experimental paradigms indexing the tendency to detect affectively salient signals in noise may be used to identify liability to psychosis.
Resumo:
Query-by-Example Spoken Term Detection (QbE STD) aims at retrieving data from a speech data repository given an acoustic query containing the term of interest as input. Nowadays, it has been receiving much interest due to the high volume of information stored in audio or audiovisual format. QbE STD differs from automatic speech recognition (ASR) and keyword spotting (KWS)/spoken term detection (STD) since ASR is interested in all the terms/words that appear in the speech signal and KWS/STD relies on a textual transcription of the search term to retrieve the speech data. This paper presents the systems submitted to the ALBAYZIN 2012 QbE STD evaluation held as a part of ALBAYZIN 2012 evaluation campaign within the context of the IberSPEECH 2012 Conference(a). The evaluation consists of retrieving the speech files that contain the input queries, indicating their start and end timestamps within the appropriate speech file. Evaluation is conducted on a Spanish spontaneous speech database containing a set of talks from MAVIR workshops(b), which amount at about 7 h of speech in total. We present the database metric systems submitted along with all results and some discussion. Four different research groups took part in the evaluation. Evaluation results show the difficulty of this task and the limited performance indicates there is still a lot of room for improvement. The best result is achieved by a dynamic time warping-based search over Gaussian posteriorgrams/posterior phoneme probabilities. This paper also compares the systems aiming at establishing the best technique dealing with that difficult task and looking for defining promising directions for this relatively novel task.
Resumo:
[EN] This paper examines how the female characters in Greek novels have recourse to false speech. Based on an analysis of female speech in Attic tragedy, which was one of the literary genres that exerted the greatest influence on speech parts of the novels, a study is conducted to find out which characters in the novel employ false speech and their purpose in doing so. Two types of false speech were identified: the defensive one, used by the female protagonists or by secondary characters of similar social and ideological status, and the offensive one, used by characters of lower rank, and blameworthy morality within the ideological love's frameword publicized through the novel.
Resumo:
[ES]El rápido aumento de las nuevas tecnologías, así como el aumento del uso tanto del transporte público como del privado ha conllevado un aumento en los niveles de ruido. Además, la proximidad de las industrias a las ciudades y los trabajos diarios a pie de calle no han mejorado esta situación, sino que la ha empeorado. Debido a este aumento de la intensidad sonora, las enfermedades que derivan de ello han aumentado. Por ello, las soluciones para disminuir los niveles de ruido que llegan a la población se han ido desarrollando cada vez más. En el caso de las viviendas cercanas a los aeropuertos, la solución adoptada comúnmente es la insonorización acústica del edificio por su relativa sencillez. En este trabajo, se estudia un caso concreto de un edificio residencial cercano al aeropuerto de Loiu, Bilbao y se resuelve la opción más sencilla posible de aislamiento acústico para la vivienda.
Resumo:
Feature-based vocoders, e.g., STRAIGHT, offer a way to manipulate the perceived characteristics of the speech signal in speech transformation and synthesis. For the harmonic model, which provide excellent perceived quality, features for the amplitude parameters already exist (e.g., Line Spectral Frequencies (LSF), Mel-Frequency Cepstral Coefficients (MFCC)). However, because of the wrapping of the phase parameters, phase features are more difficult to design. To randomize the phase of the harmonic model during synthesis, a voicing feature is commonly used, which distinguishes voiced and unvoiced segments. However, voice production allows smooth transitions between voiced/unvoiced states which makes voicing segmentation sometimes tricky to estimate. In this article, two-phase features are suggested to represent the phase of the harmonic model in a uniform way, without voicing decision. The synthesis quality of the resulting vocoder has been evaluated, using subjective listening tests, in the context of resynthesis, pitch scaling, and Hidden Markov Model (HMM)-based synthesis. The experiments show that the suggested signal model is comparable to STRAIGHT or even better in some scenarios. They also reveal some limitations of the harmonic framework itself in the case of high fundamental frequencies.
Resumo:
Study of emotions in human-computer interaction is a growing research area. This paper shows an attempt to select the most significant features for emotion recognition in spoken Basque and Spanish Languages using different methods for feature selection. RekEmozio database was used as the experimental data set. Several Machine Learning paradigms were used for the emotion classification task. Experiments were executed in three phases, using different sets of features as classification variables in each phase. Moreover, feature subset selection was applied at each phase in order to seek for the most relevant feature subset. The three phases approach was selected to check the validity of the proposed approach. Achieved results show that an instance-based learning algorithm using feature subset selection techniques based on evolutionary algorithms is the best Machine Learning paradigm in automatic emotion recognition, with all different feature sets, obtaining a mean of 80,05% emotion recognition rate in Basque and a 74,82% in Spanish. In order to check the goodness of the proposed process, a greedy searching approach (FSS-Forward) has been applied and a comparison between them is provided. Based on achieved results, a set of most relevant non-speaker dependent features is proposed for both languages and new perspectives are suggested.