32 resultados para natural language processing


Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents methods for simulating room acoustics using the finite-difference time-domain (FDTD) technique, focusing on boundary and medium modeling. A family of nonstaggered 3-D compact explicit FDTD schemes is analyzed in terms of stability, accuracy, and computational efficiency, and the most accurate and isotropic schemes based on a rectilinear grid are identified. A frequency-dependent boundary model that is consistent with locally reacting surface theory is also presented, in which the wall impedance is represented with a digital filter. For boundaries, accuracy in numerical reflection is analyzed and a stability proof is provided. The results indicate that the proposed 3-D interpolated wideband and isotropic schemes outperform directly related techniques based on Yee's staggered grid and standard digital waveguide mesh, and that the boundary formulations generally have properties that are similar to that of the basic scheme used.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper, a method for modeling diffusive boundaries in finite difference time domain (FDTD) room acoustics simulations with the use of impedance filters is presented. The proposed technique is based on the concept of phase grating diffusers, and realized by designing boundary impedance filters from normal-incidence reflection filters with added delay. These added delays, that correspond to the diffuser well depths, are varied across the boundary surface, and implemented using Thiran allpass filters. The proposed method for simulating sound scattering is suitable for modeling high frequency diffusion caused by small variations in surface roughness and, more generally, diffusers characterized by narrow wells with infinitely thin separators. This concept is also applicable to other wave-based modeling techniques. The approach is validated by comparing numerical results for Schroeder diffusers to measured data. In addition, it is proposed that irregular surfaces are modeled by shaping them with Brownian noise, giving good control over the sound scattering properties of the simulated boundary through two parameters, namely the spectral density exponent and the maximum well depth.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Approximants that can be considered weaker versions of voiced fricatives (termed here ‘frictionless continuants’) are poorly served by the IPA in terms of symbolization as compared to semi-vowel approximants. In this paper we survey the central approximants and the symbols and diacritics used to transcribe them; we focus on evidence for the use of non-rhotic frictionless continuants in both natural language (by which we mean non-clinical varieties) and disordered speech; and we suggest some possible unitary symbols for those that currently require the use of a hard-to-read lowering diacritic beneath the symbol for the corresponding voiced fricative.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Purpose: The purpose of this paper is to engage a different notion of feminism in accounting by addressing the issues of feminism, balance, and integration as a means of understanding differently the world for which one accounts. The ideas are communicated by the sharing of experiences through myth and storytelling.

Design/methodology/approach: An alternative lens for understanding the giving of accounts is proposed, drawing on earlier feminist accounting literature as well as storytelling and myth.

Findings: Including the subjective and intersubjective approaches to experiencing and understanding the world recommends an approach whereby both the feminine-intuitive and the masculine-rational processes are integrated in constructing decision models and accounts.

Research limitations/implications: Through an expanded view of values that can be included in reporting or recounting a different model is seen, and different decisions are enabled. The primary limitation is having to use words to convey one’s subjective and intersubjective understandings. The written medium is not the most natural language for such an undertaking.

Practical implications: By enabling the inclusion of more feminine values, a way is opened to engage more holistically with the society in which decisions are embedded.

Originality/value: Drawing on the storytelling tradition, a holistic model is suggested that can lead to emergence of a more balanced societal reporting.

Keywords: Feminism, Integration, Accounting, Storytelling, Myths

Paper type: Research paper

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper studies single-channel speech separation, assuming unknown, arbitrary temporal dynamics for the speech signals to be separated. A data-driven approach is described, which matches each mixed speech segment against a composite training segment to separate the underlying clean speech segments. To advance the separation accuracy, the new approach seeks and separates the longest mixed speech segments with matching composite training segments. Lengthening the mixed speech segments to match reduces the uncertainty of the constituent training segments, and hence the error of separation. For convenience, we call the new approach Composition of Longest Segments, or CLOSE. The CLOSE method includes a data-driven approach to model long-range temporal dynamics of speech signals, and a statistical approach to identify the longest mixed speech segments with matching composite training segments. Experiments are conducted on the Wall Street Journal database, for separating mixtures of two simultaneous large-vocabulary speech utterances spoken by two different speakers. The results are evaluated using various objective and subjective measures, including the challenge of large-vocabulary continuous speech recognition. It is shown that the new separation approach leads to significant improvement in all these measures.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper describes a data model for content representation of temporal media in an IP based sensor network. The model is formed by introducing the idea of semantic-role from linguistics into the underlying concepts of formal event representation with the aim of developing a common event model. The architecture of a prototype system for a multi camera surveillance system, based on the proposed model is described. The important aspects of the proposed model are its expressiveness, its ability to model content of temporal media, and its suitability for use with a natural language interface. It also provides a platform for temporal information fusion, as well as organizing sensor annotations by help of ontologies.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Decision making is an important element throughout the life-cycle of large-scale projects. Decisions are critical as they have a direct impact upon the success/outcome of a project and are affected by many factors including the certainty and precision of information. In this paper we present an evidential reasoning framework which applies Dempster-Shafer Theory and its variant Dezert-Smarandache Theory to aid decision makers in making decisions where the knowledge available may be imprecise, conflicting and uncertain. This conceptual framework is novel as natural language based information extraction techniques are utilized in the extraction and estimation of beliefs from diverse textual information sources, rather than assuming these estimations as already given. Furthermore we describe an algorithm to define a set of maximal consistent subsets before fusion occurs in the reasoning framework. This is important as inconsistencies between subsets may produce results which are incorrect/adverse in the decision making process. The proposed framework can be applied to problems involving material selection and a Use Case based in the Engineering domain is presented to illustrate the approach. © 2013 Elsevier B.V. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Temporal dynamics and speaker characteristics are two important features of speech that distinguish speech from noise. In this paper, we propose a method to maximally extract these two features of speech for speech enhancement. We demonstrate that this can reduce the requirement for prior information about the noise, which can be difficult to estimate for fast-varying noise. Given noisy speech, the new approach estimates clean speech by recognizing long segments of the clean speech as whole units. In the recognition, clean speech sentences, taken from a speech corpus, are used as examples. Matching segments are identified between the noisy sentence and the corpus sentences. The estimate is formed by using the longest matching segments found in the corpus sentences. Longer speech segments as whole units contain more distinct dynamics and richer speaker characteristics, and can be identified more accurately from noise than shorter speech segments. Therefore, estimation based on the longest recognized segments increases the noise immunity and hence the estimation accuracy. The new approach consists of a statistical model to represent up to sentence-long temporal dynamics in the corpus speech, and an algorithm to identify the longest matching segments between the noisy sentence and the corpus sentences. The algorithm is made more robust to noise uncertainty by introducing missing-feature based noise compensation into the corpus sentences. Experiments have been conducted on the TIMIT database for speech enhancement from various types of nonstationary noise including song, music, and crosstalk speech. The new approach has shown improved performance over conventional enhancement algorithms in both objective and subjective evaluations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Southern Tiwa (Tanoan) exhibits agreement with up to three arguments (ergative, absolutive, dative). This agreement is subject to certain restrictions resembling the Person-Case Constraint paradigm (Bonet 1991). Moreover, there is a correlation between agreement restrictions and conditions on (the obviation of) noun-incorporation in Southern Tiwa, as explicitly and elegantly captured by Rosen (1990) in terms of a heterogeneous feature hierarchy and rules of association. We attempt to recast Rosen’s central insights in terms of Anagnostopoulou’s probe-sharing model of Person-Case Constraint effects (Anagnostopoulou 2003, 2006), to show that the full range of Southern Tiwa agreement and (non-)incorporation restrictions can be given a single, unified analysis within the probe-goal-Agree framework of Chomsky (2001). In particular, we argue that Southern Tiwa’s triple-agreement system is characterized by (a) an independent class probe located on the heads T and v, and (b) a rule that allows this class probe to be deleted in the context of local-person T-agreement. The various restrictions on agreement and non-incorporation then reduce to a single source: failure of class-valuation with DP (as opposed to NP) arguments.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We present three natural language marking strategies based on fast and reliable shallow parsing techniques, and on widely available lexical resources: lexical substitution, adjective conjunction swaps, and relativiser switching. We test these techniques on a random sample of the British National Corpus. Individual candidate marks are checked for goodness of structural and semantic fit, using both lexical resources, and the web as a corpus. A representative sample of marks is given to 25 human judges to evaluate for acceptability and preservation of meaning. This establishes a correlation between corpus based felicity measures and perceived quality, and makes qualified predictions. Grammatical acceptability correlates with our automatic measure strongly (Pearson's r = 0.795, p = 0.001), allowing us to account for about two thirds of variability in human judgements. A moderate but statistically insignificant (Pearson's r = 0.422, p = 0.356) correlation is found with judgements of meaning preservation, indicating that the contextual window of five content words used for our automatic measure may need to be extended. © 2007 SPIE-IS&T.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In most previous research on distributional semantics, Vector Space Models (VSMs) of words are built either from topical information (e.g., documents in which a word is present), or from syntactic/semantic types of words (e.g., dependency parse links of a word in sentences), but not both. In this paper, we explore the utility of combining these two representations to build VSM for the task of semantic composition of adjective-noun phrases. Through extensive experiments on benchmark datasets, we find that even though a type-based VSM is effective for semantic composition, it is often outperformed by a VSM built using a combination of topic- and type-based statistics. We also introduce a new evaluation task wherein we predict the composed vector representation of a phrase from the brain activity of a human subject reading that phrase. We exploit a large syntactically parsed corpus of 16 billion tokens to build our VSMs, with vectors for both phrases and words, and make them publicly available.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Story understanding involves many perceptual and cognitive subprocesses, from perceiving individual words, to parsing sentences, to understanding the relationships among the story characters. We present an integrated computational model of reading that incorporates these and additional subprocesses, simultaneously discovering their fMRI signatures. Our model predicts the fMRI activity associated with reading arbitrary text passages, well enough to distinguish which of two story segments is being read with 74% accuracy. This approach is the first to simultaneously track diverse reading subprocesses during complex story processing and predict the detailed neural representation of diverse story features, ranging from visual word properties to the mention of different story characters and different actions they perform. We construct brain representation maps that replicate many results from a wide range of classical studies that focus each on one aspect of language processing and offer new insights on which type of information is processed by different areas involved in language processing. Additionally, this approach is promising for studying individual differences: it can be used to create single subject maps that may potentially be used to measure reading comprehension and diagnose reading disorders.