972 resultados para Language contact
Resumo:
A computational impact analysis methodology has been developed, based on modal analysis and a local contact force-deflection model. The contact law is based on Hertz contact theory while contact stresses are elastic, defines a modified contact theory to take account of local permanent indentation, and considers elastic recovery during unloading. The model was validated experimentally through impact testing of glass-carbon hybrid braided composite panels. Specimens were mounted in a support frame and the contact force was inferred from the deceleration of the impactor, measured by high-speed photography. A Finite Element analysis of the panel and support frame assembly was performed to compute the modal responses. The new contact model performed well in predicting the peak forces and impact durations for moderate energy impacts (15 J), where contact stresses locally exceed the linear elastic limit and damage may be deemed to have occurred. C-scan measurements revealed substantial damage for impact energies in the range of 30-50 J. For this regime the new model predictions might be improved by characterisation of the contact law hysteresis during the unloading phase, and a modification of the elastic vibration response in line with damage levels acquired during the impact. © 2011 Elsevier Ltd. All rights reserved.
Resumo:
An increasingly common scenario in building speech synthesis and recognition systems is training on inhomogeneous data. This paper proposes a new framework for estimating hidden Markov models on data containing both multiple speakers and multiple languages. The proposed framework, speaker and language factorization, attempts to factorize speaker-/language-specific characteristics in the data and then model them using separate transforms. Language-specific factors in the data are represented by transforms based on cluster mean interpolation with cluster-dependent decision trees. Acoustic variations caused by speaker characteristics are handled by transforms based on constrained maximum-likelihood linear regression. Experimental results on statistical parametric speech synthesis show that the proposed framework enables data from multiple speakers in different languages to be used to: train a synthesis system; synthesize speech in a language using speaker characteristics estimated in a different language; and adapt to a new language. © 2012 IEEE.
Resumo:
Most previous work on trainable language generation has focused on two paradigms: (a) using a statistical model to rank a set of generated utterances, or (b) using statistics to inform the generation decision process. Both approaches rely on the existence of a handcrafted generator, which limits their scalability to new domains. This paper presents BAGEL, a statistical language generator which uses dynamic Bayesian networks to learn from semantically-aligned data produced by 42 untrained annotators. A human evaluation shows that BAGEL can generate natural and informative utterances from unseen inputs in the information presentation domain. Additionally, generation performance on sparse datasets is improved significantly by using certainty-based active learning, yielding ratings close to the human gold standard with a fraction of the data. © 2010 Association for Computational Linguistics.
Resumo:
This paper describes a new approach to model the forces on a tread block for a free-rolling tyre in contact with a rough road. A theoretical analysis based on realistic tread mechanical properties and road roughness is presented, indicating partial contact between a tread block and a rough road. Hence an asperity-scale indentation model is developed using a semi-empirical formulation, taking into account both the rubber viscoelasticity and the tread block geometry. The model aims to capture the essential details of the contact at the simplest level, to make it suitable as part of a time-domain dynamic analysis of the coupled tyre-road system. The indentation model is found to have a good correlation with the finite element (FE) predictions and is validated against experimental results using a rolling contact rig. When coupled to a deformed tyre belt profile, the indentation model predicts normal and tangential force histories inside the tyre contact patch that show good agreement with FE predictions. © 2012 Elsevier B.V..
Resumo:
State-of-the-art large vocabulary continuous speech recognition (LVCSR) systems often combine outputs from multiple subsystems developed at different sites. Cross system adaptation can be used as an alternative to direct hypothesis level combination schemes such as ROVER. The standard approach involves only cross adapting acoustic models. To fully exploit the complimentary features among sub-systems, language model (LM) cross adaptation techniques can be used. Previous research on multi-level n-gram LM cross adaptation is extended to further include the cross adaptation of neural network LMs in this paper. Using this improved LM cross adaptation framework, significant error rate gains of 4.0%-7.1% relative were obtained over acoustic model only cross adaptation when combining a range of Chinese LVCSR sub-systems used in the 2010 and 2011 DARPA GALE evaluations. Copyright © 2011 ISCA.
Resumo:
We use multispeckle diffusive wave spectroscopy to probe the micron-scale dynamics of a water-saturated granular pile submitted to discrete gentle taps. The typical time scale between plastic events is found to increase dramatically with the number of applied taps. Furthermore, this microscopic dynamics weakly depends on the solid fraction of the sample. This process is largely analogous to the aging phenomenon observed in thermal glassy systems. We propose a heuristic model where this slowing-down mechanism is associated with a slow evolution of the distribution of the contact forces between particles. This model accounts for the main features of the observed dynamics.
Resumo:
Language models (LMs) are often constructed by building multiple individual component models that are combined using context independent interpolation weights. By tuning these weights, using either perplexity or discriminative approaches, it is possible to adapt LMs to a particular task. This paper investigates the use of context dependent weighting in both interpolation and test-time adaptation of language models. Depending on the previous word contexts, a discrete history weighting function is used to adjust the contribution from each component model. As this dramatically increases the number of parameters to estimate, robust weight estimation schemes are required. Several approaches are described in this paper. The first approach is based on MAP estimation where interpolation weights of lower order contexts are used as smoothing priors. The second approach uses training data to ensure robust estimation of LM interpolation weights. This can also serve as a smoothing prior for MAP adaptation. A normalized perplexity metric is proposed to handle the bias of the standard perplexity criterion to corpus size. A range of schemes to combine weight information obtained from training data and test data hypotheses are also proposed to improve robustness during context dependent LM adaptation. In addition, a minimum Bayes' risk (MBR) based discriminative training scheme is also proposed. An efficient weighted finite state transducer (WFST) decoding algorithm for context dependent interpolation is also presented. The proposed technique was evaluated using a state-of-the-art Mandarin Chinese broadcast speech transcription task. Character error rate (CER) reductions up to 7.3 relative were obtained as well as consistent perplexity improvements. © 2012 Elsevier Ltd. All rights reserved.