984 resultados para Computational-linguistic domain


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a simple technique for reducing the computational effort while solving any geotechnical stability problem by using the upper bound finite element limit analysis and linear optimization. In the proposed method, the problem domain is discretized into a number of different regions in which a particular order (number of sides) of the polygon is chosen to linearize the Mohr-Coulomb yield criterion. A greater order of the polygon needs to be selected only in that region wherein the rate of the plastic strains becomes higher. The computational effort required to solve the problem with this implementation reduces considerably. By using the proposed method, the bearing capacity has been computed for smooth and rough strip footings and the results are found to be quite satisfactory.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper discusses a novel high-speed approach for human action recognition in H. 264/AVC compressed domain. The proposed algorithm utilizes cues from quantization parameters and motion vectors extracted from the compressed video sequence for feature extraction and further classification using Support Vector Machines (SVM). The ultimate goal of our work is to portray a much faster algorithm than pixel domain counterparts, with comparable accuracy, utilizing only the sparse information from compressed video. Partial decoding rules out the complexity of full decoding, and minimizes computational load and memory usage, which can effect in reduced hardware utilization and fast recognition results. The proposed approach can handle illumination changes, scale, and appearance variations, and is robust in outdoor as well as indoor testing scenarios. We have tested our method on two benchmark action datasets and achieved more than 85% accuracy. The proposed algorithm classifies actions with speed (>2000 fps) approximately 100 times more than existing state-of-the-art pixel-domain algorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The goal of this work is to reduce the cost of computing the coefficients in the Karhunen-Loeve (KL) expansion. The KL expansion serves as a useful and efficient tool for discretizing second-order stochastic processes with known covariance function. Its applications in engineering mechanics include discretizing random field models for elastic moduli, fluid properties, and structural response. The main computational cost of finding the coefficients of this expansion arises from numerically solving an integral eigenvalue problem with the covariance function as the integration kernel. Mathematically this is a homogeneous Fredholm equation of second type. One widely used method for solving this integral eigenvalue problem is to use finite element (FE) bases for discretizing the eigenfunctions, followed by a Galerkin projection. This method is computationally expensive. In the current work it is first shown that the shape of the physical domain in a random field does not affect the realizations of the field estimated using KL expansion, although the individual KL terms are affected. Based on this domain independence property, a numerical integration based scheme accompanied by a modification of the domain, is proposed. In addition to presenting mathematical arguments to establish the domain independence, numerical studies are also conducted to demonstrate and test the proposed method. Numerically it is demonstrated that compared to the Galerkin method the computational speed gain in the proposed method is of three to four orders of magnitude for a two dimensional example, and of one to two orders of magnitude for a three dimensional example, while retaining the same level of accuracy. It is also shown that for separable covariance kernels a further cost reduction of three to four orders of magnitude can be achieved. Both normal and lognormal fields are considered in the numerical studies. (c) 2014 Elsevier B.V. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we propose a H.264/AVC compressed domain human action recognition system with projection based metacognitive learning classifier (PBL-McRBFN). The features are extracted from the quantization parameters and the motion vectors of the compressed video stream for a time window and used as input to the classifier. Since compressed domain analysis is done with noisy, sparse compression parameters, it is a huge challenge to achieve performance comparable to pixel domain analysis. On the positive side, compressed domain allows rapid analysis of videos compared to pixel level analysis. The classification results are analyzed for different values of Group of Pictures (GOP) parameter, time window including full videos. The functional relationship between the features and action labels are established using PBL-McRBFN with a cognitive and meta-cognitive component. The cognitive component is a radial basis function, while the meta-cognitive component employs self-regulation to achieve better performance in subject independent action recognition task. The proposed approach is faster and shows comparable performance with respect to the state-of-the-art pixel domain counterparts. It employs partial decoding, which rules out the complexity of full decoding, and minimizes computational load and memory usage. This results in reduced hardware utilization and increased speed of classification. The results are compared with two benchmark datasets and show more than 90% accuracy using the PBL-McRBFN. The performance for various GOP parameters and group of frames are obtained with twenty random trials and compared with other well-known classifiers in machine learning literature. (C) 2015 Elsevier B.V. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We have developed an integrated database for Mycobacterium tuberculosis H37Rv (Mtb) that collates information on protein sequences, domain assignments, functional annotation and 3D structural information along with protein-protein and protein-small molecule interactions. SInCRe (Structural Interactome Computational Resource) is developed out of CamBan (Cambridge and Bangalore) collaboration. The motivation for development of this database is to provide an integrated platform to allow easily access and interpretation of data and results obtained by all the groups in CamBan in the field of Mtb informatics. In-house algorithms and databases developed independently by various academic groups in CamBan are used to generate Mtb-specific datasets and are integrated in this database to provide a structural dimension to studies on tuberculosis. The SInCRe database readily provides information on identification of functional domains, genome-scale modelling of structures of Mtb proteins and characterization of the small-molecule binding sites within Mtb. The resource also provides structure-based function annotation, information on small-molecule binders including FDA (Food and Drug Administration)-approved drugs, protein-protein interactions (PPIs) and natural compounds that bind to pathogen proteins potentially and result in weakening or elimination of host-pathogen protein-protein interactions. Together they provide prerequisites for identification of off-target binding.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper discusses a novel high-speed approach for human action recognition in H.264/AVC compressed domain. The proposed algorithm utilizes cues from quantization parameters and motion vectors extracted from the compressed video sequence for feature extraction and further classification using Support Vector Machines (SVM). The ultimate goal of the proposed work is to portray a much faster algorithm than pixel domain counterparts, with comparable accuracy, utilizing only the sparse information from compressed video. Partial decoding rules out the complexity of full decoding, and minimizes computational load and memory usage, which can result in reduced hardware utilization and faster recognition results. The proposed approach can handle illumination changes, scale, and appearance variations, and is robust to outdoor as well as indoor testing scenarios. We have evaluated the performance of the proposed method on two benchmark action datasets and achieved more than 85 % accuracy. The proposed algorithm classifies actions with speed (> 2,000 fps) approximately 100 times faster than existing state-of-the-art pixel-domain algorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Image and video analysis requires rich features that can characterize various aspects of visual information. These rich features are typically extracted from the pixel values of the images and videos, which require huge amount of computation and seldom useful for real-time analysis. On the contrary, the compressed domain analysis offers relevant information pertaining to the visual content in the form of transform coefficients, motion vectors, quantization steps, coded block patterns with minimal computational burden. The quantum of work done in compressed domain is relatively much less compared to pixel domain. This paper aims to survey various video analysis efforts published during the last decade across the spectrum of video compression standards. In this survey, we have included only the analysis part, excluding the processing aspect of compressed domain. This analysis spans through various computer vision applications such as moving object segmentation, human action recognition, indexing, retrieval, face detection, video classification and object tracking in compressed videos.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Crowd flow segmentation is an important step in many video surveillance tasks. In this work, we propose an algorithm for segmenting flows in H.264 compressed videos in a completely unsupervised manner. Our algorithm works on motion vectors which can be obtained by partially decoding the compressed video without extracting any additional features. Our approach is based on modelling the motion vector field as a Conditional Random Field (CRF) and obtaining oriented motion segments by finding the optimal labelling which minimises the global energy of CRF. These oriented motion segments are recursively merged based on gradient across their boundaries to obtain the final flow segments. This work in compressed domain can be easily extended to pixel domain by substituting motion vectors with motion based features like optical flow. The proposed algorithm is experimentally evaluated on a standard crowd flow dataset and its superior performance in both accuracy and computational time are demonstrated through quantitative results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A quasi-steady time domain method is developed for the prediction of dynamic behavior of a mooring system under the environmental disturbances, such as regular or irregular waves, winds and currents. The mooring forces are obtained in a static sense at each instant. The dynamic feature of the mooring cables can be obtained by incorporating the extended 3-D lumped-mass method with the known ship motion history. Some nonlinear effects, such as the influence of the instantaneous change of the wetted hull surface on the hydrostatic restoring forces and Froude-Krylov forces, are included. The computational results show a satisfactory agreement with the experimental ones.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The high Reynolds number flow contains a wide range of length and time scales, and the flow domain can be divided into several sub-domains with different characteristic scales. In some sub-domains, the viscosity dissipation scale can only be considered in a certain direction; in some sub-domains, the viscosity dissipation scales need to be considered in all directions; in some sub-domains, the viscosity dissipation scales are unnecessary to be considered at all. For laminar boundary layer region, the characteristic length scales in the streamwise and normal directions are L and L Re-1/ 2 , respectively. The characteristic length scale and the velocity scale in the outer region of the boundary layer are L and U, respectively. In the neighborhood region of the separated point, the length scale l<domain decomposition method is developed for the high Reynolds number flow. First, the whole domain is decomposed to different sub-domains with the different characteristic scales. Then the different dominant equation of all sub-domains is defined according to the diffusion parabolized (DP) theory of viscous flow. Finally these different equations are solved simultaneously in whole computational region. For numerical tests of high Reynolds numerical flows, two-dimensional supersonic flows over rearward and frontward steps as well as an interaction flow between shock wave and boundary layer were solved numerically. The pressure distributions and local coefficients of skin friction on the wall are given. The numerical results obtained by the multiscale-domain decomposition algorithm are well agreement with those by NS equations. Comparing with the usual method of solving the Navier-Stokes equations in the whole flow, under the same numerical accuracy, the present multiscale domain decomposition method decreases CPU consuming about 20% and reflects the physical mechanism of practical flow more accurately.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a pseudo-time-step method to calculate a (vector) Green function for the adjoint linearised Euler equations as a scattering problem in the frequency domain, for use as a jet-noise propagation prediction tool. A method of selecting the acoustics-related solution in a truncated spatial domain while suppressing any possible shear-layer-type instability is presented. Numerical tests for 3-D axisymmetrical parallel mean flows against semi-analytical reference solutions indicate that the new iterative algorithm is capable of producing accurate solutions with modest computational requirements.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper a method to incorporate linguistic information regarding single-word and compound verbs is proposed, as a first step towards an SMT model based on linguistically-classified phrases. By substituting these verb structures by the base form of the head verb, we achieve a better statistical word alignment performance, and are able to better estimate the translation model and generalize to unseen verb forms during translation. Preliminary experiments for the English - Spanish language pair are performed, and future research lines are detailed. © 2005 Association for Computational Linguistics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Human listeners can identify vowels regardless of speaker size, although the sound waves for an adult and a child speaking the ’same’ vowel would differ enormously. The differences are mainly due to the differences in vocal tract length (VTL) and glottal pulse rate (GPR) which are both related to body size. Automatic speech recognition machines are notoriously bad at understanding children if they have been trained on the speech of an adult. In this paper, we propose that the auditory system adapts its analysis of speech sounds, dynamically and automatically to the GPR and VTL of the speaker on a syllable-to-syllable basis. We illustrate how this rapid adaptation might be performed with the aid of a computational version of the auditory image model, and we propose that an auditory preprocessor of this form would improve the robustness of speech recognisers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The computer science technique of computational complexity analysis can provide powerful insights into the algorithm-neutral analysis of information processing tasks. Here we show that a simple, theory-neutral linguistic model of syntactic agreement and ambiguity demonstrates that natural language parsing may be computationally intractable. Significantly, we show that it may be syntactic features rather than rules that can cause this difficulty. Informally, human languages and the computationally intractable Satisfiability (SAT) problem share two costly computional mechanisms: both enforce agreement among symbols across unbounded distances (Subject-Verb agreement) and both allow ambiguity (is a word a Noun or a Verb?).