980 resultados para Search problems


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Experimental and simulation studies have uncovered at least two anomalous concentration regimes in water-dimethyl sulfoxide (DMSO) binary mixture whose precise origin has remained a subject of debate. In order to facilitate time domain experimental investigation of the dynamics of such binary mixtures, we explore strength or extent of influence of these anomalies in dipolar solvation dynamics by carrying out long molecular dynamics simulations over a wide range of DMSO concentration. The solvation time correlation function so calculated indeed displays strong composition dependent anomalies, reflected in pronounced non-exponential kinetics and non-monotonous composition dependence of the average solvation time constant. In particular, we find remarkable slow-down in the solvation dynamics around 10%-20% and 35%-50% mole percentage. We investigate microscopic origin of these two anomalies. The population distribution analyses of different structural morphology elucidate that these two slowing down are reflections of intriguing structural transformations in water-DMSO mixture. The structural transformations themselves can be explained in terms of a change in the relative coordination number of DMSO and water molecules, from 1DMSO:2H(2)O to 1H(2)O:1DMSO and 1H(2)O:2DMSO complex formation. Thus, while the emergence of first slow down (at 15% DMSO mole percentage) is due to the percolation among DMSO molecules supported by the water molecules (whose percolating network remains largely unaffected), the 2nd anomaly (centered on 40%-50%) is due to the formation of the network structure where the unit of 1DMSO:1H(2)O and 2DMSO:1H(2)O dominates to give rise to rich dynamical features. Through an analysis of partial solvation dynamics an interesting negative cross-correlation between water and DMSO is observed that makes an important contribution to relaxation at intermediate to longer times.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

ZnO is a unique material with numerous applications. There has been great interest in the synthesis of long nanowires to explore new technology coupled with length. But the quest to enhance the length is limited by various experimental shortcomings such as catalytic poisoning, degradation of the precursors and growth in all possible directions. This review article focuses on the growth of ultralong ZnO nanowires along with possible methodologies to overcome these limitations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We consider the problem of finding the best features for value function approximation in reinforcement learning and develop an online algorithm to optimize the mean square Bellman error objective. For any given feature value, our algorithm performs gradient search in the parameter space via a residual gradient scheme and, on a slower timescale, also performs gradient search in the Grassman manifold of features. We present a proof of convergence of our algorithm. We show empirical results using our algorithm as well as a similar algorithm that uses temporal difference learning in place of the residual gradient scheme for the faster timescale updates.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Transductive SVM (TSVM) is a well known semi-supervised large margin learning method for binary text classification. In this paper we extend this method to multi-class and hierarchical classification problems. We point out that the determination of labels of unlabeled examples with fixed classifier weights is a linear programming problem. We devise an efficient technique for solving it. The method is applicable to general loss functions. We demonstrate the value of the new method using large margin loss on a number of multi-class and hierarchical classification datasets. For maxent loss we show empirically that our method is better than expectation regularization/constraint and posterior regularization methods, and competitive with the version of entropy regularization method which uses label constraints.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we consider the problem of finding a spectrum hole of a specified bandwidth in a given wide band of interest. We propose a new, simple and easily implementable sub-Nyquist sampling scheme for signal acquisition and a spectrum hole search algorithm that exploits sparsity in the primary spectral occupancy in the frequency domain by testing a group of adjacent subbands in a single test. The sampling scheme deliberately introduces aliasing during signal acquisition, resulting in a signal that is the sum of signals from adjacent sub-bands. Energy-based hypothesis tests are used to provide an occupancy decision over the group of subbands, and this forms the basis of the proposed algorithm to find contiguous spectrum holes. We extend this framework to a multi-stage sensing algorithm that can be employed in a variety of spectrum sensing scenarios, including non-contiguous spectrum hole search. Further, we provide the analytical means to optimize the hypothesis tests with respect to the detection thresholds, number of samples and group size to minimize the detection delay under a given error rate constraint. Depending on the sparsity and SNR, the proposed algorithms can lead to significantly lower detection delays compared to a conventional bin-by-bin energy detection scheme; the latter is in fact a special case of the group test when the group size is set to 1. We validate our analytical results via Monte Carlo simulations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Visual search in real life involves complex displays with a target among multiple types of distracters, but in the laboratory, it is often tested using simple displays with identical distracters. Can complex search be understood in terms of simple searches? This link may not be straightforward if complex search has emergent properties. One such property is linear separability, whereby search is hard when a target cannot be separated from its distracters using a single linear boundary. However, evidence in favor of linear separability is based on testing stimulus configurations in an external parametric space that need not be related to their true perceptual representation. We therefore set out to assess whether linear separability influences complex search at all. Our null hypothesis was that complex search performance depends only on classical factors such as target-distracter similarity and distracter homogeneity, which we measured using simple searches. Across three experiments involving a variety of artificial and natural objects, differences between linearly separable and nonseparable searches were explained using target-distracter similarity and distracter heterogeneity. Further, simple searches accurately predicted complex search regardless of linear separability (r = 0.91). Our results show that complex search is explained by simple search, refuting the widely held belief that linear separability influences visual search.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Compressive Sensing theory combines the signal sampling and compression for sparse signals resulting in reduction in sampling rate and computational complexity of the measurement system. In recent years, many recovery algorithms were proposed to reconstruct the signal efficiently. Look Ahead OMP (LAOMP) is a recently proposed method which uses a look ahead strategy and performs significantly better than other greedy methods. In this paper, we propose a modification to the LAOMP algorithm to choose the look ahead parameter L adaptively, thus reducing the complexity of the algorithm, without compromising on the performance. The performance of the algorithm is evaluated through Monte Carlo simulations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Three codes, that can solve three dimensional linear elastostatic problems using constant boundary elements while ignoring body forces, are provided here. The file 'bemconst.m' contains a MATLAB code for solving three dimensional linear elastostatic problems using constant boundary elements while ignoring body forces. The file 'bemconst.f90' is a Fortran translation of the MATLAB code contained in the file 'bemconst.m'. The file 'bemconstp.f90' is a parallelized version of the Fortran code contained in the file 'bemconst.f90'. The file 'inbem96.txt' is the input file for the Fortran codes contained in the files 'bemconst.f90' and 'bemconstp.f90'. Author hereby declares that the present codes are the original works of the author. Further, author hereby declares that any of the present codes, in full or in part, is not a translation or a copy of any of the existing codes written by someone else. Author's institution (Indian Institute of Science) has informed the author in writing that the institution is not interested in claiming any copyright on the present codes. Author is hereby distributing the present codes under the MIT License; full text of the license is included in each of the files that contain the codes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this article, we derive an a posteriori error estimator for various discontinuous Galerkin (DG) methods that are proposed in (Wang, Han and Cheng, SIAM J. Numer. Anal., 48: 708-733, 2010) for an elliptic obstacle problem. Using a key property of DG methods, we perform the analysis in a general framework. The error estimator we have obtained for DG methods is comparable with the estimator for the conforming Galerkin (CG) finite element method. In the analysis, we construct a non-linear smoothing function mapping DG finite element space to CG finite element space and use it as a key tool. The error estimator consists of a discrete Lagrange multiplier associated with the obstacle constraint. It is shown for non-over-penalized DG methods that the discrete Lagrange multiplier is uniformly stable on non-uniform meshes. Finally, numerical results demonstrating the performance of the error estimator are presented.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this work, we present a finite element formulation for the Saint-Venant torsion and bending problems for prismatic beams. The torsion problem formulation is based on the warping function, and can handle multiply-connected regions (including thin-walled structures), compound and anisotropic bars. Similarly, the bending formulation, which is based on linearized elasticity theory, can handle multiply-connected domains including thin-walled sections. The torsional rigidity and shear centers can be found as special cases of these formulations. Numerical results are presented to show the good coarse-mesh accuracy of both the formulations for both the displacement and stress fields. The stiffness matrices and load vectors (which are similar to those for a variable body force in a conventional structural mechanics problem) in both formulations involve only domain integrals, which makes them simple to implement and computationally efficient. (C) 2014 Elsevier Ltd. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents a simple technique for reducing the computational effort while solving any geotechnical stability problem by using the upper bound finite element limit analysis and linear optimization. In the proposed method, the problem domain is discretized into a number of different regions in which a particular order (number of sides) of the polygon is chosen to linearize the Mohr-Coulomb yield criterion. A greater order of the polygon needs to be selected only in that region wherein the rate of the plastic strains becomes higher. The computational effort required to solve the problem with this implementation reduces considerably. By using the proposed method, the bearing capacity has been computed for smooth and rough strip footings and the results are found to be quite satisfactory.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We develop iterative diffraction tomography algorithms, which are similar to the distorted Born algorithms, for inverting scattered intensity data. Within the Born approximation, the unknown scattered field is expressed as a multiplicative perturbation to the incident field. With this, the forward equation becomes stable, which helps us compute nearly oscillation-free solutions that have immediate bearing on the accuracy of the Jacobian computed for use in a deterministic Gauss-Newton (GN) reconstruction. However, since the data are inherently noisy and the sensitivity of measurement to refractive index away from the detectors is poor, we report a derivative-free evolutionary stochastic scheme, providing strictly additive updates in order to bridge the measurement-prediction misfit, to arrive at the refractive index distribution from intensity transport data. The superiority of the stochastic algorithm over the GN scheme for similar settings is demonstrated by the reconstruction of the refractive index profile from simulated and experimentally acquired intensity data. (C) 2014 Optical Society of America

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A numerical formulation has been proposed for solving an axisymmetric stability problem in geomechanics with upper bound limit analysis, finite elements, and linear optimization. The Drucker-Prager yield criterion is linearized by simulating a sphere with a circumscribed truncated icosahedron. The analysis considers only the velocities and plastic multiplier rates, not the stresses, as the basic unknowns. The formulation is simple to implement, and it has been employed for finding the collapse loads of a circular footing placed over the surface of a cohesive-frictional material. The formulation can be used to solve any general axisymmetric geomechanics stability problem.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Single features such as line orientation and length are known to guide visual search, but relatively little is known about how multiple features combine in search. To address this question, we investigated how search for targets differing in multiple features ( intensity, length, orientation) from the distracters is related to searches for targets differing in each of the individual features. We tested race models (based on reaction times) and coactivation models ( based on reciprocal of reaction times) for their ability to predict multiple feature searches. Multiple feature searches were best accounted for by a co-activation model in which feature information combined linearly (r = 0.95). This result agrees with the classic finding that these features are separable i.e., subjective dissimilarity ratings sum linearly. We then replicated the classical finding that the length and width of a rectangle are integral features-in other words, they combine nonlinearly in visual search. However, to our surprise, upon including aspect ratio as an additional feature, length and width combined linearly and this model outperformed all other models. Thus, length and width of a rectangle became separable when considered together with aspect ratio. This finding predicts that searches involving shapes with identical aspect ratio should be more difficult than searches where shapes differ in aspect ratio. We confirmed this prediction on a variety of shapes. We conclude that features in visual search co-activate linearly and demonstrate for the first time that aspect ratio is a novel feature that guides visual search.