959 resultados para mesh-free method
Resumo:
The detection of buried objects using time-domain freespace measurements was carried out in the near field. The location of a hidden object was determined from an analysis of the reflected signal. This method can be extended to detect any number of objects. Measurements were carried out in the X- and Ku-bands using ordinary rectangular pyramidal horn antennas of gain 15 dB. The same antenna was used as the transmitter and recei er. The experimental results were compared with simulated results by applying the two-dimensional finite-difference time-domain(FDTD)method, and agree well with each other. The dispersi e nature of the dielectric medium was considered for the simulation.
Resumo:
With the increasing popularity of wireless network and its application, mobile ad-hoc networks (MANETS) emerged recently. MANET topology is highly dynamic in nature and nodes are highly mobile so that the rate of link failure is more in MANET. There is no central control over the nodes and the control is distributed among nodes and they can act as either router or source. MANTEs have been considered as isolated stand-alone network. Node can add or remove at any time and it is not infrastructure dependent. So at any time at any where the network can setup and a trouble free communication is possible. Due to more chances of link failures, collisions and transmission errors in MANET, the maintenance of network became costly. As per the study more frequent link failures became an important aspect of diminishing the performance of the network and also it is not predictable. The main objective of this paper is to study the route instability in AODV protocol and suggest a solution for improvement. This paper proposes a new approach to reduce the route failure by storing the alternate route in the intermediate nodes. In this algorithm intermediate nodes are also involved in the route discovery process. This reduces the route establishment overhead as well as the time to find the reroute when a link failure occurs.
Resumo:
Using Doppler-free two-photon absorption spectroscopy, we have measured hyperfine splitting constants as well as isotopic level shifts of the 6s^2 np ^2 P_l/2,3/2 states in (n=7-10) in ^203 TI and ^205 TI. Calculations for hyperfine constants and electron density at the nucleus have been performed by the Dirac-Fock method. The experimental results are compared with these calculations as well as with the predictions of the semiempirical theory.
Resumo:
This participatory action research was based on a experience of educational intervention on La Cruz and Bello Oriente (Manrique-Medellin), a marginal zone in the northeastern part of the Commune 3 in Medellin,. Colombia. In this marginal sector, psychosocial problems seem to be associated to limited educational and employment opportunities, domestic violence, illegal armed forces, sexual abuse, social discrimination, and lack of adequate public services, among others. All these are also considered as risk factors for drug dependency. We used a structured interview designed to identify leisure tendencies, use of free time, and tendencies in recreational activities. Data from the interview were triangulated with information collected by observation and in field work and used to build a psycho-pedagogic method based on play and leisure activities. The effects of the use of this educational intervention on the satisfaction of human needs were analyzed in light of the theory of Manfred Max-Neef. Results point out the need for new educational strategies aimed to promote creativity, solidarity, mental, physical and social health, more enthusiasm and motivation and in general, positive attitudes that help prevent drug dependence.
Resumo:
Resumen tomado de la publicación. Con el apoyo económico del departamento MIDE de la UNED
Resumo:
A select-divide-and-conquer variational method to approximate configuration interaction (CI) is presented. Given an orthonormal set made up of occupied orbitals (Hartree-Fock or similar) and suitable correlation orbitals (natural or localized orbitals), a large N-electron target space S is split into subspaces S0,S1,S2,...,SR. S0, of dimension d0, contains all configurations K with attributes (energy contributions, etc.) above thresholds T0={T0egy, T0etc.}; the CI coefficients in S0 remain always free to vary. S1 accommodates KS with attributes above T1≤T0. An eigenproblem of dimension d0+d1 for S0+S 1 is solved first, after which the last d1 rows and columns are contracted into a single row and column, thus freezing the last d1 CI coefficients hereinafter. The process is repeated with successive Sj(j≥2) chosen so that corresponding CI matrices fit random access memory (RAM). Davidson's eigensolver is used R times. The final energy eigenvalue (lowest or excited one) is always above the corresponding exact eigenvalue in S. Threshold values {Tj;j=0, 1, 2,...,R} regulate accuracy; for large-dimensional S, high accuracy requires S 0+S1 to be solved outside RAM. From there on, however, usually a few Davidson iterations in RAM are needed for each step, so that Hamiltonian matrix-element evaluation becomes rate determining. One μhartree accuracy is achieved for an eigenproblem of order 24 × 106, involving 1.2 × 1012 nonzero matrix elements, and 8.4×109 Slater determinants
Resumo:
An analytic method to evaluate nuclear contributions to electrical properties of polyatomic molecules is presented. Such contributions control changes induced by an electric field on equilibrium geometry (nuclear relaxation contribution) and vibrational motion (vibrational contribution) of a molecular system. Expressions to compute the nuclear contributions have been derived from a power series expansion of the potential energy. These contributions to the electrical properties are given in terms of energy derivatives with respect to normal coordinates, electric field intensity or both. Only one calculation of such derivatives at the field-free equilibrium geometry is required. To show the useful efficiency of the analytical evaluation of electrical properties (the so-called AEEP method), results for calculations on water and pyridine at the SCF/TZ2P and the MP2/TZ2P levels of theory are reported. The results obtained are compared with previous theoretical calculations and with experimental values
Resumo:
In this paper a cell by cell anisotropic adaptive mesh technique is added to an existing staggered mesh Lagrange plus remap finite element ALE code for the solution of the Euler equations. The quadrilateral finite elements may be subdivided isotropically or anisotropically and a hierarchical data structure is employed. An efficient computational method is proposed, which only solves on the finest level of resolution that exists for each part of the domain with disjoint or hanging nodes being used at resolution transitions. The Lagrangian, equipotential mesh relaxation and advection (solution remapping) steps are generalised so that they may be applied on the dynamic mesh. It is shown that for a radial Sod problem and a two-dimensional Riemann problem the anisotropic adaptive mesh method runs over eight times faster.
Resumo:
We consider the problem of scattering of a time-harmonic acoustic incident plane wave by a sound soft convex polygon. For standard boundary or finite element methods, with a piecewise polynomial approximation space, the computational cost required to achieve a prescribed level of accuracy grows linearly with respect to the frequency of the incident wave. Recently Chandler–Wilde and Langdon proposed a novel Galerkin boundary element method for this problem for which, by incorporating the products of plane wave basis functions with piecewise polynomials supported on a graded mesh into the approximation space, they were able to demonstrate that the number of degrees of freedom required to achieve a prescribed level of accuracy grows only logarithmically with respect to the frequency. Here we propose a related collocation method, using the same approximation space, for which we demonstrate via numerical experiments a convergence rate identical to that achieved with the Galerkin scheme, but with a substantially reduced computational cost.
Resumo:
In this paper we consider the problem of time-harmonic acoustic scattering in two dimensions by convex polygons. Standard boundary or finite element methods for acoustic scattering problems have a computational cost that grows at least linearly as a function of the frequency of the incident wave. Here we present a novel Galerkin boundary element method, which uses an approximation space consisting of the products of plane waves with piecewise polynomials supported on a graded mesh, with smaller elements closer to the corners of the polygon. We prove that the best approximation from the approximation space requires a number of degrees of freedom to achieve a prescribed level of accuracy that grows only logarithmically as a function of the frequency. Numerical results demonstrate the same logarithmic dependence on the frequency for the Galerkin method solution. Our boundary element method is a discretization of a well-known second kind combined-layer-potential integral equation. We provide a proof that this equation and its adjoint are well-posed and equivalent to the boundary value problem in a Sobolev space setting for general Lipschitz domains.
Resumo:
A numerical algorithm for the biharmonic equation in domains with piecewise smooth boundaries is presented. It is intended for problems describing the Stokes flow in the situations where one has corners or cusps formed by parts of the domain boundary and, due to the nature of the boundary conditions on these parts of the boundary, these regions have a global effect on the shape of the whole domain and hence have to be resolved with sufficient accuracy. The algorithm combines the boundary integral equation method for the main part of the flow domain and the finite-element method which is used to resolve the corner/cusp regions. Two parts of the solution are matched along a numerical ‘internal interface’ or, as a variant, two interfaces, and they are determined simultaneously by inverting a combined matrix in the course of iterations. The algorithm is illustrated by considering the flow configuration of ‘curtain coating’, a flow where a sheet of liquid impinges onto a moving solid substrate, which is particularly sensitive to what happens in the corner region formed, physically, by the free surface and the solid boundary. The ‘moving contact line problem’ is addressed in the framework of an earlier developed interface formation model which treats the dynamic contact angle as part of the solution, as opposed to it being a prescribed function of the contact line speed, as in the so-called ‘slip models’. Keywords: Dynamic contact angle; finite elements; free surface flows; hybrid numerical technique; Stokes equations.
Resumo:
In this paper we consider the impedance boundary value problem for the Helmholtz equation in a half-plane with piecewise constant boundary data, a problem which models, for example, outdoor sound propagation over inhomogeneous. at terrain. To achieve good approximation at high frequencies with a relatively low number of degrees of freedom, we propose a novel Galerkin boundary element method, using a graded mesh with smaller elements adjacent to discontinuities in impedance and a special set of basis functions so that, on each element, the approximation space contains polynomials ( of degree.) multiplied by traces of plane waves on the boundary. We prove stability and convergence and show that the error in computing the total acoustic field is O( N-(v+1) log(1/2) N), where the number of degrees of freedom is proportional to N logN. This error estimate is independent of the wavenumber, and thus the number of degrees of freedom required to achieve a prescribed level of accuracy does not increase as the wavenumber tends to infinity.
Resumo:
In this paper we show stability and convergence for a novel Galerkin boundary element method approach to the impedance boundary value problem for the Helmholtz equation in a half-plane with piecewise constant boundary data. This problem models, for example, outdoor sound propagation over inhomogeneous flat terrain. To achieve a good approximation with a relatively low number of degrees of freedom we employ a graded mesh with smaller elements adjacent to discontinuities in impedance, and a special set of basis functions for the Galerkin method so that, on each element, the approximation space consists of polynomials (of degree $\nu$) multiplied by traces of plane waves on the boundary. In the case where the impedance is constant outside an interval $[a,b]$, which only requires the discretization of $[a,b]$, we show theoretically and experimentally that the $L_2$ error in computing the acoustic field on $[a,b]$ is ${\cal O}(\log^{\nu+3/2}|k(b-a)| M^{-(\nu+1)})$, where $M$ is the number of degrees of freedom and $k$ is the wavenumber. This indicates that the proposed method is especially commendable for large intervals or a high wavenumber. In a final section we sketch how the same methodology extends to more general scattering problems.
Resumo:
Flow in the world's oceans occurs at a wide range of spatial scales, from a fraction of a metre up to many thousands of kilometers. In particular, regions of intense flow are often highly localised, for example, western boundary currents, equatorial jets, overflows and convective plumes. Conventional numerical ocean models generally use static meshes. The use of dynamically-adaptive meshes has many potential advantages but needs to be guided by an error measure reflecting the underlying physics. A method of defining an error measure to guide an adaptive meshing algorithm for unstructured tetrahedral finite elements, utilizing an adjoint or goal-based method, is described here. This method is based upon a functional, encompassing important features of the flow structure. The sensitivity of this functional, with respect to the solution variables, is used as the basis from which an error measure is derived. This error measure acts to predict those areas of the domain where resolution should be changed. A barotropic wind driven gyre problem is used to demonstrate the capabilities of the method. The overall objective of this work is to develop robust error measures for use in an oceanographic context which will ensure areas of fine mesh resolution are used only where and when they are required. (c) 2006 Elsevier Ltd. All rights reserved.
Resumo:
1. Estimates of seed bank depletion rates are essential for modelling and management of plant populations. The seed bag burial method is often used to measure seed mortality in the soil. However, the density of seeds within seed bags is higher than densities in natural seed banks, which may elevate levels of pathogens and influence seed mortality. The aim of this study was to quantify the effects of fungi and seed density within buried mesh bags on the mortality of seeds. Striga hermonthica was chosen as the study species because it has been widely studied but different methods for measuring seed mortality in the soil have yielded contradictory estimates. 2. Seed bags were buried in soil and exhumed at regular time intervals to monitor mortality of the seeds in three field experiments during two rainy seasons. The effect of fungal activity on seed mortality was evaluated in a fungi exclusion experiment. Differences in seed-to-seed interaction were obtained by using two and four densities within the seed bags in consecutive years. Densities were created by mixing 1000 seeds with 0, 10, 100 or 1000 g of coarse sand. 3. The mortality rate was significantly lower when fungi were excluded, indicating the possible role of pathogenic fungi. 4. Decreasing the density of seeds in bags significantly reduced seed mortality, most probably because of decreased seed-to-seed contamination by pathogenic fungi. 5. Synthesis and applications. Models of plant populations in general and annual weeds in particular often use values from the literature for seed bank depletion rates. These depletion rates have often been estimated by the seed bag burial method, yet seed density within seed bags may be unrealistically high. Consequently, estimates of seed mortality rates may be too high because of an overestimation of the effects of soil or seed-borne pathogens. Species that have been classified from such studies as having short-lived seed banks may need to be re-assessed using realistic densities either within seed bags or otherwise. Similarly, models of seed bank dynamics based on such overestimated depletion rates may lead to incorrect conclusions regarding the seed banks and, perhaps, the management of weeds and rare species.