973 resultados para matrix function approximation
Resumo:
The nucleon spectral function in nuclear matter fulfills an energy weighted sum rule. Comparing two different realistic potentials, these sum rules are studied for Greens functions that are derived self-consistently within the T matrix approximation at finite temperature.
Resumo:
Realistic nucleon-nucleon interactions induce correlations to the nuclear many-body system, which lead to a fragmentation of the single-particle strength over a wide range of energies and momenta. We address the question of how this fragmentation affects the thermodynamical properties of nuclear matter. In particular, we show that the entropy can be computed with the help of a spectral function, which can be evaluated in terms of the self-energy obtained in the self-consistent Green's function approach. Results for the density and temperature dependences of the entropy per particle for symmetric nuclear matter are presented and compared to the results of lowest order finite-temperature Brueckner-Hartree-Fock calculations. The effects of correlations on the calculated entropy are small, if the appropriate quasiparticle approximation is used. The results demonstrate the thermodynamical consistency of the self-consistent T-matrix approximation for the evaluation of the Green's functions.
Resumo:
The method of approximate approximations, introduced by Maz'ya [1], can also be used for the numerical solution of boundary integral equations. In this case, the matrix of the resulting algebraic system to compute an approximate source density depends only on the position of a finite number of boundary points and on the direction of the normal vector in these points (Boundary Point Method). We investigate this approach for the Stokes problem in the whole space and for the Stokes boundary value problem in a bounded convex domain G subset R^2, where the second part consists of three steps: In a first step the unknown potential density is replaced by a linear combination of exponentially decreasing basis functions concentrated near the boundary points. In a second step, integration over the boundary partial G is replaced by integration over the tangents at the boundary points such that even analytical expressions for the potential approximations can be obtained. In a third step, finally, the linear algebraic system is solved to determine an approximate density function and the resulting solution of the Stokes boundary value problem. Even not convergent the method leads to an efficient approximation of the form O(h^2) + epsilon, where epsilon can be chosen arbitrarily small.
Resumo:
In this paper we consider the problem of approximating a function belonging to some funtion space Φ by a linear comination of n translates of a given function G. Ussing a lemma by Jones (1990) and Barron (1991) we show that it is possible to define function spaces and functions G for which the rate of convergence to zero of the erro is 0(1/n) in any number of dimensions. The apparent avoidance of the "curse of dimensionality" is due to the fact that these function spaces are more and more constrained as the dimension increases. Examples include spaces of the Sobolev tpe, in which the number of weak derivatives is required to be larger than the number of dimensions. We give results both for approximation in the L2 norm and in the Lc norm. The interesting feature of these results is that, thanks to the constructive nature of Jones" and Barron"s lemma, an iterative procedure is defined that can achieve this rate.
Resumo:
Data on the vibrational energy levels and rotational constants of carbon suboxide for the low-wavenumber bending mode ν7 are reviewed, in the ground-state manifold, and in the ν2-, ν3-, ν4-, and ν2 + ν4-state manifolds. Following the procedure developed by Duckett, Mills, and Robiette [J. Mol. Spectrosc. 63, 249 (1976)] the data have been inverted to give the effective bending potential in ν7 for each of these five states. Values are obtained for various other parameters in the effective vibration-rotation Hamiltonian. The potential and rotational constants in ν2 + ν4 are given to a close approximation by linear extrapolation from the ground state through the ν2 and ν4 states.
Resumo:
The rovibration partition function of CH4 was calculated in the temperature range of 100-1000 K using well-converged energy levels that were calculated by vibrational-rotational configuration interaction using the Watson Hamiltonian for total angular momenta J=0-50 and the MULTIMODE computer program. The configuration state functions are products of ground-state occupied and virtual modals obtained using the vibrational self-consistent field method. The Gilbert and Jordan potential energy surface was used for the calculations. The resulting partition function was used to test the harmonic oscillator approximation and the separable-rotation approximation. The harmonic oscillator, rigid-rotator approximation is in error by a factor of 2.3 at 300 K, but we also propose a separable-rotation approximation that is accurate within 2% from 100 to 1000 K. (C) 2004 American Institute of Physics.
Resumo:
Epidemiological studies have suggested an inverse correlation between red wine consumption and the incidence of CVD. However, Champagne wine has not been fully investigated for its cardioprotective potential. In order to assess whether acute and moderate Champagne wine consumption is capable of modulating vascular function, we performed a randomised, placebo-controlled, cross-over intervention trial. We show that consumption of Champagne wine, but not a control matched for alcohol, carbohydrate and fruit-derived acid content, induced an acute change in endothelium-independent vasodilatation at 4 and 8 h post-consumption. Although both Champagne wine and the control also induced an increase in endothelium-dependent vascular reactivity at 4 h, there was no significant difference between the vascular effects induced by Champagne or the control at any time point. These effects were accompanied by an acute decrease in the concentration of matrix metalloproteinase (MMP-9), a significant decrease in plasma levels of oxidising species and an increase in urinary excretion of a number of phenolic metabolites. In particular, the mean total excretion of hippuric acid, protocatechuic acid and isoferulic acid were all significantly greater following the Champagne wine intervention compared with the control intervention. Our data suggest that a daily moderate consumption of Champagne wine may improve vascular performance via the delivery of phenolic constituents capable of improving NO bioavailability and reducing matrix metalloproteinase activity.
Resumo:
A tunable radial basis function (RBF) network model is proposed for nonlinear system identification using particle swarm optimisation (PSO). At each stage of orthogonal forward regression (OFR) model construction, PSO optimises one RBF unit's centre vector and diagonal covariance matrix by minimising the leave-one-out (LOO) mean square error (MSE). This PSO aided OFR automatically determines how many tunable RBF nodes are sufficient for modelling. Compared with the-state-of-the-art local regularisation assisted orthogonal least squares algorithm based on the LOO MSE criterion for constructing fixed-node RBF network models, the PSO tuned RBF model construction produces more parsimonious RBF models with better generalisation performance and is computationally more efficient.
Resumo:
In this paper we consider hybrid (fast stochastic approximation and deterministic refinement) algorithms for Matrix Inversion (MI) and Solving Systems of Linear Equations (SLAE). Monte Carlo methods are used for the stochastic approximation, since it is known that they are very efficient in finding a quick rough approximation of the element or a row of the inverse matrix or finding a component of the solution vector. We show how the stochastic approximation of the MI can be combined with a deterministic refinement procedure to obtain MI with the required precision and further solve the SLAE using MI. We employ a splitting A = D – C of a given non-singular matrix A, where D is a diagonal dominant matrix and matrix C is a diagonal matrix. In our algorithm for solving SLAE and MI different choices of D can be considered in order to control the norm of matrix T = D –1C, of the resulting SLAE and to minimize the number of the Markov Chains required to reach given precision. Further we run the algorithms on a mini-Grid and investigate their efficiency depending on the granularity. Corresponding experimental results are presented.
Nonlinear system identification using particle swarm optimisation tuned radial basis function models
Resumo:
A novel particle swarm optimisation (PSO) tuned radial basis function (RBF) network model is proposed for identification of non-linear systems. At each stage of orthogonal forward regression (OFR) model construction process, PSO is adopted to tune one RBF unit's centre vector and diagonal covariance matrix by minimising the leave-one-out (LOO) mean square error (MSE). This PSO aided OFR automatically determines how many tunable RBF nodes are sufficient for modelling. Compared with the-state-of-the-art local regularisation assisted orthogonal least squares algorithm based on the LOO MSE criterion for constructing fixed-node RBF network models, the PSO tuned RBF model construction produces more parsimonious RBF models with better generalisation performance and is often more efficient in model construction. The effectiveness of the proposed PSO aided OFR algorithm for constructing tunable node RBF models is demonstrated using three real data sets.
Resumo:
An orthogonal forward selection (OFS) algorithm based on leave-one-out (LOO) criteria is proposed for the construction of radial basis function (RBF) networks with tunable nodes. Each stage of the construction process determines an RBF node, namely, its center vector and diagonal covariance matrix, by minimizing the LOO statistics. For regression application, the LOO criterion is chosen to be the LOO mean-square error, while the LOO misclassification rate is adopted in two-class classification application. This OFS-LOO algorithm is computationally efficient, and it is capable of constructing parsimonious RBF networks that generalize well. Moreover, the proposed algorithm is fully automatic, and the user does not need to specify a termination criterion for the construction process. The effectiveness of the proposed RBF network construction procedure is demonstrated using examples taken from both regression and classification applications.
Resumo:
In this letter, a Box-Cox transformation-based radial basis function (RBF) neural network is introduced using the RBF neural network to represent the transformed system output. Initially a fixed and moderate sized RBF model base is derived based on a rank revealing orthogonal matrix triangularization (QR decomposition). Then a new fast identification algorithm is introduced using Gauss-Newton algorithm to derive the required Box-Cox transformation, based on a maximum likelihood estimator. The main contribution of this letter is to explore the special structure of the proposed RBF neural network for computational efficiency by utilizing the inverse of matrix block decomposition lemma. Finally, the Box-Cox transformation-based RBF neural network, with good generalization and sparsity, is identified based on the derived optimal Box-Cox transformation and a D-optimality-based orthogonal forward regression algorithm. The proposed algorithm and its efficacy are demonstrated with an illustrative example in comparison with support vector machine regression.
Resumo:
A modified radial basis function (RBF) neural network and its identification algorithm based on observational data with heterogeneous noise are introduced. The transformed system output of Box-Cox is represented by the RBF neural network. To identify the model from observational data, the singular value decomposition of the full regression matrix consisting of basis functions formed by system input data is initially carried out and a new fast identification method is then developed using Gauss-Newton algorithm to derive the required Box-Cox transformation, based on a maximum likelihood estimator (MLE) for a model base spanned by the largest eigenvectors. Finally, the Box-Cox transformation-based RBF neural network, with good generalisation and sparsity, is identified based on the derived optimal Box-Cox transformation and an orthogonal forward regression algorithm using a pseudo-PRESS statistic to select a sparse RBF model with good generalisation. The proposed algorithm and its efficacy are demonstrated with numerical examples.
Resumo:
An efficient model identification algorithm for a large class of linear-in-the-parameters models is introduced that simultaneously optimises the model approximation ability, sparsity and robustness. The derived model parameters in each forward regression step are initially estimated via the orthogonal least squares (OLS), followed by being tuned with a new gradient-descent learning algorithm based on the basis pursuit that minimises the l(1) norm of the parameter estimate vector. The model subset selection cost function includes a D-optimality design criterion that maximises the determinant of the design matrix of the subset to ensure model robustness and to enable the model selection procedure to automatically terminate at a sparse model. The proposed approach is based on the forward OLS algorithm using the modified Gram-Schmidt procedure. Both the parameter tuning procedure, based on basis pursuit, and the model selection criterion, based on the D-optimality that is effective in ensuring model robustness, are integrated with the forward regression. As a consequence the inherent computational efficiency associated with the conventional forward OLS approach is maintained in the proposed algorithm. Examples demonstrate the effectiveness of the new approach.
Resumo:
New construction algorithms for radial basis function (RBF) network modelling are introduced based on the A-optimality and D-optimality experimental design criteria respectively. We utilize new cost functions, based on experimental design criteria, for model selection that simultaneously optimizes model approximation, parameter variance (A-optimality) or model robustness (D-optimality). The proposed approaches are based on the forward orthogonal least-squares (OLS) algorithm, such that the new A-optimality- and D-optimality-based cost functions are constructed on the basis of an orthogonalization process that gains computational advantages and hence maintains the inherent computational efficiency associated with the conventional forward OLS approach. The proposed approach enhances the very popular forward OLS-algorithm-based RBF model construction method since the resultant RBF models are constructed in a manner that the system dynamics approximation capability, model adequacy and robustness are optimized simultaneously. The numerical examples provided show significant improvement based on the D-optimality design criterion, demonstrating that there is significant room for improvement in modelling via the popular RBF neural network.