884 resultados para Dunkl Kernel
Resumo:
Ce mémoire, composé d'un article en collaboration avec Monsieur Luc Vinet et Vincent X. Genest, est la suite du travail effectué sur les systèmes quantiques super-intégrables définis par des Hamiltoniens de type Dunkl. Plus particulièrement, ce mémoire vise l'analyse du problème de Coulomb-Dunkl dans le plan qui est une généralisation du système quantique de l'atome d'hydrogène impliquant des opérateurs de réflexion sur les variables x et y. Le modèle est défini par un potentiel en 1/r. Nous avons tout d'abord remarqué que l'Hamiltonien est séparable en coordonnées polaires et que les fonctions d'onde s'écrivent en termes de produits de polynômes de Laguerre généralisés et des harmoniques de Dunkl sur le cercle. L'algèbre générée par les opérateurs de symétrie nous a également permis de confirmer le caractère maximalement super-intégrable du problème de Coulomb-Dunkl. Nous avons aussi pu écrire explicitement les représentations de cette même algèbre. Nous avons finalement trouvé le spectre de l'énergie de manière algébrique.
Resumo:
An improved color video super-resolution technique using kernel regression and fuzzy enhancement is presented in this paper. A high resolution frame is computed from a set of low resolution video frames by kernel regression using an adaptive Gaussian kernel. A fuzzy smoothing filter is proposed to enhance the regression output. The proposed technique is a low cost software solution to resolution enhancement of color video in multimedia applications. The performance of the proposed technique is evaluated using several color videos and it is found to be better than other techniques in producing high quality high resolution color videos
Resumo:
Ricinodendron heudelotii (Baill.) Pierre ex Pax. kernel (njansang) commercialization has been promoted by the World Agroforestry Centre (ICRAF) in project villages in Cameroon with the aim to alleviate poverty for small-scale farmers. We evaluated to what extent development interventions improved the financial situation of households by comparing project and control households. The financial importance of njansang to household livelihoods between 2005 and 2010 was investigated through semi-structured questionnaires with retrospective questions, focus group discussions, interviews and wealth-ranking exercises. The importance of njansang increased strongly in the entire study region and the increase was significantly larger in project households. Moreover, absolute numbers of income from njansang commercialization as well as relative importance of njansang in total cash income, increased significantly more in project households (p < 0.05). Although the lower wealth class households could increase their income through njansang trade, the upper wealth class households benefited more from the projects' interventions. Group sales as conducted in project villages did not lead to significantly higher prices and should be reconsidered. Hence, promotion of njansang had a positive effect on total cash income and can still be improved. The corporative actors for njansang commercialization are encouraged to adapt their strategies to ensure that also the lower wealth class households benefit from the conducted project interventions. In this respect, frequent project monitoring and impact analysis are important tools to accomplish this adaptation.
Resumo:
We present distribution independent bounds on the generalization misclassification performance of a family of kernel classifiers with margin. Support Vector Machine classifiers (SVM) stem out of this class of machines. The bounds are derived through computations of the $V_gamma$ dimension of a family of loss functions where the SVM one belongs to. Bounds that use functions of margin distributions (i.e. functions of the slack variables of SVM) are derived.
Resumo:
This paper presents a new paradigm for signal reconstruction and superresolution, Correlation Kernel Analysis (CKA), that is based on the selection of a sparse set of bases from a large dictionary of class- specific basis functions. The basis functions that we use are the correlation functions of the class of signals we are analyzing. To choose the appropriate features from this large dictionary, we use Support Vector Machine (SVM) regression and compare this to traditional Principal Component Analysis (PCA) for the tasks of signal reconstruction, superresolution, and compression. The testbed we use in this paper is a set of images of pedestrians. This paper also presents results of experiments in which we use a dictionary of multiscale basis functions and then use Basis Pursuit De-Noising to obtain a sparse, multiscale approximation of a signal. The results are analyzed and we conclude that 1) when used with a sparse representation technique, the correlation function is an effective kernel for image reconstruction and superresolution, 2) for image compression, PCA and SVM have different tradeoffs, depending on the particular metric that is used to evaluate the results, 3) in sparse representation techniques, L_1 is not a good proxy for the true measure of sparsity, L_0, and 4) the L_epsilon norm may be a better error metric for image reconstruction and compression than the L_2 norm, though the exact psychophysical metric should take into account high order structure in images.
Resumo:
This paper presents a computation of the $V_gamma$ dimension for regression in bounded subspaces of Reproducing Kernel Hilbert Spaces (RKHS) for the Support Vector Machine (SVM) regression $epsilon$-insensitive loss function, and general $L_p$ loss functions. Finiteness of the RV_gamma$ dimension is shown, which also proves uniform convergence in probability for regression machines in RKHS subspaces that use the $L_epsilon$ or general $L_p$ loss functions. This paper presenta a novel proof of this result also for the case that a bias is added to the functions in the RKHS.
Resumo:
In a seminal paper, Aitchison and Lauder (1985) introduced classical kernel density estimation techniques in the context of compositional data analysis. Indeed, they gave two options for the choice of the kernel to be used in the kernel estimator. One of these kernels is based on the use the alr transformation on the simplex SD jointly with the normal distribution on RD-1. However, these authors themselves recognized that this method has some deficiencies. A method for overcoming these dificulties based on recent developments for compositional data analysis and multivariate kernel estimation theory, combining the ilr transformation with the use of the normal density with a full bandwidth matrix, was recently proposed in Martín-Fernández, Chacón and Mateu- Figueras (2006). Here we present an extensive simulation study that compares both methods in practice, thus exploring the finite-sample behaviour of both estimators
Resumo:
The antioxidant and tyrosinase inhibitory properties of extracts of mango seed kernel (Mangifera indica L.), which is normally discarded when the fruit is processed, were studied. Extracts contained phenolic components by a high antioxidant activity, which was assessed in homogeneous solution by the 2,2-diphenyt-1-picrylhydrazyl radical and 2,2'-azinobis (3-ethylbenzothialozinesulfonic acid) radical cation-scavenging assays and in an emulsion with the ferric thiocyanate test. The extracts also possessed tyrosinase inhibitory activity. Drying conditions and extraction solvent were varied, and optimum conditions for preparation of mango seed kernel extract were found to be sun-drying with ethanol extraction at room temperature. Refluxing in acidified ethanol gave an increase in yield and the obtained extract had the highest content of total phenolics, and also was the most effective antioxidant with the highest radical-scavenging, metal-chelating and tyrosinase inhibitory activity. The extracts did not cause acute irritation of rabbit skins. Our study for the first time reveals the high total phenol content, radical-scavenging, metal-chelating and tyrosinase inhibitory activities of the extract from mango seed kernel. This extract may be suitable for use in food, cosmetic, nutraceutical and pharmaceutical applications. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
Using the classical Parzen window (PW) estimate as the target function, the sparse kernel density estimator is constructed in a forward constrained regression manner. The leave-one-out (LOO) test score is used for kernel selection. The jackknife parameter estimator subject to positivity constraint check is used for the parameter estimation of a single parameter at each forward step. As such the proposed approach is simple to implement and the associated computational cost is very low. An illustrative example is employed to demonstrate that the proposed approach is effective in constructing sparse kernel density estimators with comparable accuracy to that of the classical Parzen window estimate.
Resumo:
A unified approach is proposed for sparse kernel data modelling that includes regression and classification as well as probability density function estimation. The orthogonal-least-squares forward selection method based on the leave-one-out test criteria is presented within this unified data-modelling framework to construct sparse kernel models that generalise well. Examples from regression, classification and density estimation applications are used to illustrate the effectiveness of this generic sparse kernel data modelling approach.
Resumo:
A novel sparse kernel density estimator is derived based on a regression approach, which selects a very small subset of significant kernels by means of the D-optimality experimental design criterion using an orthogonal forward selection procedure. The weights of the resulting sparse kernel model are calculated using the multiplicative nonnegative quadratic programming algorithm. The proposed method is computationally attractive, in comparison with many existing kernel density estimation algorithms. Our numerical results also show that the proposed method compares favourably with other existing methods, in terms of both test accuracy and model sparsity, for constructing kernel density estimates.
Resumo:
The note proposes an efficient nonlinear identification algorithm by combining a locally regularized orthogonal least squares (LROLS) model selection with a D-optimality experimental design. The proposed algorithm aims to achieve maximized model robustness and sparsity via two effective and complementary approaches. The LROLS method alone is capable of producing a very parsimonious model with excellent generalization performance. The D-optimality design criterion further enhances the model efficiency and robustness. An added advantage is that the user only needs to specify a weighting for the D-optimality cost in the combined model selecting criterion and the entire model construction procedure becomes automatic. The value of this weighting does not influence the model selection procedure critically and it can be chosen with ease from a wide range of values.
Resumo:
An automatic algorithm is derived for constructing kernel density estimates based on a regression approach that directly optimizes generalization capability. Computational efficiency of the density construction is ensured using an orthogonal forward regression, and the algorithm incrementally minimizes the leave-one-out test score. Local regularization is incorporated into the density construction process to further enforce sparsity. Examples are included to demonstrate the ability of the proposed algorithm to effectively construct a very sparse kernel density estimate with comparable accuracy to that of the full sample Parzen window density estimate.
Resumo:
This paper presents an efficient construction algorithm for obtaining sparse kernel density estimates based on a regression approach that directly optimizes model generalization capability. Computational efficiency of the density construction is ensured using an orthogonal forward regression, and the algorithm incrementally minimizes the leave-one-out test score. A local regularization method is incorporated naturally into the density construction process to further enforce sparsity. An additional advantage of the proposed algorithm is that it is fully automatic and the user is not required to specify any criterion to terminate the density construction procedure. This is in contrast to an existing state-of-art kernel density estimation method using the support vector machine (SVM), where the user is required to specify some critical algorithm parameter. Several examples are included to demonstrate the ability of the proposed algorithm to effectively construct a very sparse kernel density estimate with comparable accuracy to that of the full sample optimized Parzen window density estimate. Our experimental results also demonstrate that the proposed algorithm compares favorably with the SVM method, in terms of both test accuracy and sparsity, for constructing kernel density estimates.