9 resultados para Tikhonov regularization
em Cambridge University Engineering Department Publications Database
Restoration of images and 3D data to higher resolution by deconvolution with sparsity regularization
Restoration of images and 3D data to higher resolution by deconvolution with sparsity regularization
Resumo:
Image convolution is conventionally approximated by the LTI discrete model. It is well recognized that the higher the sampling rate, the better is the approximation. However sometimes images or 3D data are only available at a lower sampling rate due to physical constraints of the imaging system. In this paper, we model the under-sampled observation as the result of combining convolution and subsampling. Because the wavelet coefficients of piecewise smooth images tend to be sparse and well modelled by tree-like structures, we propose the L0 reweighted-L2 minimization (L0RL2 ) algorithm to solve this problem. This promotes model-based sparsity by minimizing the reweighted L2 norm, which approximates the L0 norm, and by enforcing a tree model over the weights. We test the algorithm on 3 examples: a simple ring, the cameraman image and a 3D microscope dataset; and show that good results can be obtained. © 2010 IEEE.
Resumo:
We consider a straight cylindrical duct with a steady subsonic axial flow and a reacting boundary (e.g. an acoustic lining). The wave modes are separated into ordinary acoustic duct modes, and surface modes confined to a small neighbourhood of the boundary. Many researchers have used a mass-spring-damper boundary model, for which one surface mode has previously been identified as a convective instability; however, we show the stability analysis used in such cases to be questionable. We investigate instead the stability of the surface modes using the Briggs-Bers criterion for a Flügge thin-shell boundary model. For modest frequencies and wavenumbers the thin-shell has an impedance which is effectively that of a mass-spring-damper, although for the large wavenumbers needed for the stability analysis the thin-shell and mass-spring-damper impedances diverge, owing to the thin shell's bending stiffness. The thin shell model may therefore be viewed as a regularization of the mass-spring-damper model which accounts for nonlocally-reacting effects. We find all modes to be stable for realistic thin-shell parameters, while absolute instabilities are demonstrated for extremely thin boundary thicknesses. The limit of vanishing bending stiffness is found to be a singular limit, yielding absolute instabilities of arbitrarily large temporal growth rate. We propose that the problems with previous stability analyses are due to the neglect of something akin to bending stiffness in the boundary model. Our conclusion is that the surface mode previously identified as a convective instability may well be stable in reality. Finally, inspired by Rienstra's recent analysis, we investigate the scattering of an acoustic mode as it encounters a sudden change from a hard-wall to a thin-shell boundary, using a Wiener-Hopf technique. The thin-shell is considered to be clamped to the hard-wall. The acoustic mode is found to scatter into transmitted and reflected acoustic modes, and surface modes strongly linked to the solid waves in the boundary, although no longitudinal or transverse waves within the boundary are excited. Examples are provided that demonstrate total transmission, total reflection, and a combination of the two. This thin-shell scattering problem is preferable to the mass-spring-damper scattering problem presented by Rienstra, since the thin-shell problem is fully determined and does not need to appeal to a Kutta-like condition or the inclusion of an instability in order to avoid a surface-streamline cusp at the boundary change.
Resumo:
It is shown that a new mixed nonlinear/eddy viscosity LES model reproduces profiles better than a number of competing nonlinear and mixed models for plane channel flow. The objective is an LES method that produces a fully resolved turbulent boundary layer and could be applied to a variety of aerospace problems that are currently studied with RANS, RANS-LES, or DES methods that lack a true turbulent boundary layer. There are two components to the new model. One an eddy viscosity based upon the advected subgrid scale energy and a relatively small coefficient. Second, filtered nonlinear terms based upon the Leray regularization. Coefficients for the eddy viscosity and nonlinear terms come from LES tests in decaying, isotropic turbulence. Using these coefficients, the velocity profile matches measurements data at Reτ ≈ 1000 exactly. Profiles of the components of kinetic energy have the same shape as in the experiment, but the magnitudes differ by about 25%. None of the competing LES gets the shape correct. This method does not require extra operations at the transition between the boundary layer and the interior flow.
Resumo:
The paper addresses the problem of low-rank trace norm minimization. We propose an algorithm that alternates between fixed-rank optimization and rank-one updates. The fixed-rank optimization is characterized by an efficient factorization that makes the trace norm differentiable in the search space and the computation of duality gap numerically tractable. The search space is nonlinear but is equipped with a Riemannian structure that leads to efficient computations. We present a second-order trust-region algorithm with a guaranteed quadratic rate of convergence. Overall, the proposed optimization scheme converges superlinearly to the global solution while maintaining complexity that is linear in the number of rows and columns of the matrix. To compute a set of solutions efficiently for a grid of regularization parameters we propose a predictor-corrector approach that outperforms the naive warm-restart approach on the fixed-rank quotient manifold. The performance of the proposed algorithm is illustrated on problems of low-rank matrix completion and multivariate linear regression. © 2013 Society for Industrial and Applied Mathematics.
Resumo:
Choosing appropriate architectures and regularization strategies of deep networks is crucial to good predictive performance. To shed light on this problem, we analyze the analogous problem of constructing useful priors on compositions of functions. Specifically, we study the deep Gaussian process, a type of infinitely-wide, deep neural network. We show that in standard architectures, the representational capacity of the network tends to capture fewer degrees of freedom as the number of layers increases, retaining only a single degree of freedom in the limit. We propose an alternate network architecture which does not suffer from this pathology. We also examine deep covariance functions, obtained by composing infinitely many feature transforms. Lastly, we characterize the class of models obtained by performing dropout on Gaussian processes.