286 research outputs found
Curvelets and Ridgelets
International audienceDespite the fact that wavelets have had a wide impact in image processing, they fail to efficiently represent objects with highly anisotropic elements such as lines or curvilinear structures (e.g. edges). The reason is that wavelets are non-geometrical and do not exploit the regularity of the edge curve. The Ridgelet and the Curvelet [3, 4] transforms were developed as an answer to the weakness of the separable wavelet transform in sparsely representing what appears to be simple building atoms in an image, that is lines, curves and edges. Curvelets and ridgelets take the form of basis elements which exhibit high directional sensitivity and are highly anisotropic [5, 6, 7, 8]. These very recent geometric image representations are built upon ideas of multiscale analysis and geometry. They have had an important success in a wide range of image processing applications including denoising [8, 9, 10], deconvolution [11, 12], contrast enhancement [13], texture analysis [14, 15], detection [16], watermarking [17], component separation [18], inpainting [19, 20] or blind source separation[21, 22]. Curvelets have also proven useful in diverse fields beyond the traditional image processing application. Let’s cite for example seismic imaging [10, 23, 24], astronomical imaging [25, 26, 27], scientific computing and analysis of partial differential equations [28, 29]. Another reason for the success of ridgelets and curvelets is the availability of fast transform algorithms which are available in non-commercial software packages following the philosophy of reproducible research, see [30, 31]
Representation Learning via Cauchy Convolutional Sparse Coding
In representation learning, Convolutional Sparse Coding (CSC) enables
unsupervised learning of features by jointly optimising both an -norm
fidelity term and a sparsity enforcing penalty. This work investigates using a
regularisation term derived from an assumed Cauchy prior for the coefficients
of the feature maps of a CSC generative model. The sparsity penalty term
resulting from this prior is solved via its proximal operator, which is then
applied iteratively, element-wise, on the coefficients of the feature maps to
optimise the CSC cost function. The performance of the proposed Iterative
Cauchy Thresholding (ICT) algorithm in reconstructing natural images is
compared against the common choice of -norm optimised via soft and
hard thresholding. ICT outperforms IHT and IST in most of these reconstruction
experiments across various datasets, with an average PSNR of up to 11.30 and
7.04 above ISTA and IHT respectively.Comment: 19 pages, 9 figures, journal draf
Representation Learning via Cauchy Convolutional Sparse Coding
In representation learning, Convolutional Sparse Coding (CSC) enables
unsupervised learning of features by jointly optimising both an -norm
fidelity term and a sparsity enforcing penalty. This work investigates using a
regularisation term derived from an assumed Cauchy prior for the coefficients
of the feature maps of a CSC generative model. The sparsity penalty term
resulting from this prior is solved via its proximal operator, which is then
applied iteratively, element-wise, on the coefficients of the feature maps to
optimise the CSC cost function. The performance of the proposed Iterative
Cauchy Thresholding (ICT) algorithm in reconstructing natural images is
compared against the common choice of -norm optimised via soft and
hard thresholding. ICT outperforms IHT and IST in most of these reconstruction
experiments across various datasets, with an average PSNR of up to 11.30 and
7.04 above ISTA and IHT respectively.Comment: 19 pages, 9 figures, journal draf
- …