Figure 2 - uploaded by George Papandreou
Content may be subject to copyright.
Image denoising example. Up: The noisy image and its denoised version at an automatically selected scale. Down: Model Error ME( t ) (ground truth) and Prediction Error PE + ( t ) (estimated by each of the two cross-validation algorithms). 

Image denoising example. Up: The noisy image and its denoised version at an automatically selected scale. Down: Model Error ME( t ) (ground truth) and Prediction Error PE + ( t ) (estimated by each of the two cross-validation algorithms). 

Source publication
Conference Paper
Full-text available
Scale-spaces induced by diffusion processes play an important role in many computer vision tasks. Automatically selecting the most appropriate scale for a particular problem is a central issue for the practical applicability of such scale-space techniques. This paper concentrates on automatic scale selection when nonlinear diffusion scale-spaces ar...

Context in source publication

Context 1
... [8] and the references therein). For some model selection problems, like choosing the regularization parameter of smoothing splines, leave-one-out cross-validation (where K = M and T i only contains ( x i , y i ) ) can be approxi- mated analytically, leading to fast computations [7]. However, adopting leave-one-out cross-validation in our case would require building nonlinear diffusion scale-spaces of M images of size ( M − 1) -pixels each, which clearly is un- acceptably expensive. Hence we have explored two other data resampling configurations for our problem. Fig. 1 de- picts for each of these two alternative strategies the pixels D − T i used to build the scale-space model (in black) and the pixels T i used to estimate its prediction error (in white) for the case i = 1 . The first of these configurations (called quadruple-cv from now on), which was also used by Nason in the context of wavelet shrinkage [11], creates K = 4 nonlinear diffusion scale-spaces, each built up using ≈ 1 / 4 of the noisy image data and used to estimate the prediction error of the model on the remaining | T i | ≈ 3 / 4 of the data. The second (called double-cv in the following) selects the members of T i in a chessboard-like fashion, utilizing half for building the scale-space and half for testing it, repeat- ing K = 2 times. In more detail, in the case of quadruple-cv configuration, by selecting four different values { (0 , 0) , (0 , 1) , (1 , 0) , (1 , 1) } for the shift vector ( s i , s j ) , we get four subsampled by a factor of two in each direction versions of the noisy image y . Each consists of roughly M/ 4 pixels with coordi- nates { (2 i + s i , 2 j + s j ) : 0 ≤ i < M x / 2 , 0 ≤ j < M y / 2 } . We then build the nonlinear diffusion scale-space of each of these subsampled images, appropriately scal- ing the diffusion PDE. For example, if Eq.(1) is utilized, we must enforce τ = τ / 4 (for the time-step), g ( · ) = g ( · / 2) and σ = σ/ 2 . We subsequently use Eq. (6) to get PE CV quadruple − cv ( t ) , computing each of the four terms { PE + ( f t − T k ) : 1 ≤ k ≤ 4 } as follows: We take the t = t/ 4 snapshot of the corresponding auxiliary scale-space, with dimensions M / 2 by M / 2 and interpolate from it the values at the remaining 3 M/ 4 pixels p i ∈ T k . A nonlinear, edge-preserving procedure should ideally be utilized for the interpolation, although, in practice, we have noticed little difference when using simple bi- linear interpolation. We then penalize (by means of L ) the difference between the interpolated value and the initial noisy value y i at the same pixel p i and average over the 3 M/ 4 pixels of T k to get PE + ( f t − T k ) . The procedure just described adds automatic scale selection to nonlinear diffusion-based denoising procedures at roughly double the computational cost of plain nonlinear diffusion, since 4 auxiliary scale-spaces of size M/ 4 pixels each need to be built in parallel with the standard scale-space. The concept in the case of double-cv resampling is similar and won’t be described in detail. The main difference is that the pixels that build each of the auxiliary scale-spaces are not located on a rectangular lattice any more (see Fig. 1, right). Therefore, it is convenient to first interpolate the values at the M/ 2 “white” pixels of T k from the values at the remaining M/ 2 “black” pixels of the noisy image y i and then build the auxiliary scale-spaces. Since these auxiliary scale-spaces are full-sized, the PDE (6) doesn’t need any rescaling. The overall cost of the procedure is three times the cost of the standard scale-space, since two auxiliary full- sized scale-spaces evolve in parallel with the main one. An example of image denoising with automatic scale selection by cross-validation techniques can be seen in Fig. 2. At the first row one can see the noisy image and its denoised version at scale t ∗ double − cv determined by the double- cv cross-validation algorithm. The corresponding plots de- picting the ME( t ) (ground truth) and PE + ( t ) (as estimated by the two cross-validation algorithms) can be seen at the second row. Notice that, as we discussed in Sec. 2, ME( t ) is smaller than PE + ( t ) . Nonetheless, both quantities attain their minimum at roughly the same scale (after about 8 iter- ations). Some further examples of automatic denoising, one of an MRI scan and one of an aerial image, utilizing cross- validation scale selection techniques, can be seen in Fig. 3. In order to systematically assess the performance of the proposed algorithms and compare them with existing techniques, we run a series of denoising experiments on a dataset of 39 natural grayscale images (the ko- dak, aerial and misc1 collections available from http: //www.cipr.rpi.edu/resource/stills/ ), cor- rupted by artificial noise so that the ground truth f ∗ would be available. Apart from the two cross-validation algorithms we have proposed ( double-cv and quadruple-cv ), we have also implemented the snr-based method of We- ickert (with ground-truth snr value), and the decorrelation method of Mrazek, in the sequel dubbed snr and dec , respectively. The nonlinear diffusion scale-space used in the experiments reported here was generated by the PDE of Eq. (1) with σ = 0 . 1 pixels. We employed the diffusivity function g ( r ) = 1 / [1 + ( r/λ ) 2 ] [12], with λ = 0 . 01 in all of the tests (the intensity values of the images were in [0 , 1] ) and the AOS numerical scheme [16]. Bilinear interpolation was utilized for upsampling. We experimented with three different noise types nt ∈ { gaussian , salt&pepper , speckle } . The degraded images were respectively generated by y i = f ∗ ( x i ) + i (gaussian), y i = (1 + i ) f ∗ ( x i ) (speckle), and y i = f ∗ ( x i ) , with probability 1 − p and y i = 0 or 1 with probability p/ 2 each (salt&pepper). In the gaussian and speckle cases, the i were i.i.d. sampled from N (0 , σ 2 ) . We conducted tests for varying noise levels nl ∈ { 0 . 05 , 0 . 1 , 0 . 15 , 0 . 2 , 0 . 3 , 0 . 4 } , where nl = σ in the case of gaussian or speckle noise and nl = p in the case of salt&pepper noise and for two different choices of the loss function L ∈ { L 1 , L 2 } . To evaluate the performance of the algorithms under consideration for each of the 2 · 3 · 6 = 36 combinations of ( L, nt, nl ) we run experiments on all 39 images in the database and averaged the performance results of each algorithm on the different images. We totally performed 36 · 39 = 1404 experiments. At each experiment and for each scale selection algorithm alg ∈ { double − cv , quadruple − cv , dec , snr } , we eval- uated two quantities, namely the relative increase in model error , defined as ME = ME( t ∗ ) and the rela- ME tive scale error , given by ∆ t ∗ t ∗ = | t ∗ alg t ∗ − t ∗ ME | . ME In Figs. 4 and 5 we present average benchmark scores for ∆(ME) ME and ∆ t ∗ t ∗ , respectively, acquired by the experimental process just described. In these figures one can see that the two algorithms based on cross-validation consistently out- perform the dec and snr methods. The robustness of both cross-validation algorithms, irrespectively of the noise type or the utilized loss function is particularly noteworthy. ∆(ME) In more detail, as far as is concerned (Fig. 4), ME double-cv and quadruple-cv give results that sometimes are one order of magnitude better than the results given by the snr and dec algorithms. This is particularly true in the ∆(ME) case of gaussian noise, where the of the two cross- ME validation algorithms is almost always less than 1% , ap- proaching sometimes 0 . 1% , practically meaning that the automatically selected scale t ∗ CV in these cases coincided with the ground truth optimal scale for almost every image in the dataset. The performance of double-cv and quadruple-cv in the two cases of non-gaussian noise is less impressive, although the error still remains under 10% . Among the other two algorithms, snr seems to perform better than dec . It is noteworthy in Fig. 4 that the error made by the snr algorithm consistently lies around 10% . The decorrelation ...

Similar publications

Preprint
Full-text available
Computing the marginal likelihood (also called the Bayesian model evidence) is an important task in Bayesian model selection, providing a principled quantitative way to compare models. The learned harmonic mean estimator solves the exploding variance problem of the original harmonic mean estimation of the marginal likelihood. The learned harmonic m...
Article
Full-text available
The large majority of the criteria for model selection are functions of the usual variance estimate for a regression model. The validity of the usual variance estimate depends on some assumptions, most critically the validity of the model being estimated. This is often violated in model selection contexts, where model search takes place over invali...
Conference Paper
Full-text available
Multiscale, i.e. scale-space image analysis is a powerful framework for many image processing tasks. A fundamental issue with such scale-space techniques is the automatic selection of the most salient scale for a particular application. This paper considers optimal scale selection when nonlinear diffusion and morphological scale-spaces are utilized...
Article
Full-text available
We present a new regression model for the determination of parton distribution functions (PDF) using techniques inspired from deep learning projects. In the context of the NNPDF methodology, we implement a new efficient computing framework based on graph generated models for PDF parametrization and gradient descent optimization. The best model conf...
Preprint
Full-text available
K-fold cross validation (CV) is a popular method for estimating the true performance of machine learning models, allowing model selection and parameter tuning. However, the very process of CV requires random partitioning of the data and so our performance estimates are in fact stochastic, with variability that can be substantial for natural languag...

Citations

... Since the anisotropic diffusion is an iterative process, the problem of choosing the optimal stopping time and preventing an over smoothed result is crucial. Therefore, various methods estimating this parameter were proposed [69][70][71]. For instance, Tsiotsios and Petrou, in their work [69], have been proposed a version of anisotropic diffusion filter where an automatic stopping criterion is used which takes into consideration the quality of the preserved edges. ...
Article
Full-text available
In image processing and computer vision, the denoising process is an important step before several processing tasks. This paper presents a new adaptive noise-reducing anisotropic diffusion (ANRAD) method to improve the image quality, which can be considered as a modified version of a speckle-reducing anisotropic diffusion (SRAD) filter. The SRAD works very well for monochrome images with speckle noise. However, in the case of images corrupted with other types of noise, it cannot provide optimal image quality due to the inaccurate noise model. The ANRAD method introduces an automatic RGB noise model estimator in a partial differential equation system similar to the SRAD diffusion, which estimates at each iteration an upper bound of the real noise level function by fitting a lower envelope to the standard deviations of pre-segment image variances. Compared to the conventional SRAD filter, the proposed filter has the advantage of being adapted to the color noise produced by today’s CCD digital camera. The simulation results show that the ANRAD filter can reduce the noise while preserving image edges and fine details very well. Also, it is favorably compared to the fast non-local means filter, showing an improvement in the quality of the restored image. A quantitative comparison measure is given by the parameters like the mean structural similarity index and the peak signal-to-noise ratio.
... Papandreou and Maragos [10] emphasized that the problem of choosing the optimal number of iterations of a non-linear diffusion filter could be treated as a statistical model selection problem. In their work, the number of iterations should be chosen to minimize the expected distance to the (generally unknown) noise-free image, under a given problem specific loss function. ...
... In their work, the number of iterations should be chosen to minimize the expected distance to the (generally unknown) noise-free image, under a given problem specific loss function. Papandreou and Maragos [10] also proposed two novel algorithms based on statistical cross-validation techniques that approximate this ideal optimal criterion. In their method, the image region is divided into training and test sets, and the number of iterations is selected by minimizing L 1 or L 2 error measurement on the test set. ...
Article
Full-text available
Although anisotropic diffusion filters have been used extensively and with great success in medical image denoising, one limitation of this iterative approach, when used on fully automatic medical image processing schemes, is that the quality of the resulting denoised image is highly dependent on the number of iterations of the algorithm. Using many iterations may excessively blur the edges of the anatomical structures, while a few may not be enough to remove the undesirable noise. In this work, a mathematical model is proposed to automatically determine the number of iterations of the robust anisotropic diffusion filter applied to the problem of denoising three common human brain magnetic resonance (MR) images (T1-weighted, T2-weighted and proton density). The model is determined off-line by means of the maximization of the mean structural similarity index, which is used in this work as metric for quantitative assessment of the resulting processed images obtained after each iteration of the algorithm. After determining the model parameters, the optimal number of iterations of the algorithm is easily determined without requiring any extra computation time. The proposed method was tested on 3D synthetic and clinical human brain MR images and the results of qualitative and quantitative evaluation have shown its effectiveness.
... There are many alternative regularization parameter-choice methods, see [18, 25]; this CC measure itself is not universally accepted within the image processing community [44]. However, an exhaustive comparison of these heuristics is beyond the scope of this paper. ...
Article
Anisotropic diffusion methods are well known for giving good qualitative results for image denoising. This paper gives a review of the anisotropic diffusion methodology and its application to image denoising. We propose a fixed-point iteration using a multigrid solver to solve a regularized anisotropic diffusion equation, which is not only well-posed, but also has a nontrivial steady-state solution. A new regularization parameter-choice method (Brent-NCP), combining Brent's method and the normalized cumulative periodogram information of the misfit, is also introduced. We test our algorithm on several common test images with different noise levels. The experimental results demonstrate the effectiveness of the anisotropic diffusion with a multigrid approach and the broad applicability of the Brent-NCP parameter-choice algorithm.
... The Perona-Malik equation [5] deals with the scale space and performs anisotropic diffusion for edgepreserving smoothing. It is observed in [4] that the proper selection of scale is instrumental in grouping relevant information, suppressing irrelevant noises while preserving the necessary image details. Because in general image features occur at multiple scales, the optimal scale of analysis should be different at each pixel in order to achieve the simulta-neous grouping of smooth features without over-smoothing fine details and the stopping of any smoothing at discontinuities. ...
Conference Paper
Full-text available
Abstract In this paper we propose a method,that simultaneously performs image denoising and compression by using mul- tiscale tensor voting. Given a real color image, the pixels are r st converted into a set of tokens to be grouped by ten- sor voting, where optimal scales are automatically selected among,others for perceptual grouping and faithful recon- struction. Tensor voting at multiple scales are performed at all input tokens to infer the feature grouping attributes such as region-ness, curve-ness, and junction-ness with their op- timal scales. We perform experiments on complex real im- ages to demonstrate the robustness of our method.
... Our approach is on the same spirit with [13] , where similar cross-validation techniques are utilized for selecting the optimal threshold for denoising by wavelet thresholding. This paper extends our recent work in [14], which was confined to scale selection in diffusion scale-spaces, by studying the performance of the proposed cross-validation scale selection algorithms in denoising schemes utilizing a wider class of nonlinear scale-spaces, including morphologically-induced ones. ...
Conference Paper
Full-text available
Multiscale, i.e. scale-space image analysis is a powerful framework for many image processing tasks. A fundamental issue with such scale-space techniques is the automatic selection of the most salient scale for a particular application. This paper considers optimal scale selection when nonlinear diffusion and morphological scale-spaces are utilized for image denoising. The problem is studied from a statistical model selection viewpoint and cross-validation techniques are utilized to address it in a principled way. The proposed novel algorithms do not require knowledge of the noise variance, have acceptable computational cost and are readily integrated with a wide class of scale-space inducing processes which require setting of a scale parameter. Our experiments show that this methodology leads to robust algorithms, which outperform existing scale-selection techniques for a wide range of noise types and noise levels.
Chapter
Noise often corrupts images during their acquisition, storage, and transmission processes, resulting in a degradation of image quality. The main challenge in image denoising is to remove the corrupted information while preserving the fine details of the original image. Anisotropic Diffusion (AD) is a well-established method for noise removal in digital images that effectively preserves the image edges. However, selecting the appropriate parameters for the AD operation significantly influences the filtering outcomes. Despite its importance, the automatic determination of AD parameters based on image requirements has received limited attention in the literature. This chapter introduces a novel multi-objective approach for obtaining optimal AD parameters that yield effective filtering results. The presented methodology aims to strike the best possible balance between two conflicting objectives: minimizing noise content in the image while maximizing contrast. Improving one objective without compromising the other is not feasible. To address this, the presented approach employs the Non-dominated Sorting Genetic Algorithm based on reference points (NSGA-III), which is known as a robust and powerful algorithm for multi-objective optimization. The final solution is determined by analyzing the Optimal Pareto front. Experimental results demonstrate that the presented method outperforms existing filtering algorithms in terms of both visual quality and standard performance metrics.
Article
Images are often corrupted by noise during their process of acquisition, storage and transmission. This alteration usually deteriorates the perception quality of the image. The central challenge in image denoising corresponds to eliminating the corrupted information while it is maintained the fine features of the original image. Anisotropic Diffusion (AD) is considered a well-established scheme for removing noise in digital images without deteriorating their edges. The selection of parameters that define the AD operation presents a very decisive implication in the filtering results. In spite of the importance of the AD operation, the problem of automatically calculating its parameter according to the image requirements has been scarcely considered in the literature. In this paper, a new multi-objective methodology to obtain the AD parameters for effective filter results is presented. In the proposed approach, the appropriate AD parameter values are obtained as the solution that involves the best possible trade-off between two conflicting scenarios: To minimize the noise content in the image while the contrast is maximized. Both objectives cannot be individually improved without worsening another. The proposed scheme uses to solve this formulation of the Non-dominated sorting genetic algorithm based on reference points (NSGA-III), which represent one of the most robust and powerful algorithms for multi-objective optimization. Then, the final solution is obtained from an analysis of the Optimal Pareto front. Experimental results demonstrate that the proposed method presents better results than existing filtering algorithms in terms of visual quality and standard performance metrics.
Chapter
In previous chapter, the weaknesses of conventional segmentation methods have been identified. This concludes the desired segmentation criteria in order to guide the mechanism of the proposed framework of segmentation. The segmentation is performed to partition the hand bone from its background and soft-tissue region in the beginning of this chapter. The challenges of hand bone segmentation are the overlapping intensity between the soft-tissue region and the spongy bone region within the hand bone. Three modules of techniques will be discussed and implemented to solve the problem in this chapter.
Article
Anisotropic diffusion filtering is highly dependent on some crucial parameters, such as the conductance function, the gradient threshold parameter and the stopping time of the iterative process. The various alternative options at each stage of the algorithm are examined and evaluated and the best choice is selected. An automatic stopping criterion is proposed, that takes into consideration the quality of the preserved edges as opposed to just the level of smoothing achieved. The proposed scheme is evaluated with the help of real and simulated images, and compared with other state of the art schemes using objective criteria.
Article
The selection of stopping time (i.e., scale) significantly affects the performance of anisotropic diffusion filter for image denoising. This paper designs a Markov random field (MRF) scale selection model, which selects scales for image segments, then the denoised image is the composition of segments at their optimal scales in the scale space. Firstly, statistics-based scale selection criteria are proposed for image segments. Then we design a scale selection energy function in the MRF framework by considering the scale coherence between neighboring segments. A segment-based noise estimation algorithm is also developed to estimate the noise statistics efficiently. Experiments show that the performance of MRF scale selection model is much better than the previous global scale selection schemes. Combined with this scale selection model, the anisotropic diffusion filter is comparable to or even outperform the state-of-the-art denoising methods in performance.