935 lines
37 KiB
Python
935 lines
37 KiB
Python
import scipy.stats
|
||
import numpy as np
|
||
from math import ceil
|
||
from .. import img_as_float
|
||
from ._denoise_cy import _denoise_bilateral, _denoise_tv_bregman
|
||
from .._shared.utils import warn
|
||
import pywt
|
||
import skimage.color as color
|
||
from skimage.color.colorconv import ycbcr_from_rgb
|
||
import numbers
|
||
|
||
|
||
def _gaussian_weight(array, sigma_squared, *, dtype=float):
|
||
"""Helping function. Define a Gaussian weighting from array and
|
||
sigma_square.
|
||
|
||
Parameters
|
||
----------
|
||
array : ndarray
|
||
Input array.
|
||
sigma_squared : float
|
||
The squared standard deviation used in the filter.
|
||
dtype : data type object, optional (default : float)
|
||
The type and size of the data to be returned.
|
||
|
||
Returns
|
||
-------
|
||
gaussian : ndarray
|
||
The input array filtered by the Gaussian.
|
||
"""
|
||
return np.exp(-0.5 * (array ** 2 / sigma_squared), dtype=dtype)
|
||
|
||
|
||
def _compute_color_lut(bins, sigma, max_value, *, dtype=float):
|
||
"""Helping function. Define a lookup table containing Gaussian filter
|
||
values using the color distance sigma.
|
||
|
||
Parameters
|
||
----------
|
||
bins : int
|
||
Number of discrete values for Gaussian weights of color filtering.
|
||
A larger value results in improved accuracy.
|
||
sigma : float
|
||
Standard deviation for grayvalue/color distance (radiometric
|
||
similarity). A larger value results in averaging of pixels with larger
|
||
radiometric differences. Note, that the image will be converted using
|
||
the `img_as_float` function and thus the standard deviation is in
|
||
respect to the range ``[0, 1]``. If the value is ``None`` the standard
|
||
deviation of the ``image`` will be used.
|
||
max_value : float
|
||
Maximum value of the input image.
|
||
dtype : data type object, optional (default : float)
|
||
The type and size of the data to be returned.
|
||
|
||
Returns
|
||
-------
|
||
color_lut : ndarray
|
||
Lookup table for the color distance sigma.
|
||
"""
|
||
values = np.linspace(0, max_value, bins, endpoint=False)
|
||
return _gaussian_weight(values, sigma**2, dtype=dtype)
|
||
|
||
|
||
def _compute_spatial_lut(win_size, sigma, *, dtype=float):
|
||
"""Helping function. Define a lookup table containing Gaussian filter
|
||
values using the spatial sigma.
|
||
|
||
Parameters
|
||
----------
|
||
win_size : int
|
||
Window size for filtering.
|
||
If win_size is not specified, it is calculated as
|
||
``max(5, 2 * ceil(3 * sigma_spatial) + 1)``.
|
||
sigma : float
|
||
Standard deviation for range distance. A larger value results in
|
||
averaging of pixels with larger spatial differences.
|
||
dtype : data type object
|
||
The type and size of the data to be returned.
|
||
|
||
Returns
|
||
-------
|
||
spatial_lut : ndarray
|
||
Lookup table for the spatial sigma.
|
||
"""
|
||
grid_points = np.arange(-win_size // 2, win_size // 2 + 1)
|
||
rr, cc = np.meshgrid(grid_points, grid_points, indexing='ij')
|
||
distances = np.hypot(rr, cc)
|
||
return _gaussian_weight(distances, sigma**2, dtype=dtype).ravel()
|
||
|
||
|
||
def denoise_bilateral(image, win_size=None, sigma_color=None, sigma_spatial=1,
|
||
bins=10000, mode='constant', cval=0, multichannel=False):
|
||
"""Denoise image using bilateral filter.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray, shape (M, N[, 3])
|
||
Input image, 2D grayscale or RGB.
|
||
win_size : int
|
||
Window size for filtering.
|
||
If win_size is not specified, it is calculated as
|
||
``max(5, 2 * ceil(3 * sigma_spatial) + 1)``.
|
||
sigma_color : float
|
||
Standard deviation for grayvalue/color distance (radiometric
|
||
similarity). A larger value results in averaging of pixels with larger
|
||
radiometric differences. Note, that the image will be converted using
|
||
the `img_as_float` function and thus the standard deviation is in
|
||
respect to the range ``[0, 1]``. If the value is ``None`` the standard
|
||
deviation of the ``image`` will be used.
|
||
sigma_spatial : float
|
||
Standard deviation for range distance. A larger value results in
|
||
averaging of pixels with larger spatial differences.
|
||
bins : int
|
||
Number of discrete values for Gaussian weights of color filtering.
|
||
A larger value results in improved accuracy.
|
||
mode : {'constant', 'edge', 'symmetric', 'reflect', 'wrap'}
|
||
How to handle values outside the image borders. See
|
||
`numpy.pad` for detail.
|
||
cval : string
|
||
Used in conjunction with mode 'constant', the value outside
|
||
the image boundaries.
|
||
multichannel : bool
|
||
Whether the last axis of the image is to be interpreted as multiple
|
||
channels or another spatial dimension.
|
||
|
||
Returns
|
||
-------
|
||
denoised : ndarray
|
||
Denoised image.
|
||
|
||
Notes
|
||
-----
|
||
This is an edge-preserving, denoising filter. It averages pixels based on
|
||
their spatial closeness and radiometric similarity [1]_.
|
||
|
||
Spatial closeness is measured by the Gaussian function of the Euclidean
|
||
distance between two pixels and a certain standard deviation
|
||
(`sigma_spatial`).
|
||
|
||
Radiometric similarity is measured by the Gaussian function of the
|
||
Euclidean distance between two color values and a certain standard
|
||
deviation (`sigma_color`).
|
||
|
||
References
|
||
----------
|
||
.. [1] C. Tomasi and R. Manduchi. "Bilateral Filtering for Gray and Color
|
||
Images." IEEE International Conference on Computer Vision (1998)
|
||
839-846. :DOI:`10.1109/ICCV.1998.710815`
|
||
|
||
Examples
|
||
--------
|
||
>>> from skimage import data, img_as_float
|
||
>>> astro = img_as_float(data.astronaut())
|
||
>>> astro = astro[220:300, 220:320]
|
||
>>> noisy = astro + 0.6 * astro.std() * np.random.random(astro.shape)
|
||
>>> noisy = np.clip(noisy, 0, 1)
|
||
>>> denoised = denoise_bilateral(noisy, sigma_color=0.05, sigma_spatial=15,
|
||
... multichannel=True)
|
||
"""
|
||
if multichannel:
|
||
if image.ndim != 3:
|
||
if image.ndim == 2:
|
||
raise ValueError("Use ``multichannel=False`` for 2D grayscale "
|
||
"images. The last axis of the input image "
|
||
"must be multiple color channels not another "
|
||
"spatial dimension.")
|
||
else:
|
||
raise ValueError("Bilateral filter is only implemented for "
|
||
"2D grayscale images (image.ndim == 2) and "
|
||
"2D multichannel (image.ndim == 3) images, "
|
||
"but the input image has {0} dimensions. "
|
||
"".format(image.ndim))
|
||
elif image.shape[2] not in (3, 4):
|
||
if image.shape[2] > 4:
|
||
msg = ("The last axis of the input image is interpreted as "
|
||
"channels. Input image with shape {0} has {1} channels "
|
||
"in last axis. ``denoise_bilateral`` is implemented "
|
||
"for 2D grayscale and color images only")
|
||
warn(msg.format(image.shape, image.shape[2]))
|
||
else:
|
||
msg = "Input image must be grayscale, RGB, or RGBA; " \
|
||
"but has shape {0}."
|
||
warn(msg.format(image.shape))
|
||
else:
|
||
if image.ndim > 2:
|
||
raise ValueError("Bilateral filter is not implemented for "
|
||
"grayscale images of 3 or more dimensions, "
|
||
"but input image has {0} dimension. Use "
|
||
"``multichannel=True`` for 2-D RGB "
|
||
"images.".format(image.shape))
|
||
|
||
if win_size is None:
|
||
win_size = max(5, 2 * int(ceil(3 * sigma_spatial)) + 1)
|
||
|
||
min_value = image.min()
|
||
max_value = image.max()
|
||
|
||
if min_value == max_value:
|
||
return image
|
||
|
||
# if image.max() is 0, then dist_scale can have an unverified value
|
||
# and color_lut[<int>(dist * dist_scale)] may cause a segmentation fault
|
||
# so we verify we have a positive image and that the max is not 0.0.
|
||
if min_value < 0.0:
|
||
raise ValueError("Image must contain only positive values")
|
||
|
||
if max_value == 0.0:
|
||
raise ValueError("The maximum value found in the image was 0.")
|
||
|
||
image = np.atleast_3d(img_as_float(image))
|
||
image = np.ascontiguousarray(image)
|
||
|
||
sigma_color = sigma_color or image.std()
|
||
|
||
color_lut = _compute_color_lut(bins, sigma_color, max_value,
|
||
dtype=image.dtype)
|
||
|
||
range_lut = _compute_spatial_lut(win_size, sigma_spatial, dtype=image.dtype)
|
||
|
||
out = np.empty(image.shape, dtype=image.dtype)
|
||
|
||
dims = image.shape[2]
|
||
|
||
# There are a number of arrays needed in the Cython function.
|
||
# It's easier to allocate them outside of Cython so that all
|
||
# arrays are in the same type, then just copy the empty array
|
||
# where needed within Cython.
|
||
empty_dims = np.empty(dims, dtype=image.dtype)
|
||
|
||
return _denoise_bilateral(image, image.max(), win_size, sigma_color,
|
||
sigma_spatial, bins, mode, cval, color_lut,
|
||
range_lut, empty_dims, out)
|
||
|
||
|
||
def denoise_tv_bregman(image, weight, max_iter=100, eps=1e-3, isotropic=True,
|
||
*, multichannel=False):
|
||
"""Perform total-variation denoising using split-Bregman optimization.
|
||
|
||
Total-variation denoising (also know as total-variation regularization)
|
||
tries to find an image with less total-variation under the constraint
|
||
of being similar to the input image, which is controlled by the
|
||
regularization parameter ([1]_, [2]_, [3]_, [4]_).
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray
|
||
Input data to be denoised (converted using img_as_float`).
|
||
weight : float
|
||
Denoising weight. The smaller the `weight`, the more denoising (at
|
||
the expense of less similarity to the `input`). The regularization
|
||
parameter `lambda` is chosen as `2 * weight`.
|
||
eps : float, optional
|
||
Relative difference of the value of the cost function that determines
|
||
the stop criterion. The algorithm stops when::
|
||
|
||
SUM((u(n) - u(n-1))**2) < eps
|
||
|
||
max_iter : int, optional
|
||
Maximal number of iterations used for the optimization.
|
||
isotropic : boolean, optional
|
||
Switch between isotropic and anisotropic TV denoising.
|
||
multichannel : bool, optional
|
||
Apply total-variation denoising separately for each channel. This
|
||
option should be true for color images, otherwise the denoising is
|
||
also applied in the channels dimension.
|
||
|
||
Returns
|
||
-------
|
||
u : ndarray
|
||
Denoised image.
|
||
|
||
References
|
||
----------
|
||
.. [1] https://en.wikipedia.org/wiki/Total_variation_denoising
|
||
.. [2] Tom Goldstein and Stanley Osher, "The Split Bregman Method For L1
|
||
Regularized Problems",
|
||
ftp://ftp.math.ucla.edu/pub/camreport/cam08-29.pdf
|
||
.. [3] Pascal Getreuer, "Rudin–Osher–Fatemi Total Variation Denoising
|
||
using Split Bregman" in Image Processing On Line on 2012–05–19,
|
||
https://www.ipol.im/pub/art/2012/g-tvd/article_lr.pdf
|
||
.. [4] https://web.math.ucsb.edu/~cgarcia/UGProjects/BregmanAlgorithms_JacquelineBush.pdf
|
||
|
||
"""
|
||
image = np.atleast_3d(img_as_float(image))
|
||
|
||
rows = image.shape[0]
|
||
cols = image.shape[1]
|
||
dims = image.shape[2]
|
||
|
||
shape_ext = (rows + 2, cols + 2, dims)
|
||
|
||
out = np.zeros(shape_ext, image.dtype)
|
||
|
||
if multichannel:
|
||
channel_out = np.zeros(shape_ext[:2] + (1,), dtype=out.dtype)
|
||
for c in range(image.shape[-1]):
|
||
# the algorithm below expects 3 dimensions to always be present.
|
||
# slicing the array in this fashion preserves the channel dimension for us
|
||
channel_in = np.ascontiguousarray(image[..., c:c+1])
|
||
|
||
_denoise_tv_bregman(channel_in, image.dtype.type(weight),
|
||
max_iter, eps, isotropic, channel_out)
|
||
|
||
out[..., c] = channel_out[..., 0]
|
||
|
||
else:
|
||
image = np.ascontiguousarray(image)
|
||
|
||
_denoise_tv_bregman(image, image.dtype.type(weight), max_iter, eps,
|
||
isotropic, out)
|
||
|
||
return np.squeeze(out[1:-1, 1:-1])
|
||
|
||
|
||
def _denoise_tv_chambolle_nd(image, weight=0.1, eps=2.e-4, n_iter_max=200):
|
||
"""Perform total-variation denoising on n-dimensional images.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray
|
||
n-D input data to be denoised.
|
||
weight : float, optional
|
||
Denoising weight. The greater `weight`, the more denoising (at
|
||
the expense of fidelity to `input`).
|
||
eps : float, optional
|
||
Relative difference of the value of the cost function that determines
|
||
the stop criterion. The algorithm stops when:
|
||
|
||
(E_(n-1) - E_n) < eps * E_0
|
||
|
||
n_iter_max : int, optional
|
||
Maximal number of iterations used for the optimization.
|
||
|
||
Returns
|
||
-------
|
||
out : ndarray
|
||
Denoised array of floats.
|
||
|
||
Notes
|
||
-----
|
||
Rudin, Osher and Fatemi algorithm.
|
||
"""
|
||
|
||
ndim = image.ndim
|
||
p = np.zeros((image.ndim, ) + image.shape, dtype=image.dtype)
|
||
g = np.zeros_like(p)
|
||
d = np.zeros_like(image)
|
||
i = 0
|
||
while i < n_iter_max:
|
||
if i > 0:
|
||
# d will be the (negative) divergence of p
|
||
d = -p.sum(0)
|
||
slices_d = [slice(None), ] * ndim
|
||
slices_p = [slice(None), ] * (ndim + 1)
|
||
for ax in range(ndim):
|
||
slices_d[ax] = slice(1, None)
|
||
slices_p[ax+1] = slice(0, -1)
|
||
slices_p[0] = ax
|
||
d[tuple(slices_d)] += p[tuple(slices_p)]
|
||
slices_d[ax] = slice(None)
|
||
slices_p[ax+1] = slice(None)
|
||
out = image + d
|
||
else:
|
||
out = image
|
||
E = (d ** 2).sum()
|
||
|
||
# g stores the gradients of out along each axis
|
||
# e.g. g[0] is the first order finite difference along axis 0
|
||
slices_g = [slice(None), ] * (ndim + 1)
|
||
for ax in range(ndim):
|
||
slices_g[ax+1] = slice(0, -1)
|
||
slices_g[0] = ax
|
||
g[tuple(slices_g)] = np.diff(out, axis=ax)
|
||
slices_g[ax+1] = slice(None)
|
||
|
||
norm = np.sqrt((g ** 2).sum(axis=0))[np.newaxis, ...]
|
||
E += weight * norm.sum()
|
||
tau = 1. / (2.*ndim)
|
||
norm *= tau / weight
|
||
norm += 1.
|
||
p -= tau * g
|
||
p /= norm
|
||
E /= float(image.size)
|
||
if i == 0:
|
||
E_init = E
|
||
E_previous = E
|
||
else:
|
||
if np.abs(E_previous - E) < eps * E_init:
|
||
break
|
||
else:
|
||
E_previous = E
|
||
i += 1
|
||
return out
|
||
|
||
|
||
def denoise_tv_chambolle(image, weight=0.1, eps=2.e-4, n_iter_max=200,
|
||
multichannel=False):
|
||
"""Perform total-variation denoising on n-dimensional images.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray of ints, uints or floats
|
||
Input data to be denoised. `image` can be of any numeric type,
|
||
but it is cast into an ndarray of floats for the computation
|
||
of the denoised image.
|
||
weight : float, optional
|
||
Denoising weight. The greater `weight`, the more denoising (at
|
||
the expense of fidelity to `input`).
|
||
eps : float, optional
|
||
Relative difference of the value of the cost function that
|
||
determines the stop criterion. The algorithm stops when:
|
||
|
||
(E_(n-1) - E_n) < eps * E_0
|
||
|
||
n_iter_max : int, optional
|
||
Maximal number of iterations used for the optimization.
|
||
multichannel : bool, optional
|
||
Apply total-variation denoising separately for each channel. This
|
||
option should be true for color images, otherwise the denoising is
|
||
also applied in the channels dimension.
|
||
|
||
Returns
|
||
-------
|
||
out : ndarray
|
||
Denoised image.
|
||
|
||
Notes
|
||
-----
|
||
Make sure to set the multichannel parameter appropriately for color images.
|
||
|
||
The principle of total variation denoising is explained in
|
||
https://en.wikipedia.org/wiki/Total_variation_denoising
|
||
|
||
The principle of total variation denoising is to minimize the
|
||
total variation of the image, which can be roughly described as
|
||
the integral of the norm of the image gradient. Total variation
|
||
denoising tends to produce "cartoon-like" images, that is,
|
||
piecewise-constant images.
|
||
|
||
This code is an implementation of the algorithm of Rudin, Fatemi and Osher
|
||
that was proposed by Chambolle in [1]_.
|
||
|
||
References
|
||
----------
|
||
.. [1] A. Chambolle, An algorithm for total variation minimization and
|
||
applications, Journal of Mathematical Imaging and Vision,
|
||
Springer, 2004, 20, 89-97.
|
||
|
||
Examples
|
||
--------
|
||
2D example on astronaut image:
|
||
|
||
>>> from skimage import color, data
|
||
>>> img = color.rgb2gray(data.astronaut())[:50, :50]
|
||
>>> img += 0.5 * img.std() * np.random.randn(*img.shape)
|
||
>>> denoised_img = denoise_tv_chambolle(img, weight=60)
|
||
|
||
3D example on synthetic data:
|
||
|
||
>>> x, y, z = np.ogrid[0:20, 0:20, 0:20]
|
||
>>> mask = (x - 22)**2 + (y - 20)**2 + (z - 17)**2 < 8**2
|
||
>>> mask = mask.astype(np.float)
|
||
>>> mask += 0.2*np.random.randn(*mask.shape)
|
||
>>> res = denoise_tv_chambolle(mask, weight=100)
|
||
|
||
"""
|
||
|
||
im_type = image.dtype
|
||
if not im_type.kind == 'f':
|
||
image = img_as_float(image)
|
||
|
||
if multichannel:
|
||
out = np.zeros_like(image)
|
||
for c in range(image.shape[-1]):
|
||
out[..., c] = _denoise_tv_chambolle_nd(image[..., c], weight, eps,
|
||
n_iter_max)
|
||
else:
|
||
out = _denoise_tv_chambolle_nd(image, weight, eps, n_iter_max)
|
||
return out
|
||
|
||
|
||
def _bayes_thresh(details, var):
|
||
"""BayesShrink threshold for a zero-mean details coeff array."""
|
||
# Equivalent to: dvar = np.var(details) for 0-mean details array
|
||
dvar = np.mean(details*details)
|
||
eps = np.finfo(details.dtype).eps
|
||
thresh = var / np.sqrt(max(dvar - var, eps))
|
||
return thresh
|
||
|
||
|
||
def _universal_thresh(img, sigma):
|
||
""" Universal threshold used by the VisuShrink method """
|
||
return sigma*np.sqrt(2*np.log(img.size))
|
||
|
||
|
||
def _sigma_est_dwt(detail_coeffs, distribution='Gaussian'):
|
||
"""Calculate the robust median estimator of the noise standard deviation.
|
||
|
||
Parameters
|
||
----------
|
||
detail_coeffs : ndarray
|
||
The detail coefficients corresponding to the discrete wavelet
|
||
transform of an image.
|
||
distribution : str
|
||
The underlying noise distribution.
|
||
|
||
Returns
|
||
-------
|
||
sigma : float
|
||
The estimated noise standard deviation (see section 4.2 of [1]_).
|
||
|
||
References
|
||
----------
|
||
.. [1] D. L. Donoho and I. M. Johnstone. "Ideal spatial adaptation
|
||
by wavelet shrinkage." Biometrika 81.3 (1994): 425-455.
|
||
:DOI:`10.1093/biomet/81.3.425`
|
||
"""
|
||
# Consider regions with detail coefficients exactly zero to be masked out
|
||
detail_coeffs = detail_coeffs[np.nonzero(detail_coeffs)]
|
||
|
||
if distribution.lower() == 'gaussian':
|
||
# 75th quantile of the underlying, symmetric noise distribution
|
||
denom = scipy.stats.norm.ppf(0.75)
|
||
sigma = np.median(np.abs(detail_coeffs)) / denom
|
||
else:
|
||
raise ValueError("Only Gaussian noise estimation is currently "
|
||
"supported")
|
||
return sigma
|
||
|
||
|
||
def _wavelet_threshold(image, wavelet, method=None, threshold=None,
|
||
sigma=None, mode='soft', wavelet_levels=None):
|
||
"""Perform wavelet thresholding.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray (2d or 3d) of ints, uints or floats
|
||
Input data to be denoised. `image` can be of any numeric type,
|
||
but it is cast into an ndarray of floats for the computation
|
||
of the denoised image.
|
||
wavelet : string
|
||
The type of wavelet to perform. Can be any of the options
|
||
pywt.wavelist outputs. For example, this may be any of ``{db1, db2,
|
||
db3, db4, haar}``.
|
||
method : {'BayesShrink', 'VisuShrink'}, optional
|
||
Thresholding method to be used. The currently supported methods are
|
||
"BayesShrink" [1]_ and "VisuShrink" [2]_. If it is set to None, a
|
||
user-specified ``threshold`` must be supplied instead.
|
||
threshold : float, optional
|
||
The thresholding value to apply during wavelet coefficient
|
||
thresholding. The default value (None) uses the selected ``method`` to
|
||
estimate appropriate threshold(s) for noise removal.
|
||
sigma : float, optional
|
||
The standard deviation of the noise. The noise is estimated when sigma
|
||
is None (the default) by the method in [2]_.
|
||
mode : {'soft', 'hard'}, optional
|
||
An optional argument to choose the type of denoising performed. It
|
||
noted that choosing soft thresholding given additive noise finds the
|
||
best approximation of the original image.
|
||
wavelet_levels : int or None, optional
|
||
The number of wavelet decomposition levels to use. The default is
|
||
three less than the maximum number of possible decomposition levels
|
||
(see Notes below).
|
||
|
||
Returns
|
||
-------
|
||
out : ndarray
|
||
Denoised image.
|
||
|
||
References
|
||
----------
|
||
.. [1] Chang, S. Grace, Bin Yu, and Martin Vetterli. "Adaptive wavelet
|
||
thresholding for image denoising and compression." Image Processing,
|
||
IEEE Transactions on 9.9 (2000): 1532-1546.
|
||
:DOI:`10.1109/83.862633`
|
||
.. [2] D. L. Donoho and I. M. Johnstone. "Ideal spatial adaptation
|
||
by wavelet shrinkage." Biometrika 81.3 (1994): 425-455.
|
||
:DOI:`10.1093/biomet/81.3.425`
|
||
"""
|
||
wavelet = pywt.Wavelet(wavelet)
|
||
if not wavelet.orthogonal:
|
||
warn(("Wavelet thresholding was designed for use with orthogonal "
|
||
"wavelets. For nonorthogonal wavelets such as {}, results are "
|
||
"likely to be suboptimal.").format(wavelet.name))
|
||
|
||
# original_extent is used to workaround PyWavelets issue #80
|
||
# odd-sized input results in an image with 1 extra sample after waverecn
|
||
original_extent = tuple(slice(s) for s in image.shape)
|
||
|
||
# Determine the number of wavelet decomposition levels
|
||
if wavelet_levels is None:
|
||
# Determine the maximum number of possible levels for image
|
||
dlen = wavelet.dec_len
|
||
wavelet_levels = pywt.dwtn_max_level(image.shape, wavelet)
|
||
|
||
# Skip coarsest wavelet scales (see Notes in docstring).
|
||
wavelet_levels = max(wavelet_levels - 3, 1)
|
||
|
||
coeffs = pywt.wavedecn(image, wavelet=wavelet, level=wavelet_levels)
|
||
# Detail coefficients at each decomposition level
|
||
dcoeffs = coeffs[1:]
|
||
|
||
if sigma is None:
|
||
# Estimate the noise via the method in [2]_
|
||
detail_coeffs = dcoeffs[-1]['d' * image.ndim]
|
||
sigma = _sigma_est_dwt(detail_coeffs, distribution='Gaussian')
|
||
|
||
if method is not None and threshold is not None:
|
||
warn(("Thresholding method {} selected. The user-specified threshold "
|
||
"will be ignored.").format(method))
|
||
|
||
if threshold is None:
|
||
var = sigma**2
|
||
if method is None:
|
||
raise ValueError(
|
||
"If method is None, a threshold must be provided.")
|
||
elif method == "BayesShrink":
|
||
# The BayesShrink thresholds from [1]_ in docstring
|
||
threshold = [{key: _bayes_thresh(level[key], var) for key in level}
|
||
for level in dcoeffs]
|
||
elif method == "VisuShrink":
|
||
# The VisuShrink thresholds from [2]_ in docstring
|
||
threshold = _universal_thresh(image, sigma)
|
||
else:
|
||
raise ValueError("Unrecognized method: {}".format(method))
|
||
|
||
if np.isscalar(threshold):
|
||
# A single threshold for all coefficient arrays
|
||
denoised_detail = [{key: pywt.threshold(level[key],
|
||
value=threshold,
|
||
mode=mode) for key in level}
|
||
for level in dcoeffs]
|
||
else:
|
||
# Dict of unique threshold coefficients for each detail coeff. array
|
||
denoised_detail = [{key: pywt.threshold(level[key],
|
||
value=thresh[key],
|
||
mode=mode) for key in level}
|
||
for thresh, level in zip(threshold, dcoeffs)]
|
||
denoised_coeffs = [coeffs[0]] + denoised_detail
|
||
return pywt.waverecn(denoised_coeffs, wavelet)[original_extent]
|
||
|
||
|
||
def _scale_sigma_and_image_consistently(image, sigma, multichannel,
|
||
rescale_sigma):
|
||
"""If the ``image`` is rescaled, also rescale ``sigma`` consistently.
|
||
|
||
Images that are not floating point will be rescaled via ``img_as_float``.
|
||
"""
|
||
if multichannel:
|
||
if isinstance(sigma, numbers.Number) or sigma is None:
|
||
sigma = [sigma] * image.shape[-1]
|
||
elif len(sigma) != image.shape[-1]:
|
||
raise ValueError(
|
||
"When multichannel is True, sigma must be a scalar or have "
|
||
"length equal to the number of channels")
|
||
if image.dtype.kind != 'f':
|
||
if rescale_sigma:
|
||
range_pre = image.max() - image.min()
|
||
image = img_as_float(image)
|
||
if rescale_sigma:
|
||
range_post = image.max() - image.min()
|
||
# apply the same magnitude scaling to sigma
|
||
scale_factor = range_post / range_pre
|
||
if multichannel:
|
||
sigma = [s * scale_factor if s is not None else s
|
||
for s in sigma]
|
||
elif sigma is not None:
|
||
sigma *= scale_factor
|
||
return image, sigma
|
||
|
||
|
||
def _rescale_sigma_rgb2ycbcr(sigmas):
|
||
"""Convert user-provided noise standard deviations to YCbCr space.
|
||
|
||
Notes
|
||
-----
|
||
If R, G, B are linearly independent random variables and a1, a2, a3 are
|
||
scalars, then random variable C:
|
||
C = a1 * R + a2 * G + a3 * B
|
||
has variance, var_C, given by:
|
||
var_C = a1**2 * var_R + a2**2 * var_G + a3**2 * var_B
|
||
"""
|
||
if sigmas[0] is None:
|
||
return sigmas
|
||
sigmas = np.asarray(sigmas)
|
||
rgv_variances = sigmas * sigmas
|
||
for i in range(3):
|
||
scalars = ycbcr_from_rgb[i, :]
|
||
var_channel = np.sum(scalars * scalars * rgv_variances)
|
||
sigmas[i] = np.sqrt(var_channel)
|
||
return sigmas
|
||
|
||
|
||
def denoise_wavelet(image, sigma=None, wavelet='db1', mode='soft',
|
||
wavelet_levels=None, multichannel=False,
|
||
convert2ycbcr=False, method='BayesShrink',
|
||
rescale_sigma=None):
|
||
"""Perform wavelet denoising on an image.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray ([M[, N[, ...P]][, C]) of ints, uints or floats
|
||
Input data to be denoised. `image` can be of any numeric type,
|
||
but it is cast into an ndarray of floats for the computation
|
||
of the denoised image.
|
||
sigma : float or list, optional
|
||
The noise standard deviation used when computing the wavelet detail
|
||
coefficient threshold(s). When None (default), the noise standard
|
||
deviation is estimated via the method in [2]_.
|
||
wavelet : string, optional
|
||
The type of wavelet to perform and can be any of the options
|
||
``pywt.wavelist`` outputs. The default is `'db1'`. For example,
|
||
``wavelet`` can be any of ``{'db2', 'haar', 'sym9'}`` and many more.
|
||
mode : {'soft', 'hard'}, optional
|
||
An optional argument to choose the type of denoising performed. It
|
||
noted that choosing soft thresholding given additive noise finds the
|
||
best approximation of the original image.
|
||
wavelet_levels : int or None, optional
|
||
The number of wavelet decomposition levels to use. The default is
|
||
three less than the maximum number of possible decomposition levels.
|
||
multichannel : bool, optional
|
||
Apply wavelet denoising separately for each channel (where channels
|
||
correspond to the final axis of the array).
|
||
convert2ycbcr : bool, optional
|
||
If True and multichannel True, do the wavelet denoising in the YCbCr
|
||
colorspace instead of the RGB color space. This typically results in
|
||
better performance for RGB images.
|
||
method : {'BayesShrink', 'VisuShrink'}, optional
|
||
Thresholding method to be used. The currently supported methods are
|
||
"BayesShrink" [1]_ and "VisuShrink" [2]_. Defaults to "BayesShrink".
|
||
rescale_sigma : bool or None, optional
|
||
If False, no rescaling of the user-provided ``sigma`` will be
|
||
performed. The default of ``None`` rescales sigma appropriately if the
|
||
image is rescaled internally. A ``DeprecationWarning`` is raised to
|
||
warn the user about this new behaviour. This warning can be avoided
|
||
by setting ``rescale_sigma=True``.
|
||
|
||
.. versionadded:: 0.16
|
||
``rescale_sigma`` was introduced in 0.16
|
||
|
||
Returns
|
||
-------
|
||
out : ndarray
|
||
Denoised image.
|
||
|
||
Notes
|
||
-----
|
||
The wavelet domain is a sparse representation of the image, and can be
|
||
thought of similarly to the frequency domain of the Fourier transform.
|
||
Sparse representations have most values zero or near-zero and truly random
|
||
noise is (usually) represented by many small values in the wavelet domain.
|
||
Setting all values below some threshold to 0 reduces the noise in the
|
||
image, but larger thresholds also decrease the detail present in the image.
|
||
|
||
If the input is 3D, this function performs wavelet denoising on each color
|
||
plane separately.
|
||
|
||
.. versionchanged:: 0.16
|
||
For floating point inputs, the original input range is maintained and
|
||
there is no clipping applied to the output. Other input types will be
|
||
converted to a floating point value in the range [-1, 1] or [0, 1]
|
||
depending on the input image range. Unless ``rescale_sigma = False``,
|
||
any internal rescaling applied to the ``image`` will also be applied
|
||
to ``sigma`` to maintain the same relative amplitude.
|
||
|
||
Many wavelet coefficient thresholding approaches have been proposed. By
|
||
default, ``denoise_wavelet`` applies BayesShrink, which is an adaptive
|
||
thresholding method that computes separate thresholds for each wavelet
|
||
sub-band as described in [1]_.
|
||
|
||
If ``method == "VisuShrink"``, a single "universal threshold" is applied to
|
||
all wavelet detail coefficients as described in [2]_. This threshold
|
||
is designed to remove all Gaussian noise at a given ``sigma`` with high
|
||
probability, but tends to produce images that appear overly smooth.
|
||
|
||
Although any of the wavelets from ``PyWavelets`` can be selected, the
|
||
thresholding methods assume an orthogonal wavelet transform and may not
|
||
choose the threshold appropriately for biorthogonal wavelets. Orthogonal
|
||
wavelets are desirable because white noise in the input remains white noise
|
||
in the subbands. Biorthogonal wavelets lead to colored noise in the
|
||
subbands. Additionally, the orthogonal wavelets in PyWavelets are
|
||
orthonormal so that noise variance in the subbands remains identical to the
|
||
noise variance of the input. Example orthogonal wavelets are the Daubechies
|
||
(e.g. 'db2') or symmlet (e.g. 'sym2') families.
|
||
|
||
References
|
||
----------
|
||
.. [1] Chang, S. Grace, Bin Yu, and Martin Vetterli. "Adaptive wavelet
|
||
thresholding for image denoising and compression." Image Processing,
|
||
IEEE Transactions on 9.9 (2000): 1532-1546.
|
||
:DOI:`10.1109/83.862633`
|
||
.. [2] D. L. Donoho and I. M. Johnstone. "Ideal spatial adaptation
|
||
by wavelet shrinkage." Biometrika 81.3 (1994): 425-455.
|
||
:DOI:`10.1093/biomet/81.3.425`
|
||
|
||
Examples
|
||
--------
|
||
>>> from skimage import color, data
|
||
>>> img = img_as_float(data.astronaut())
|
||
>>> img = color.rgb2gray(img)
|
||
>>> img += 0.1 * np.random.randn(*img.shape)
|
||
>>> img = np.clip(img, 0, 1)
|
||
>>> denoised_img = denoise_wavelet(img, sigma=0.1, rescale_sigma=True)
|
||
|
||
"""
|
||
if method not in ["BayesShrink", "VisuShrink"]:
|
||
raise ValueError(
|
||
('Invalid method: {}. The currently supported methods are '
|
||
'"BayesShrink" and "VisuShrink"').format(method))
|
||
|
||
# floating-point inputs are not rescaled, so don't clip their output.
|
||
clip_output = image.dtype.kind != 'f'
|
||
|
||
if convert2ycbcr and not multichannel:
|
||
raise ValueError("convert2ycbcr requires multichannel == True")
|
||
|
||
if rescale_sigma is None:
|
||
msg = (
|
||
"As of scikit-image 0.16, automated rescaling of sigma to match "
|
||
"any internal rescaling of the image is performed. Setting "
|
||
"rescale_sigma to False, will disable this new behaviour. To "
|
||
"avoid this warning the user should explicitly set rescale_sigma "
|
||
"to True or False."
|
||
)
|
||
warn(msg, FutureWarning, stacklevel=2)
|
||
rescale_sigma = True
|
||
image, sigma = _scale_sigma_and_image_consistently(image,
|
||
sigma,
|
||
multichannel,
|
||
rescale_sigma)
|
||
if multichannel:
|
||
if convert2ycbcr:
|
||
out = color.rgb2ycbcr(image)
|
||
# convert user-supplied sigmas to the new colorspace as well
|
||
if rescale_sigma:
|
||
sigma = _rescale_sigma_rgb2ycbcr(sigma)
|
||
for i in range(3):
|
||
# renormalizing this color channel to live in [0, 1]
|
||
_min, _max = out[..., i].min(), out[..., i].max()
|
||
scale_factor = _max - _min
|
||
if scale_factor == 0:
|
||
# skip any channel containing only zeros!
|
||
continue
|
||
channel = out[..., i] - _min
|
||
channel /= scale_factor
|
||
sigma_channel = sigma[i]
|
||
if sigma_channel is not None:
|
||
sigma_channel /= scale_factor
|
||
out[..., i] = denoise_wavelet(channel,
|
||
wavelet=wavelet,
|
||
method=method,
|
||
sigma=sigma_channel,
|
||
mode=mode,
|
||
wavelet_levels=wavelet_levels,
|
||
rescale_sigma=rescale_sigma)
|
||
out[..., i] = out[..., i] * scale_factor
|
||
out[..., i] += _min
|
||
out = color.ycbcr2rgb(out)
|
||
else:
|
||
out = np.empty_like(image)
|
||
for c in range(image.shape[-1]):
|
||
out[..., c] = _wavelet_threshold(image[..., c],
|
||
wavelet=wavelet,
|
||
method=method,
|
||
sigma=sigma[c], mode=mode,
|
||
wavelet_levels=wavelet_levels)
|
||
else:
|
||
out = _wavelet_threshold(image, wavelet=wavelet, method=method,
|
||
sigma=sigma, mode=mode,
|
||
wavelet_levels=wavelet_levels)
|
||
|
||
if clip_output:
|
||
clip_range = (-1, 1) if image.min() < 0 else (0, 1)
|
||
out = np.clip(out, *clip_range, out=out)
|
||
return out
|
||
|
||
|
||
def estimate_sigma(image, average_sigmas=False, multichannel=False):
|
||
"""
|
||
Robust wavelet-based estimator of the (Gaussian) noise standard deviation.
|
||
|
||
Parameters
|
||
----------
|
||
image : ndarray
|
||
Image for which to estimate the noise standard deviation.
|
||
average_sigmas : bool, optional
|
||
If true, average the channel estimates of `sigma`. Otherwise return
|
||
a list of sigmas corresponding to each channel.
|
||
multichannel : bool
|
||
Estimate sigma separately for each channel.
|
||
|
||
Returns
|
||
-------
|
||
sigma : float or list
|
||
Estimated noise standard deviation(s). If `multichannel` is True and
|
||
`average_sigmas` is False, a separate noise estimate for each channel
|
||
is returned. Otherwise, the average of the individual channel
|
||
estimates is returned.
|
||
|
||
Notes
|
||
-----
|
||
This function assumes the noise follows a Gaussian distribution. The
|
||
estimation algorithm is based on the median absolute deviation of the
|
||
wavelet detail coefficients as described in section 4.2 of [1]_.
|
||
|
||
References
|
||
----------
|
||
.. [1] D. L. Donoho and I. M. Johnstone. "Ideal spatial adaptation
|
||
by wavelet shrinkage." Biometrika 81.3 (1994): 425-455.
|
||
:DOI:`10.1093/biomet/81.3.425`
|
||
|
||
Examples
|
||
--------
|
||
>>> import skimage.data
|
||
>>> from skimage import img_as_float
|
||
>>> img = img_as_float(skimage.data.camera())
|
||
>>> sigma = 0.1
|
||
>>> img = img + sigma * np.random.standard_normal(img.shape)
|
||
>>> sigma_hat = estimate_sigma(img, multichannel=False)
|
||
"""
|
||
if multichannel:
|
||
nchannels = image.shape[-1]
|
||
sigmas = [estimate_sigma(
|
||
image[..., c], multichannel=False) for c in range(nchannels)]
|
||
if average_sigmas:
|
||
sigmas = np.mean(sigmas)
|
||
return sigmas
|
||
elif image.shape[-1] <= 4:
|
||
msg = ("image is size {0} on the last axis, but multichannel is "
|
||
"False. If this is a color image, please set multichannel "
|
||
"to True for proper noise estimation.")
|
||
warn(msg.format(image.shape[-1]))
|
||
coeffs = pywt.dwtn(image, wavelet='db2')
|
||
detail_coeffs = coeffs['d' * image.ndim]
|
||
return _sigma_est_dwt(detail_coeffs, distribution='Gaussian')
|