Important: Please read the installation page for details about how to install the toolboxes. $\newcommand{\dotp}[2]{\langle #1, #2 \rangle}$ $\newcommand{\enscond}[2]{\lbrace #1, #2 \rbrace}$ $\newcommand{\pd}[2]{ \frac{ \partial #1}{\partial #2} }$ $\newcommand{\umin}[1]{\underset{#1}{\min}\;}$ $\newcommand{\umax}[1]{\underset{#1}{\max}\;}$ $\newcommand{\umin}[1]{\underset{#1}{\min}\;}$ $\newcommand{\uargmin}[1]{\underset{#1}{argmin}\;}$ $\newcommand{\norm}[1]{\|#1\|}$ $\newcommand{\abs}[1]{\left|#1\right|}$ $\newcommand{\choice}[1]{ \left\{ \begin{array}{l} #1 \end{array} \right. }$ $\newcommand{\pa}[1]{\left(#1\right)}$ $\newcommand{\diag}[1]{{diag}\left( #1 \right)}$ $\newcommand{\qandq}{\quad\text{and}\quad}$ $\newcommand{\qwhereq}{\quad\text{where}\quad}$ $\newcommand{\qifq}{ \quad \text{if} \quad }$ $\newcommand{\qarrq}{ \quad \Longrightarrow \quad }$ $\newcommand{\ZZ}{\mathbb{Z}}$ $\newcommand{\CC}{\mathbb{C}}$ $\newcommand{\RR}{\mathbb{R}}$ $\newcommand{\EE}{\mathbb{E}}$ $\newcommand{\Zz}{\mathcal{Z}}$ $\newcommand{\Ww}{\mathcal{W}}$ $\newcommand{\Vv}{\mathcal{V}}$ $\newcommand{\Nn}{\mathcal{N}}$ $\newcommand{\NN}{\mathcal{N}}$ $\newcommand{\Hh}{\mathcal{H}}$ $\newcommand{\Bb}{\mathcal{B}}$ $\newcommand{\Ee}{\mathcal{E}}$ $\newcommand{\Cc}{\mathcal{C}}$ $\newcommand{\Gg}{\mathcal{G}}$ $\newcommand{\Ss}{\mathcal{S}}$ $\newcommand{\Pp}{\mathcal{P}}$ $\newcommand{\Ff}{\mathcal{F}}$ $\newcommand{\Xx}{\mathcal{X}}$ $\newcommand{\Mm}{\mathcal{M}}$ $\newcommand{\Ii}{\mathcal{I}}$ $\newcommand{\Dd}{\mathcal{D}}$ $\newcommand{\Ll}{\mathcal{L}}$ $\newcommand{\Tt}{\mathcal{T}}$ $\newcommand{\si}{\sigma}$ $\newcommand{\al}{\alpha}$ $\newcommand{\la}{\lambda}$ $\newcommand{\ga}{\gamma}$ $\newcommand{\Ga}{\Gamma}$ $\newcommand{\La}{\Lambda}$ $\newcommand{\si}{\sigma}$ $\newcommand{\Si}{\Sigma}$ $\newcommand{\be}{\beta}$ $\newcommand{\de}{\delta}$ $\newcommand{\De}{\Delta}$ $\newcommand{\phi}{\varphi}$ $\newcommand{\th}{\theta}$ $\newcommand{\om}{\omega}$ $\newcommand{\Om}{\Omega}$
This numerical tour study image denoising using non-local means. This algorithm has been introduced for denoising purposes in BuaCoMoA05
using PyPlot
using NtToolBox
# using Autoreload
# arequire("NtToolBox")
This numerical tour is dedicated to the study of the structure of patches in images.
Size $N = n \times n$ of the image.
n = 128;
We load a noisy image $f_0\in \RR^N$.
c = [100, 200]
f0 = load_image("NtToolBox/src/data/lena.png")
f0 = rescale(f0[c[1] - Base.div(n, 2) + 1 : c[1] + Base.div(n, 2), c[2] - Base.div(n, 2) + 1 : c[2] + Base.div(n, 2)]);
Display $f_0$.
figure(figsize = (5, 5))
imageplot(f0)
Noise level $\si$.
sigma = .04;
Generate a noisy image $f=f_0+\epsilon$ where $\epsilon \times \Nn(0,\si^2\text{Id}_N)$.
using Distributions
f = f0 .+ sigma.*rand(Normal(), n, n);
Display $f$.
figure(figsize = (5,5))
imageplot(clamP(f))
We denote $w$ to be the half width of the patches, and $w_1=2w+1$ the full width.
w = 3
w1 = 2*w + 1;
We set up large $(n,n,w_1,w_1)$ matrices to index the the X and Y position of the pixel to extract.
Location of pixels to extract.
include("NtToolBox/src/ndgrid.jl")
(Y, X, dX, dY) = ndgrid(1 : n, 1 : n, -w : w, -w : w)
X = X + dX
Y = Y + dY;
We handle boundary condition by reflexion
X[X .< 1] = 2 .- X[X .< 1]
Y[Y .< 1] = 2 .- Y[Y .< 1]
X[X .> n] = 2*n .- X[X .> n]
Y[Y .> n] = 2*n .- Y[Y .> n];
Patch extractor operator.
I = X .+ (Y .- 1)*n
for k in 1 : Base.div(n, w)
for l in 1 : Base.div(n, w)
I[k, l, :, :] = transpose(I[k, l, :, :])
end
end
function patch(f)
return f'[I]
end;
Define the patch matrix $P$ of size $(n,n,w_1,w_1)$. Each $P(i,j,:,:)$ represent an $(w_1,w_1)$ patch extracted around pixel $(i,j)$ in the image.
P = patch(f);
Display some example of patches.
figure(figsize = (5,5))
for i in 1:16
x = rand(1 : n)
y = rand(1 : n)
imageplot(P[x, y, :, :], "", [4, 4, i])
end
Since NL-means type algorithms require the computation of many distances between patches, it is advantagous to reduce the dimensionality of the patch while keeping as much as possible of information.
Target dimensionality $d$.
d = 25;
A linear dimensionality reduction is obtained by Principal Component Analysis (PCA) that projects the data on a small number of leading direction of the covariance matrix of the patches.
Turn the patch matrix into an $(w_1*w_1,n*n)$ array, so that each $P(:,i)$ is a $w_1*w_1$ vector representing a patch.
resh = P -> transpose((reshape(P, (n*n,w1*w1))));
Operator to remove the mean of the patches to each patch.
remove_mean = Q -> Q - repeat(mean(Q, 1), inner = (w1*w1, 1));
Compute the mean and the covariance of the points cloud representing the patches.
P1 = remove_mean(resh(P))
C = P1*transpose(P1);
Extract the eigenvectors, sorted by decreasing amplitude.
(D, V) = eig(C)
D = D[end : -1 : 1]
V = V[:, end : -1 : 1];
Display the decaying amplitude of the eigenvalues.
plot(D, linewidth = 2)
ylim(0, maximum(D))
show()
Display the leading eigenvectors - they look like Fourier modes.
figure(figsize = (5,5))
for i in 1 : 16
imageplot(abs(reshape(V[:, i], (w1,w1))'), "", [4, 4, i])
end
Patch dimensionality reduction operator.
iresh = Q -> reshape(Q', (n, n, d))
descriptor = f -> iresh(V[: , 1 : d]'*remove_mean(resh(P)));
Each $H(i,j,:)$ is a $d$-dimensional descriptor of a patch.
H = descriptor(f);
NL-means applies, an adaptive averaging kernel is computed from patch distances to each pixel location.
We denote $H_{i} \in \RR^d$ the descriptor at pixel $i$. We define the distance matrix $$ D_{i,j} = \frac{1}{w_1^2}\norm{H_i-H_j}^2. $$
Operator to compute the distances $(D_{i,j})_j$ between the patch around $i=(i_1,i_2)$ and all the other ones.
distance = i -> sum((H - repeat(reshape(H[i[1], i[2], :], (1, 1, 25)), inner = [n, n, 1])).^2, 3)./(w1*w1);
The non-local mean filter computes a denoised image $\tilde f$ as :
$$ \tilde f_i = \sum_j K_{i,j} f_j $$where the weights $K$ are computed as : $$ K_{i,j} = \frac{ \tilde K_{i,j} }{ \sum_{j'} \tilde K_{i,j'} } \qandq \tilde K_{i,j} = e^{-\frac{D_{i,j}}{2\tau^2}} . $$
The width $\tau$ of the Gaussian is very important and should be adapted to match the noise level.
Compute and normalize the weight.
normalize = K -> K./sum(K)
kernel = (i, tau) -> normalize(exp(-distance(i)./(2*tau^2)));
Compute a typical example of kernel for some pixel position $(x,y)$.
tau = .05
i = [84, 73]
D = distance(i)
K = kernel(i, tau);
Display the squared distance and the kernel.
figure(figsize = (10,10))
imageplot(D[:, :], "D", [1, 2, 1])
imageplot(K[:, :], "K", [1, 2, 2])
PyObject <matplotlib.text.Text object at 0x000000002E345F98>
We set a "locality constant" $q$ that set the maximum distance between patches to compare. This allows to speed up computation, and makes NL-means type methods semi-global (to avoid searching in all the image).
q = 14;
Using this locality constant, we compute the distance between patches only within a window. Once again, one should be careful about boundary conditions.
selection = i -> [clamP(i[1] - q : i[1] + q, 1, n)'; clamP(i[2] - q : i[2] + q, 1, n)'];
Compute distance and kernel only within the window.
function distance_0(i,sel)
H1 = (H[sel[1, :], :, :])
H2 = (H1[:, sel[2, :], :])
return sum((H2 - repeat(reshape(H[i[1], i[2], :], (1, 1, length(H[i[1], i[2], :]))), inner = [length(sel[1, :]), length(sel[1, :]), 1])).^2, 3)/w1*w1
end
distance = i -> distance_0(i, selection(i))
kernel = (i, tau) -> normalize(exp(-distance(i)./(2*tau^2)));
Compute a typical example of kernel for some pixel position $(x,y)$.
Display the squared distance and the kernel.
sel = selection(i)
D = distance(i)
K = kernel(i, tau);
figure(figsize = (10,10))
imageplot(D[:, :], "D", [1, 2, 1])
imageplot(K[:, :], "K", [1, 2, 2])
PyObject <matplotlib.text.Text object at 0x000000002E31E240>
The NL-filtered value at pixel $(x,y)$ is obtained by averaging the values of $f$ with the weight $K$.
function NLval_0(K,sel)
f_temp = f[sel[1, :], :]
return sum(K.*f_temp[:, sel[2, :]])
end
NLval = (i, ta) -> NLval_0(kernel(i, tau), selection(i));
We apply the filter to each pixel location to perform the NL-means algorithm.
(Y, X) = meshgrid(0 : n - 1, 0 : n - 1)
function arrayfun(f, X, Y)
n = size(X)[1]
p = size(Y)[1]
R = zeros(n, p)
for k in 1:n
for l in 1:p
R[k,l] = f(k,l)
end
end
return R
end
NLmeans = tau -> arrayfun((i1, i2) -> NLval([i1,i2], tau), X, Y);
Display the result for some value of $\tau$.
tau = .03
figure(figsize = (5,5))
imageplot(NLmeans(tau))
tau = .03;
Exercise 1
Compute the denoising result for several values of $\tau$ in order to determine the optimal denoising that minimizes $\norm{\tilde f - f_0}$.
include("NtSolutions/denoisingadv_6_nl_means/exo1.jl")