Discrete Universal Denoiser Explained
In information theory and signal processing, the Discrete Universal Denoiser (DUDE) is a denoising scheme for recovering sequences over a finite alphabet, which have been corrupted by a discretememoryless channel. The DUDE was proposed in 2005 by Tsachy Weissman, Erik Ordentlich, Gadiel Seroussi, Sergio Verdú and Marcelo J. Weinberger.[1]
Overview
The Discrete Universal Denoiser (DUDE) is a denoising scheme that estimates anunknown signal
xn=\left(x1\ldotsxn\right)
over a finitealphabet from a noisy version
zn=\left(z1\ldotszn\right)
.While most
denoising schemes in the signal processingand statistics literature deal with
signals overan infinite alphabet (notably, real-valued signals), the DUDE addresses thefinite alphabet case. The noisy version
is assumed to be generated by transmitting
through a known
discretememoryless channel.
For a fixed context length parameter
, the DUDE counts of the occurrences of all the strings of length
appearing in
. The estimated value
is determined based the two-sided length-
context \left(zi-k,\ldots,zi-1,zi+1,\ldots,zi+k\right)
of
, taking into account all the other tokens in
with the same context, as well as the known channel matrix and the loss function being used.
The idea underlying the DUDE is best illustrated when
is arealization of a random vector
. If the conditional distribution
Xi|Zi-k,\ldots,Zi-1,Zi+1,\ldots,Zi+k
, namelythe distribution of the noiseless symbol
conditional on its noisy context
\left(Zi-k,\ldots,
Zi-1,Zi+1,\ldots,Zi+k\right)
was available, the optimalestimator
would be the
Bayes Response to
Xi|Zi-k,\ldots,Zi-1,Zi+1,\ldots,Zi+k
.Fortunately, whenthe channel matrix is known and non-degenerate, this conditional distributioncan be expressed in terms of the conditional distribution
Zi|Zi-k,\ldots,Zi-1,Zi+1,\ldots,Zi+k
, namelythe distribution of the noisy symbol
conditional on its noisycontext. This conditional distribution, in turn, can be estimated from anindividual observed noisy signal
by virtue of the
Law of Large Numbers, provided
is “large enough”.
Applying the DUDE scheme with a context length
to a sequence oflength
over a finite alphabet
requires
operations and space
O\left(min(n,|l{Z}|2k)
\right)
.
Under certain assumptions, the DUDE is a universal scheme in the sense of asymptotically performing as well as an optimal denoiser, which has oracle access to the unknown sequence. More specifically, assume that the denoising performance is measured using a given single-character fidelity criterion, and consider the regime where the sequence length
tends to infinity and the context length
tends to infinity “not too fast”. In the stochastic setting, where a doubly infinite sequence noiseless sequence
is a realization of a stationary process
, the DUDE asymptotically performs, in expectation, as well as the best denoiser, which has oracle access to the source distribution
. In the single-sequence, or “semi-stochastic” setting with a
fixed doubly infinite sequence
, the DUDE asymptotically performs as well as the best “sliding window” denoiser, namely any denoiser that determines
from the window
\left(zi-k,\ldots,zi+k\right)
, which has oracle access to
.
The discrete denoising problem
Let
be the finite alphabet of a fixed but unknown original “noiseless” sequence
xn=\left(x1,\ldots,xn\right)\inl{X}n
. The sequence is fed into a
discretememoryless channel (DMC). The DMC operates on each symbol
independently, producing a corresponding random symbol
in a finite alphabet
. The DMC is known and given as a
-by-
Markov matrix
, whose entries are
\pi(x,z)=P\left(Z=z|X=x\right)
. It is convenient to write
for the
-column of
. The DMC produces a random noisy sequence
Zn=\left(z1,\ldots,zn\right)\inl{Z}n
. A specific realization of this random vector will be denoted by
.A denoiser is a function
that attempts to recover the noiseless sequence
from a distorted version
. A specific denoised sequence is denoted by
\hat{x}n=\hat{X}n\left(zn
\right)=\left(\hat{X}1(zn),\ldots,
\right)
.The problem of choosing the denoiser
is known as signalestimation,
filtering or
smoothing. To compare candidate denoisers, we choose a single-symbol fidelity criterion
Λ:l{X} x l{X}\to[0,infty)
(for example, the Hamming loss) and define the per-symbol loss of the denoiser
at
by
\begin{align}
xn,zn\right)=
xi,
\right).
\end{align}
Ordering the elements of the alphabet
by
l{X}=\left(a1,\ldots,
a|l{X|}\right)
, the fidelity criterion can be given by a
-by-
matrix, with columns of the form
} = \left(\begin \Lambda(a_1,\hat) \\ \vdots \\ \Lambda(a_
,\hat) \end \right) \,. \end
The DUDE scheme
Step 1: Calculating the empirical distribution in each context
The DUDE corrects symbols according to their context. The context length
used is a tuning parameter of the scheme. For
, define the left context of the
-th symbol in
by
and the corresponding right context as
rk(zn,i)=\left(zi+1,\ldots,zi+k\right)
. A two-sided context is a combination
of a left and a right context.
The first step of the DUDE scheme is to calculate the empirical distribution of symbols in each possible two-sided context along the noisy sequence
. Formally, a given two-sided context
that appears once or more along
determines an empirical probability distribution over
, whose value at the symbol
is
\begin{align}
\mu\left(zn,lk,rk\right)[z]=
| |
\left\{k+1\leqi\leqn-k|(zi-k,\ldots,zi+k)=lkzrk\right\ |
|}
|
{|
\left\{k+1\leqi\leqn-k|lk(zn,i)=lkandrk(zn,i)=rk\right\}|}.
\end{align}
Thus, the first step of the DUDE scheme with context length
is to scan the input noisy sequence
once, and store the length-
empirical distribution vector
(or its non-normalized version, the count vector) for each two-sided context found along
. Since there are at most
Nn,k=min\left(n,|l{Z}|2k\right)
possible two-sided contexts along
, this step requires
operations and storage
.
Step 2: Calculating the Bayes response to each context
Denote the column of single-symbol fidelity criterion
, corresponding to the symbol
, by
}. We define the
Bayes Response to any vector
of length
with non-negative entries as
\begin{align}
\hat{X}Bayes(v)=
argmin\hat{x\inl{X}}λ\hat{x
}^\top\mathbf\,. \end
This definition is motivated in the background below.
The second step of the DUDE scheme is to calculate, for each two-sided context
observed in the previous step along
, and for each symbol
observed in each context (namely, any
such that
is a substring of
) the Bayes response to the vector
\Pi-\top\mu\left(zn,lk,rk\right)\odot\piz
, namely
\begin{align}
g(lk,z,rk):=\hat{X}Bayes\left(\Pi-\top\mu\left(zn,lk,rk\right)\odot\piz\right).\end{align}
Note that the sequence
and the context length
are implicit. Here,
is the
-column of
and for vectors
and
,
denotes their Schur (entrywise) product, defined by
\left(a\odotb\right)i=aibi
. Matrix multiplication is evaluated before the Schur product, so that
stands for
.
This formula assumed that the channel matrix
is square (
) and invertible. When
and
is not invertible, under the reasonable assumption that it has full row rank, we replace
above with its Moore-Penrose pseudo-inverse
\left(\Pi\Pi\top\right)-1\Pi
and calculate instead
\begin{align}
g(lk,z,r
\left((\Pi\Pi\top)-1\Pi\mu\left(zn,lk,rk\right)\odot
\piz\right).
\end{align}
By caching the inverse or pseudo-inverse
, and the values
}\odot \pi_z for the relevant pairs
(\hat{x},z)\inl{X} x l{Z}
, this step requires
operations and
storage.
Step 3: Estimating each symbol by the Bayes response to its context
The third and final step of the DUDE scheme is to scan
again and compute the actual denoised sequence
\hat{X}n(zn)=\left(
\ldots
\right)
. The denoised symbol chosen to replace
is the Bayes response to the two-sided context of the symbol, namely
\begin{align}
:=g\left(lk(zn,i),zi,rk(zn,i)\right).\end{align}
This step requires
operations and used the data structure constructed in the previous step.
In summary, the entire DUDE requires
operations and
storage.
Asymptotic optimality properties
The DUDE is designed to be universally optimal, namely optimal (is some sense, under some assumptions) regardless of the original sequence
.
Let
denote a sequence of DUDE schemes, as described above, where
uses a context length
that is implicit in the notation. We only require that
and that
.
For a stationary source
Denote by
the set of all
-block denoisers, namely all maps
.
Let
be an unknown stationary source and
be the distribution of the corresponding noisy sequence. Then
\begin{align}
\limn\toinftyE\left[
}\left(X^n,Z^n \right) \right]= \lim_\min_\mathbf \left[L_{\hat{X}^n}\left(X^n,Z^n
\right)\right]\,, \end
and both limits exist. If, in addition the source
is ergodic, then
\begin{align}
\limsupn\toinfty
}\left(X^n,Z^n \right) = \lim_\min_\mathbf \left[L_{\hat{X}^n}\left(X^n,Z^n
\right)\right]\,,\,\text\,. \end
For an individual sequence
Denote by
the set of all
-block
-th order sliding window denoisers, namely all maps
of the form
=f\left(zi-k,\ldots,zi+k\right)
with
arbitrary.
Let
be an unknown noiseless sequence stationary source and
be the distribution of the corresponding noisy sequence. Then
\begin{align}
\limn\toinfty\left[
}\left(x^n,Z^n \right) - \min_ L_\left(x^n,Z^n \right) \right ] =0 \,,\,\text\,. \end
Non-asymptotic performance
Let
denote the DUDE on with context length
defined on
-blocks. Then there exist explicit constants
and
that depend on
alone, such that for any
and any
we have
}B^k\,\leq \mathbf \left[L_{\hat{X}^n_{k}}\left(x^n,Z^n \right) -
\min_{\hat{X}^n\in\mathcal{D}_{n,k}} L_{\hat{X}^n}\left(x^n,Z^n \right)
\right] \leq \sqrt\frac |\mathcal|^ \,, \end
where
is the noisy sequence corresponding to
(whose randomness is due to the channel alone)
[2] .
In fact holds with the same constants
as above for
any
-block denoiser
. The lower bound proof requires that the channel matrix
be square and the pair
satisfies a certain technical condition.
Background
To motivate the particular definition of the DUDE using the Bayes response to a particular vector, we now find the optimal denoiser in the non-universal case, where the unknown sequence
is a realization of a random vector
, whose distribution is known.
Consider first the case
. Since the joint distribution of
is known, given the observed noisy symbol
, the unknown symbol
is distributed according to the known distribution
. By ordering the elements of
, we can describe this conditional distribution on
using a probability vector
, indexed by
, whose
-entry is
. Clearly the expected loss for the choice of estimated symbol
is
}^\top \mathbf_.
Define the Bayes Envelope of a probability vector
, describing a probability distribution on
, as the minimal expected loss
U(v)=
min\hat{x\inl{X}}v\topλ\hat{x
}, and the
Bayes Response to
as the prediction that achieves this minimum,
\hat{X}Bayes(v)=argmin\hat{x
}. Observe that the Bayes response is scale invariant in the sense that
\hat{X}Bayes(v)=\hat{X}Bayes(\alphav)
for
.
For the case
, then, the optimal denoiser is
\hat{X}(z)=\hat{X}Bayes\left(PX|z\right)
. This optimal denoiser can be expressed using the marginal distribution of
alone, as follows. When the channel matrix
is invertible, we have
PX|z\propto\Pi-\topPZ\odot\piz
where
is the
-th column of
. This implies that the optimal denoiser is given equivalently by
\hat{X}(z)=\hat{X}Bayes\left(\Pi-\topPZ\odot\piz\right)
. When
and
is not invertible, under the reasonable assumption that it has full row rank, we can replace
with its Moore-Penrose pseudo-inverse and obtain
\hat{X}(z)=\hat{X}Bayes\left((\Pi\Pi\top)-1\PiPZ\odot\piz
\right).
Turning now to arbitrary
, the optimal denoiser
(with minimal expected loss) is therefore given by the Bayes response to
\begin{align}
=\hat{X}Bayes
=
argmin\hat{x\inl{X}}λ\hat{x
}^\top \mathbf_\,, \end
where
is a vector indexed by
, whose
-entry is
. The conditional probability vector
is hard to compute. A derivation analogous to the case
above shows that the optimal denoiser admits an alternative representation, namely
, where
zn=\left(z1,\ldots,zi-1,zi+1,\ldots,zn\right)\in
l{Z}n-1
is a given vector and
is the probability vector indexed by
whose
-entry is
P\left((Z1,\ldots,Zn)=
(z1,\ldots,zi-1,z,zi+1,\ldots,zn)\right).
Again,
is replaced by a pseudo-inverse if
is not square or not invertible.
When the distribution of
(and therefore, of
) isnot available, the DUDE replaces the unknown vector
with an empirical estimateobtained along the noisy sequence
itself, namely with
\mu\left(Zi,lk(Zn,i),rk(Zn,i)\right)
. This leads to theabove definition of the DUDE.
While the convergence arguments behind the optimality properties above are moresubtle, we note that the above, combined with theBirkhoff Ergodic Theorem, is enough to prove that for a stationary ergodic source, the DUDE with context-length
is asymptotically optimal all
-th order sliding window denoisers.
Extensions
The basic DUDE as described here assumes a signal with a one-dimensional indexset over a finite alphabet, a known memorylesschannel and a context length that is fixed in advance. Relaxations of each of these assumptions have been considered in turn.[3] Specifically:
Applications
Application to image denoising
A DUDE-based framework for grayscale image denoising achieves state-of-the-art denoising for impulse-type noise channels (e.g., "salt and pepper" or "M-ary symmetric" noise), and good performance on the Gaussian channel (comparable to the Non-local means image denoising scheme on this channel). A different DUDE variant applicable to grayscale images is presented in.
Application to channel decoding of uncompressed sources
The DUDE has led to universal algorithms for channel decoding of uncompressed sources.[17]
Notes and References
- T. Weissman, E. Ordentlich, G. Seroussi, S. Verdu ́, and M.J. Weinberger. Universal discrete denoising: Known channel. IEEE Transactions on Information Theory,, 51(1):5–28, 2005.
- K. Viswanathan and E. Ordentlich. Lower limits of discrete universal denoising. IEEE Transactions on Information Theory, 55(3):1374–1386, 2009.
- E. . Ordentlich . G. . Seroussi . S. . Verd´u . M. J. . Weinberger . T. . Weissman . Reflections on the DUDE .
- A. Dembo and T. Weissman. Universal denoising for the finite-input-general-output channel.IEEE Trans. Inf. Theory, 51(4):1507–1517, April 2005.
- K. Sivaramakrishnan and T. Weissman. Universal denoising of discrete-time continuous amplitude signals. In Proc. of the 2006 IEEE Intl. Symp. on Inform. Theory, (ISIT’06),Seattle, WA, USA, July 2006.
- G. Motta, E. Ordentlich, I. Ramírez, G. Seroussi, and M. Weinberger, “TheDUDE framework for continuous tone image denoising,” IEEE Transactions onImage Processing, 20, No. 1, January 2011.
- K. Sivaramakrishnan and T. Weissman. Universal denoising of continuous amplitude signalswith applications to images. In Proc. of IEEE International Conference on Image Processing,Atlanta, GA, USA, October 2006, pp. 2609–2612
- C. D. Giurcaneanu and B. Yu. Efficient algorithms for discrete universal denoising for channelswith memory. In Proc. of the 2005 IEEE Intl. Symp. on Inform. Theory, (ISIT’05), Adelaide,Australia, Sept. 2005.
- R. Zhang and T. Weissman. Discrete denoising for channels with memory. Communicationsin Information and Systems (CIS), 5(2):257–288, 2005.
- G. M. Gemelos, S. Sigurjonsson, T. Weissman. Universal minimax discrete denoising underchannel uncertainty. IEEE Trans. Inf. Theory, 52:3476–3497, 2006.
- G. M. Gemelos, S. Sigurjonsson and T. Weissman. Algorithms for discrete denoising underchannel uncertainty. IEEE Trans. Signal Process., 54(6):2263–2276, June 2006.
- E. Ordentlich, M.J. Weinberger, and T. Weissman. Multi-directional context sets with applications to universal denoising and compression. In Proc. of the 2005 IEEE Intl. Symp. onInform. Theory, (ISIT’05), Adelaide, Australia, Sept. 2005.
- J. Yu and S. Verd´u. Schemes for bidirectional modeling of discrete stationary sources. IEEETrans. Inform. Theory, 52(11):4789–4807, 2006.
- S. Chen, S. N. Diggavi, S. Dusad and S. Muthukrishnan. Efficient string matching algorithmsfor combinatorial universal denoising. In Proc. of IEEE Data Compression Conference (DCC),Snowbird, Utah, March 2005.
- G. Gimel’farb. Adaptive context for a discrete universal denoiser. In Proc. Structural, Syntactic, and Statistical Pattern Recognition, Joint IAPR International Workshops, SSPR 2004and SPR 2004, Lisbon, Portugal, August 18–20, pp. 477–485
- E. Ordentlich, G. Seroussi, S. Verd´u, M.J. Weinberger, and T. Weissman. A universal discreteimage denoiser and its application to binary images. In Proc. IEEE International Conferenceon Image Processing, Barcelona, Catalonia, Spain, September 2003.
- E. Ordentlich, G. Seroussi, S. Verdú, and K. Viswanathan, "UniversalAlgorithms for Channel Decoding of Uncompressed Sources," IEEE Trans.Information Theory, vol. 54, no. 5, pp. 2243–2262, May 2008