TITLE: Discrete Denoising for Channels with Memory
SPEAKER: Rui Zhang (Stanford University)
DATE: 2:00 - 3:00 P.M., Tuesday March 1, 2005
LOCATION: Pi, 1L (PA)
ABSTRACT:
The problem of estimating a finite-alphabet signal corrupted by a
finite-alphabet noise process, aka "discrete denoising", is arising in an
increasing variety of applications spanning engineering, statistics, computer
science, and biology.
For the case where the corruption mechanism (channel) is memoryless (independent
noise components), a practical (linear time) algorithm dubbed DUDE (Discrete
Universal DEnoiser) was recently introduced. This denoiser was shown to achieve
optimum performance (in the limit of large data sets), with no a priori
knowledge of statistical (or any other) properties of the noiseless data. This
talk will present an extension of the algorithm that accommodates possible
memory (dependence) in the noise.
Specifically, we consider the problem of estimating a discrete signal X = (X1, .
. . ,Xn) based on its noise-corrupted observation signal Z = (Z1, . . . ,Zn).
The noise-free, noisy, and reconstruction signals are all assumed to have
components taking values in the same finite M-ary alphabet. For concreteness we
focus first on the additive noise channel Zi = Xi +Ni, where addition is
modulo-M, and {Ni} is the noise process. The cumulative loss is measured by a
given loss function. The distribution of the noise is assumed known, and may
have memory restricted only to stationarity and a mild mixing condition. We
develop a sequence of denoisers (indexed by the block length n) which we show to
be asymptotically universal in both a semi-stochastic setting (where the
noiseless signal is an individual sequence) and in a fully stochastic setting
(where the noiseless signal is emitted from a stationary source). It is detailed
how the problem formulation, denoising schemes, and performance guarantees carry
over to more general (non-additive) channels, as well as to higher-dimensional
data arrays.
The proposed schemes are shown to be computationally implementable. We also
discuss a variation on these schemes that is likely to do well on data of
moderate size. We conclude with a report of experimental results for the binary
burst noise channel, where the noise is a finite-state hidden Markov process
(FS-HMP), and a finite-state hidden Markov random field (FS-HMRF), in the
respective cases of one- and two-dimensional data. These support the theoretical
predictions and show that, in practice, there is much to be gained by taking the
channel memory into account.
This is joint work with Tsachy Weissman.
|
|
|