Sie sind auf Seite 1von 3

International Symposium on Information Theory and Its Applications

Xi’an, PRC, October 7–11, 2002

Optimal Quantization for Soft-Decision Decoding Revisited


Xiao Ma† , Xiaoyi Zhang‡ , Hongyi Yu‡ and Aleksandar Kavčić†
† ‡
Division of Engineering and Applied Sciences Department of Communication Engineering
Harvard University Zhengzhou Information Engineering University
33 Oxford Street, Cambridge, MA 02138, USA Zhengzhou, China

Abstract The quantities C and R defined above represent the trans-


mission limits and the compression limits (under certain
A new criterion for quantization is proposed. The criterion constraints), respectively. In this paper, we attempt to in-
is optimal in the sense that the resulting discrete channel has terpret the following two problems,
the highest information rate.
P3. For given PY |X ,

1. Mutual Information R3 = min I(PX , PY |X ) (4)


PX ∈P
Let X = {a1 , a2 , · · · , aJ } and Y = {b1 , b2 , · · · , bK }
where P is a given subset of probability distributions.
be two finite sets. Let X be the “source” (a random vari-
able) characterized by a probability mass function PX (aj ), P4. For given PX ,
1 ≤ j ≤ J. Let PY |X (bk |aj ), 1 ≤ k ≤ K, 1 ≤ j ≤ J
be the “channel” (a probability transition matrix). The mu- R4 = max I(PX , PY |X ) (5)
tual information (MI) is defined as (for general definition, PY |X ∈W

see [1], pp.33-37)


where W is a given subset of probability transition
K
J X matrices.
X PY |X (bk |aj )
I(X; Y ) = PX (aj )PY |X (bk |aj ) log
j=1 k=1
PY (bk ) For most problems of interest, problems P1 and P2 can be
(1) solved by the Blahut-Arimoto algorithm [2] [3]. On the
P
where PY (bk ) = 1≤j≤J PX (aj )PY |X (bk |aj ). By defi- other hand, it is not clear how to solve problems P3 and
nition, we may denote I(X; Y ) by I(PX , PY |X ) to empha- P4. However, if the subset W (P, respectively) can be pa-
size that the MI is a function of the “source” distribution PX rameterized by one or two parameters, one can always find
and the “channel” transition probability matrix PY |X . The satisfactory numerical solutions.
following two problems are well-known,
2. Optimal Quantization
P1. For given PY |X ,

C = max I(PX , PY |X ) (2)


PX ∈P 2.1. Problem Statement

where P is a given subset of probability distributions. As an example, we choose the additive white Gaus-
sian noise (AWGN) channel. Assume that the input is con-
P2. For given PX , strained to X = {−1, +1}, and the output has to be quan-
tized to Y = {−1, 0, +1}. The original channel (binary
R= min I(PX , PY |X ) (3) input/continuous output) is described as
PY |X ∈W

Zt = Xt + Nt , t = 0, 1, ... (6)
where W is a given subset of probability transition
matrices. where Nt is an independent identically distributed (i.i.d.)
† This work was supported by the USA NSF under Grant No. CCR- Gaussian noise with mean zero and variance σ 2 . The signal-
0118701. to-noise-ratio (SNR) is defined as SNR = 10 log 10 (1/σ 2 ).
The quantizer is described as
1−α − β

 −1, Zt < −T +1 +1
Yt = 0, −T ≤ Zt ≤ T (7)
α

+1, Zt > T β
where the threshold T ≥ 0 is a parameter. An immediate 0
question is how to choose T .
β
α
2.2. Forney’s Criterion 1−α − β
−1 −1
The following argument is based on an Exercise in [4].
Let C be a binary linear block code [n, k, d] with d > 1. As- Figure 1: The binary-input/ternary-output DMC model.
sume that a codeword is transmitted over an AWGN channel
with BPSK signaling and the received signals are quantized The symmetric cutoff rate R̃0 for this DMC is evaluated
to ternary outputs according to Equation (7), where a zero as [5]
“0” is interpreted as an erasure. Assume that there occur s  p 
erasures and t errors. It is well known that such a code is R̃0 (T ) = 1 − log2 1 + α + 2 (1 − α − β)β . (14)
able to recover the transmitted codeword if 2t + s < d. On
the other hand, for any integers t and s such that 2t + s ≥ d The cutoff rate criterion is to choose threshold T that
and for any decoding rule, there exists some pattern of t er- maximizes the cutoff rate R0 , for details, see [5][6][7][8].
rors and s erasures that will cause a decoding error. There- A simpler criterion [5] is to choose T = Tc such that
fore, the minimum squared distance from any codeword to
Tc = arg max R̃0 (T ). (15)
its decoding boundary is equal to T ≥0

L(T ) = min {s(1 − T )2 + t(1 + T )2 }. (8)


2t+s≥d 2.4. The MI criterion
Forney’s criterion [4] is to choose T = Tf such that Let X be an independent uniformly distributed (i.u.d.)
random variable over X = {−1, +1}. Applying Equa-
Tf = arg max L(T ) (9) tion (1) to the DMC shown in Figure 1, we obtain the mutual
0≤T ≤1
information (rate) between X and Y as follows
It can be verified that the solution is
√ 1−α
2−1 R(T ) = (1−α−β) log(1−α−β)+β log β−(1−α) log .
Tf = √ (10) 2
2+1 (16)
We propose to choose T = Tm according to the i.u.d. mu-
which is independent of SNR. tual information (MI) criterion as

2.3. The Cutoff Rate Criterion Tm = arg max R(T ). (17)


T ≥0
Evidently, the original AWGN channel is transformed
The MI criterion is optimal in the sense that the resulting
into a discrete memoryless channel (DMC) parameterized
DMC has the highest (i.u.d.) information rate.
by the quantization threshold T . The resulting DMC is
shown in Figure 1. The erasure probability is
    3. Numerical Results
1−T 1+T
α=Q −Q (11) Unlike Forney’s criterion, closed-form solutions to both
σ σ
the cutoff rate criterion and the MI criterion (see Equa-
and the error probability is tions (15) and (17)) are not available. The numerical so-

1+T
 lutions to these two criterion are plotted in Figure 2. From
β=Q , (12) Figure 2, a simple approximation for the MI criterion can
σ
be written as
where Q function is defined as
Tm ≈ exp(a · SNR + b), (18)
Z∞  2

1 t
Q(x) = √ exp − dt. (13) where a and b are two yet-to-be determined parameters
2π 2 which are independent of SNR.
x
1 1
Forney’s criterion continuos output
cutoff rate criterion hard−decision
MI criterion 0.9 MI criterion
0.5 Forney criterion
cutoff rate criterion
0.8

Information Rates (bits / channel−use)


0
0.7

−0.5 0.6
log(Threshold)

−1 0.5

0.4
−1.5
0.3

−2
0.2

−2.5 0.1

0
−3 −10 −5 0 5 10 15
−10 −5 0 5 10 15
SNR (dB) SNR (dB)

Figure 2: The quantization threshold versus SNR. Figure 3: Information rates for different quantization crite-
ria.

For a given SNR, we may first determine the thresh-


olds T using different criteria, and then evaluate the mu- [3] S. Arimoto, “An algorithm for computing the capacity of
arbitrary discrete memoryless channels,” IEEE Trans. In-
tual information rates using Equation (16). Numerical re-
form. Theory, vol. IT-18, pp. 14–20, Jan. 1972.
sults (see Fig. 3) show that, 1) the cut-off rate criterion is
almost equivalent to the optimal MI criterion at low SNRs, [4] G. D. Forney Jr., “Principles of digital communication II.”
and 2) Forney’s criterion is almost equivalent to the optimal MIT course, Spring 2002.
MI criterion at high SNRs. [5] L.-N. Lee, “On optimal soft-decision demodulation,” IEEE
Transactions on Information Theory, vol. IT-22, pp. 437–
444, July 1976.
4. Conclusions
[6] J. M. Wozencraft and R. S. Kennedy, “Modulattion and de-
A quantizer can be viewed as a post-channel proces- modulation for probabilistic coding,” IEEE Transactions on
sor. The MI criterion is to choose the post-channel pro- Information Theory, vol. IT-12, pp. 291–297, July 1966.
cessor (under given complexity constraints) such that the [7] S. G. Wilson, Digital Modulation and Coding. Prentice Hall,
resulting “channel” has the highest information rate. Sim- 1996.
ilar ideas appear in [9] [10] [11], where our goal was to [8] G. Jeong and D. Hsia, “Optimal quantization for soft-
choose a “pre-channel” processor such that the resulting decision turbo decoder,” in IEEE Vehicular Technology Con-
super-channel has a high (i.u.d.) mutual information rate. ference, Volume 3, pp. 1620 –1624, Sept. 1999.
The (i.u.d.) MI criterion is optimal in the sense that it gives [9] X. Ma, N. Varnica, and A. Kavčić, “Matched information
the highest information rate. Furthermore, since the in- rate codes for binary ISI channels,” in IEEE Int. Symp. In-
puts are (i.u.d.), this information rate can be achieved by form. Theory, (Lausanne, Switzerland), p. 269, June 30-July
an (outer) linear code. 5, 2002.
[10] N. Varnica, X. Ma, and A. Kavčić, “Capacity of power con-
Acknowledgments strained memoryless AWGN channels with fixed input con-
stellations,” in Proc. IEEE GLOBECOM’2002, (Taipei, Tai-
Xiao Ma wishes to thank Mr. Shaohua Yang for his help. wan, China), Nov. 2002. to appear.
[11] A. Kavčić, X. Ma, and N. Varnica, “Matched information
rate codes for partial response channels.” submitted to IEEE
References
Trans. Inform. Theory, June 2002.
[1] R. G. Gallager, Information Theory and Reliable Communi-
cation. New York: John Wiley and Sons, Inc., 1968.
[2] R. E. Blahut, “Computation of channel capacity and rate-
distortion functions,” IEEE Trans. Inform. Theory, vol. IT-
18, pp. 460–473, July 1972.

Das könnte Ihnen auch gefallen