Beruflich Dokumente
Kultur Dokumente
E-mail: csiszar@renyi.hu. Partially supported by the Hungarian Research Grant OTKA T068258.
E-mail: hiai@math.is.tohoku.ac.jp. Partially supported by Grant-in-Aid for Scientific Research
(B)17340043.
3
E-mail: petz@math.bme.hu. Partially supported by the Hungarian Research Grant OTKA T068258.
2
Introduction
1
(n1) + (n2) + + (n1)
n
In this section we assume that supp supp for the support projections of and .
One can simply compute:
S(Rn kn ) = Tr(Rn log Rn Rn log n )
= S(Rn ) (n 1)Tr log Tr log .
Hence the identity
S(Rn kn ) = S(Rn ) + (n 1)S() + S(k) + S()
holds. It follows that the conjecture (1) is equivalent to the statement
S(Rn kn ) 0 as n
2
(3)
1X i
(X) I
n i=0
in the strong operator topology in B(H), since (X) = Tr 1/2 1/2 = 1.
Since the continuous functional calculus preserves the strong convergence (simply due
to approximation by polynomials on a compact set), we obtain
!
n1
1X i
This shows that the upper bound (3) converges to 0 and the proof is complete.
By the same proof one can obtain that for
Rm,n :=
1 m
(n1) + m (n2) + + (n1) m ,
n
(4)
() =
i1 ,...,in
(i1 /i1 ) + + (in /in )
i1 in .
n
(5)
i1 in
n
i ,...,i
1
X (i /i ) + + (i /i )
n
n
1
1
(
=
i1 in log(i1 in ) + Qn
n
i ,...,i
n
n
1X
=
n k=1
i1 i2 in log ik +
i1 i2 in log ik
i1 ,...,in
i1 ,...,in
++
i1 i2 in log ik
i1 ,...,in
n
X
X
1X
ik log ik
ik log ik +
(n 1)
=
n k=1
i
i
k
= (n 1)S()
+ Qn
+ Qn
i log i + Qn ,
i=1
where
Qn :=
i1 ,...,in
(i1 /i1 ) + + (in /in )
(i1 in )
.
n
.
n
The strong law of large numbers says that
X
X
i
X1 + + Xn
i =
i almost surely.
E(X1 ) =
n
i
i=1
i=1
Since ((X1 + + Xn )/n) is uniformly bounded, the Lebesgue bounded convergence
theorem implies that
X
Qn
i
i=1
as n .
i=1
i log i +
i=1
i log i + Qn = S(k) + Qn
i=1
X d i i
d i2
in
n
2
log
n
n
i2 ,...,in
X
d
d
d
(i2 in ) log(i2 in )
log
=
n i ,...,i
n
n
n
2
d
d
d
= (n 1)S()
log .
n
n
n
() = d (n 1)S() d log
d
.
n
P
We make a lower estimate to the entropy of Rn in such a way that we compute ()
when runs over A and B. It is clear now that
X
X
S(Rn ) (n 1)S() S()
() +
() (n 1)S() S()
A
as n .
Interpretation as capacity
(x, y X ).
(6)
p(x)S(x ),
(7)
xX
which is actually a simple consequence of the monotonicity of the relative entropy under state transformation [7], see also [11]. I(X, W ) is the so-called Holevo quantity or
classical-quantum mutual information, and it satises the identity
X
p(x)S(x k) = I(X, W ) + S(X k),
(8)
xX
The concept related to the conjecture we study is the channel capacity per unit cost
which is dened next for simplicity only in the case where X = {0, 1}, the cost of a
character 0 X is 1, while the cost of 1 X is 0.
For a memoryless channel with a binary input alphabet X = {0, 1} and an > 0, a
number R > 0 is called an -achievable rate per unit cost if for every > 0 and for any
suciently large T , there exists a code of length n > T with at least eT (R) codewords
such that each of the codewords contains at most T 0s and the error probability is at
most . The largest R which is an -achievable per unit cost for every > 0 is the
channel capacity per unit cost.
7
1 X
|{i : xi = 0}|.
|A| xA
Proof: Let c(x) := |{i : xi = 0}| for x A. Since I(A, 0 , 1 ) is a particular Holevo
quantity I(X, W n ), we can use the identity (8) to get an upper bound
1 X
1 X
S(x kn
)
=
c(x)S(0 k1 ) = c(A)S(0 k1 )
1
|A| xA
|A| xA
for I(A, 0 , 1 ).
Lemma 2. If A {0, 1}n is a code of the channel W n , whose probability of error (for
some decoding scheme) does not exceed a given 0 < < 1, then
(1 ) log |A| log 2 I(A, 0 , 1 ).
Proof: The right-hand side is a bound for the classical mutual information I(X Y ) =
H(Y ) H(Y |X), where Y is the channel output, see (7). Since the error probability
Prob(X 6= Y ) is smaller than , application of the Fano inequality (see [3]) gives
H(X|Y ) log |A| + log 2.
Therefore
I(X Y ) = H(X) H(X|Y ) (1 ) log |A| log 2,
and the proof is complete.
The above two lemmas shows that the relative entropy S(0 k1 ) is an upper bound
for the channel capacity per unit cost of the channel W (0) = 0 and W (1) = 1 with
a binary input alphabet. In fact, assume that R > 0 is an -achievable rate. For every
> 0 and T > 0 there is a code A {0, 1}n for which we get by Lemmas 1 and 2
T S(0 k1 ) c(A)S(0 k1 ) I(A, 0 , 1 )
(1 ) log |A| log 2
(1 )T (R ) log 2.
Since T is arbitrarily large and , are arbitrarily small, R S(0 k1 ) follows. That
S(0 k1 ) equals the channel capacity per unit cost will be veried below.
Theorem 3. Let the classical-quantum channel W : X = {0, 1} B(K) be defined as
W (0) = 0 and W (1) = 1 . Assume that An {0, 1}n is chosen such that
8
1 X
|{i : xi = 0}| c
|An | xA
as n
for some real number c > 0 and for some natural number . If the random variable Xn
has a uniform distribution on An , then
1 X
lim S(Xn )
S(x ) = cS(k).
n
|An | xA
n
The proof of the theorem is divided into lemmas. We need the direct part of the
so-called quantum Stein lemma obtained in [6], see also [2, 5, 9, 12].
Lemma 3. Let 0 and 1 be density matrices. For every > 0 and 0 < R < S(0 k1 ),
if N is sufficiently large, then there is a projection E B(KN ) such that
N [E] := Tr N
0 (I E) <
and for N [E] := Tr N
1 E the estimate
1
log N [E] < R
N
holds.
Note that N is called the error of the rst kind, while N is the error of the second
kind.
Lemma 4. Assume that > 0, 0 < R < S(0 k1 ), is a positive integer and the
sequences x in An {0, 1}n contain at most copies of 0. Let the codewords be the
N-fold repetitions xN = (x, x, . . . , x) of the sequences x An . If N is the integer part
of
2n
1
log
R
and n is large enough, then there is a decoding scheme such that the error probability is
smaller than .
Proof: We follow the probabilistic construction in [13]. Let the codewords be the Nfold repetitions xN = (x, x, . . . , x) of the sequences x An . The corresponding output
density matrices act on the Hilbert space KN n (Kn )N . We decompose this Hilbert
space into an N-fold product in a dierent way. For each 1 i n, let Ki be the
9
Tr N
x Fy
y:y6=x
n
X
X
n
X Y
Tr N
xi Fyi
y:y6=x i=1
n
Y
Tr N
xj Fyj
n
X
Tr N
xi (I Fxi ).
i=1
If xi = 0, then
N
Tr N
xi (I Fxi ) = Tr 0 (I Ei ) ,
from the error of the second kind. It follows that + neN R is a bound for the error
probability. The rst term will be small if is small. The second term will be small
if N is large enough. If both terms are majorized by /2, then the statement of the
lemma holds. We can choose n so large that N dened by the statement should be large
enough.
Proof of Theorem 3: Since Lemma 1 gives an upper bound, that is,
1 X
lim sup S(Xn )
S(x ) cS(k),
|An | xA
n
n
10
Lemma 4 is about the N-times repeated input X N and describes a decoding scheme
with error probability at most . According to Lemma 2 we have
(1 ) log |An | 1 S(X N )
1 X
S(xN ).
|A| xA
n
log |An |
1
1 X
S(x ).
S(X )
N
N
|An | xA
n
log |An |
log n
log |An |
and the lower bound is arbitrarily close to cR. Since R < S(0 k1 ) was arbitrary, the
proof is complete.
References
[1] V.P. Belavkin and P. Staszewski, C*-algebraic generalization of relative entropy and
entropy, Ann. Inst. Henri Poincare, Sec. A 37(1982), 5158.
[2] I. Bjelakovic, J. Deuschel, T. Kr
uger, R. Seiler, R. Siegmund-Schultze and A. Szkola,
A quantum version of Sanovs theorem, Comm. Math. Phys. 260(2005), 659671.
[3] T. M. Cover and J. A. Thomas, Elements of Information Theory, Second edition,
Wiley-Interscience, Hoboken, NJ, 2006.
[4] L. Diosi, T. Feldmann and R. Koslo, On the exact identity between thermodynamic
and informatic entropies in a unitary model of friction, Int. J. Quantum Information,
4(2006), 99104.
[5] M. Hayashi, Quantum information. An introduction, Springer, 2006.
[6] F. Hiai and D. Petz, The proper formula for relative entropy and its asymptotics in
quantum probability, Comm. Math. Phys. 143(1991), 99114.
11
[7] A.S. Holevo, Some estimates for the amount of information transmittable by a quantum communication channel (in Russian), Problemy Peredachi Informacii, 9(1973),
311.
[8] M.A. Nielsen and I.L. Chuang, Quantum computation and quantum information,
Cambridge University Press, Cambridge, 2000.
[9] T. Ogawa and H. Nagaoka, Strong converse and Steins lemma in quantum hypothesis testing, IEEE Tans. Inf. Theory 46(2000), 24282433.
[10] M. Ohya and D. Petz, Quantum Entropy and its Use, Springer, 1993.
[11] M. Ohya, D. Petz and N. Watanabe, On capacities of quantum channels, Prob.
Math. Stat. 17(1997), 179196.
[12] D. Petz, Lectures on quantum information theory and quantum statistics, book
manuscript in preparation.
[13] S. Verdu, On channel capacity per unit cost, IEEE Trans. Inform. Theory 36(1990),
10191030.
12