Beruflich Dokumente
Kultur Dokumente
This talk is dedicated to David Slepian who taught me all that I know about informa*on theory and a lot more.
Introduc*on
A very short pictorial history of my rela*onship with Roberto Padovani and how I ended up teaching at UCSD.
USAF 1960-1963
UMASS 1973-1984
UCSD
In
1983
a
new
interdisciplinary
research
center
was
being
formed
at
UCSD.
It
was
called
the
Center
for
Magne*c
Recording
Research
(CMRR)
and
was
concerned
with
educa*ng
students
and
pursuing
research
in
magne*c
recording.
It
sounded
interes*ng
to
me
because:
Our
kids
had
all
le_
Amherst
and
we
were
looking
for
something
new.
I
had
worked
with
Goeried
Ungerboeck
at
IBM
Zurich
on
coding
for
a
par*al
response
channel
which
I
learned
was
a
model
for
the
magne*c
recording
channel.
UCSD
was
in
San
Diego.
A
Minor
Problem
I
knew
nothing
about
magne*c
recording.
Not
only
did
I
not
know
how
to
spell
coercivity
but
the
rst
*me
I
men*oned
it
in
a
talk
I
mispronounced
it.
But
UCSD
reluctantly
made
me
an
oer
as
the
rst
faculty
member
in
CMRR.
But
$119.00 Amazon.com
2 kbits/in2 $10,000
*
Areal
density
has
been
increased
more
than
250
million
:mes
with
respect
to
the
rst
RAMAC
in
1956
from
0.002
Mbit/in2
to
500
Gbit/in2
in
today
We
expect
much
higher
areal
density
in
the
future,
i.e.,
1
Tbit/in2
and
10
Tbit/in2
25
The total capacity of all of the drives shown on this factory oor was less than 20 Gigabytes. The total selling price of all of these drives was about $4,000,000!
Lower capacity consumer products (e.g., IPODs) have transi*oned to ash memory. But ash is s*ll more expensive than HDDs for computer applica*ons. Flash memory is NOT a direct replacement for HDDs because of several dierences to be discussed later.
Because of the success of Flash, the 1 HDD was discon*nued a few years ago.
Blocks a digital recording designer cant control Channel * * Blocks an informa*on theorist cant control Sink * Source Decoder Channel Decoder
Write Equalization
The
error
correc*ng
code
used
for
the
last
25
years
is
a
Reed
Solomon
code
in
conjunc*on
with
a
hard
input
decoder.
But
LDPC
codes
and
itera*ve
decoding
are
on
the
way!!
The
purpose
of
the
modula*on
code
is
to
prevent
certain
bad
sequences
from
being
wri\en.
To
an
informa*on
theorist,
this
is
coding
for
the
noiseless
channel.
An AWGN channel is o_en used as a rst order approxima*on for the channel model. But the actual channel is really much more complex. At low recording densi*es there is essen*ally no ISI so matched lter (bit by bit) detec*on is op*mal (for an AWGN channel).
One es*mate is that about 20% of the progress was due to advances in signal processing. However, advances in all elds were required to make the system work.
2007 Nobel Prize in Physics was awarded to the inventors of the GMR head
MFM FM
Peak Detection
PMR technology
- High anisotropy material - Ver:cal alignment of magne:za:on Writing is bit is possible - Much smaller due to flux leaking
from the write head to the disk. Reading is due from flux leaking from the disk to the read head.
41
100 nm
disk
Modula*on
Codes
The
purpose
of
modula*on
codes
is
to
prohibit
the
occurrence
of
certain
troublesome
sequences
such
as
sequences
which
cause
excessive
ISI
or
which
make
*ming
recovery
dicult.
The
most
well
known
example
of
a
modula*on
code
is
the
so-called
(d,k)
code,
where
no
run
of
0s
longer
than
k
or
less
than
d
is
permi\ed.
d
and
k
are
nonnega*ve
integers
for
which
k
>
d.
In
early
Gbyte
drives
(circa
1980),
(2,7)
and
(1,7)
codes
were
used.
Today,
varia*ons
on
(0,k)
codes
are
used.
Shannon
discussed
such
codes
in
the
very
beginning
of
his
1948
paper
in
a
sec*on
called
en*tled
The
Discrete
Noiseless
Channel.
Claude
Shannon
In
his
classic
1948
paper,
Shannon
showed
that
for
large
n,
the
number
of
length
n
constrained
sequences,
N(n),
is
approximately
2Cn.
The
quan*ty
C
is
called
the
capacity
of
the
constrained
system.
Said
in
another
way
The rate of a code, R, is the (average) ra*o of the number of unconstrained digits to constrained digits. Shannon showed that there exists codes at rate R, if and only if R < C.
For nite k, N(n) sa*ses the linear dierence equa*on: N(n)=N(n-(d+1))+N(n-(d+2))+ +N(n-(k+1)).
Second Method:
Shannon showed that the capacity is equal to the base 2 logarithm of the largest eigenvalue of the adjacency matrix of a graph which generates the code symbols. We illustrate these two methods for a (1,2) code (i.e., d=1 and k=2).
If d=1 and k=2, the equa*on xk+2 - xk+1 - xk-d+1 +1 = 0 becomes x4 - x3 - x2 +1 = 0. The largest real root of this equa*on is 1.3247 and its base 2 logarithm is 0.4057.
000
0010 0011
Coding theorists are also interested in 2-dimensional constrained binary codes: i.e., constrained binary arrays where the binary digits are arranged in an array of rows and columns. Such codes might have applica*on in 2-dimensional storage.
0
1
0
0
1
0
1
0
0
1
0
0
1
.
.
.
1
0
0
1
0
1
0
0
1
0
0
1
0
.
.
.
0
0
1
0
1
0
0
1
0
0
1
0
1
.
.
.
0
1
0
1
0
0
1
0
0
1
0
1
0
.
.
.
1
0
1
0
0
1
0
0
1
0
1
0
1
.
.
.
.
.
.
For 2-dimensional (d,k) constraints, C2 exists but Shannon didnt tell us how to compute it. To this day, for rectangular constraints, the exact value of the capacity is unknown except for the trivial cases where C2=0 or C2= 1.
ECC
Codes
Reed
Solomon
codes
are
used
in
todays
hard
disk
drives.
We
are
on
the
verge
of
seeing
the
introduc*on
of
LDPC
codes
with
itera*ve
decoding
in
HDD.
Future
Technology?
HAMR-Heat
Assisted
Magne*c
Recording
Pa\erned Media
In ordinary media, one can write a bit anywhere on the magne*c surface. In pa\erned media one must write each bit on a magne*c island. This is a dicult task since one cannot read and write simultaneously.
5 0 (written late) 0 1 1 1 0 0 0 0 0 Note that if the data bit written late is the same as the previous bit, there is no error in the recorded bit!!!
At *me i = 1, 2, 3, . . . Xi data bit {0,1} Yi recorded bit {0,1} Zi state of channel {0,1} Zi = 0 if data bit is wri\en on correct island Zi = 1 if data bit is wri\en late Then: Thus: Yi = Xi if Zi = 0 Yi = Xi-1 if Zi = 1. Yi = Xi (Xi Xi-1) Zi
Mathema*cal Model
Previous
Example
X Z Y
0
0
0
1
0
1
1
0
1
0
1
1
0
1
0
1
0
1
0
1
1
.
.
.
.
.
.
.
.
.
A decoder can decode this rate code with zero error probability just by observing the values of Y with even indices and thus the zero error capacity of this channel is at least . As a result, a lower bound to the capacity of the channel is independent of the sta*s*cal model assumed for the Z process.
1-p1,0
p1,0
Channel
Capacity
For
any
model
for
the
Z
process,
the
channel
capacity
is
dened
in
the
usual
manner.
We
call
the
capacity
of
the
Bernoulli
state
model
with
parameter
p,
CB(p),
and
we
call
the
capacity
of
the
Gilbert
state
model
with
parameters
(p0,1
,
p1,0),
CG
(p0,
1
,
p1,0).
For
the
Bernouli
state
model
one
can
prove
that:
CB
(p)=CB
(1-p)
and
for
Gilbert
state
model,
one
can
prove
that:
CG
(p0,1
,
p1,0)
=
CG
(p1,0
,
p0,1).
Details are given in the paper Write Channel Model for Bit-Patterned Media Recording which will appear in the IEEE Transactions of Magnetics.
Bernoulli
Model
The
parameter
space
can
therefore
be
reduced
to
the
interval
p
[0,
1/2
].
Furthermore
the
same
symmetry
argument
holds
for
not
just
the
rate-maximizing
input
distribu*on,
but
for
all
input
distribu*ons.
The
capacity
of
the
Bernoulli
model
is
upper
bounded
by
the
achievable
rate
for
a
genie-aided
decoder,
i.e.,
one
with
the
{Z}
process
realiza*on
known.
The capacity of a correlated symmetric erasure channel is the same as that of a memoryless symmetric erasure channel with the same erasure probability. Therefore, CB (p) < [1 p(1-p)].
Example
Z Y X
0
0
0
1
0
1
1
1
1
0
0
1
.
.
.
.
.
.
Example
Z Y X
0
0
0
0 1
0 1
1 1
1 0
0 1
. . . . . .
Example
Z Y X
0
0
0
0
1
1
0
1
1
1
1
0
0
1
.
.
.
.
.
.
Example
Z Y X
0
0
0
0
1
1
0
1
1
1
1
1
0
0
1
.
.
.
.
.
.
Example
Z Y X
0
0
0
0
1
1
0
1
1
1
1
0
1
0
0
1
.
.
.
.
.
.
Example
Z Y X
0
0
0
0
1
1
0
1
1
1
1
0
1
0
?
0
1
1
.
.
.
.
.
.
Bernouli
Model
The
input
distribu*on
that
maximizes
the
mutual
informa*on
is
unknown
so
no
closed
form
expression
has
been
found
for
the
capacity
of
the
Bernouli
model.
However
lower
bounds
on
the
capacity
can
be
found
by
assuming
par*cular
forms
for
the
input
distribu*on
(e.g.,
an
i.i.d.
input
process).
Very
*ght
upper
and
lower
bounds
have
been
found
for
the
symmetric
informa*on
rate
when
the
input
is
uniform
and
i.i.d.
An
accurate
es*mate
of
the
symmetric
informa*on
rate
has
been
obtained
using
the
BCJR
algorithm.
Bernouli
Model
To
explore
the
loss
in
achievable
rate
due
to
an
i.i.d.
input
we
considered
a
symmetric
rst
order
binary
Markov
input
where
Pr{Xi
=
1|Xi-1
=
0}
=
Pr{Xi
=
0|Xi-1
=
1}
=
.
Upper
and
lower
bounds
were
found
for
the
mutual
informa*on
for
a
Markovian
input
as
a
func*on
of
and
p.
Comparison of the Symmetric Informa*on Rate and the Informa*on Rate for a Markovian Source
It was found that for the Bernouli model for Z, considerable gains in the reliable transfer rate are possible by using an input with memory.
Gilbert
Model
We
know
less
about
compu*ng
the
capacity
for
this
model
than
for
the
Bernouli
model
By
using
a
genie
to
inform
the
decoder
of
the
Z
process
we
can
obtain
an
upper
bound
to
the
capacity.
Again
the
result
is
a
correlated
erasure
channel
with
average
erasure
probability
Pr{Zi-1=1,
Zi=0}
=p1,0
p0,1/(p1,0+p0,1)
resul*ng
in
the
upper
bound
for
the
capacity:
CG(p0,1
,
p1,0)
<
1
[p1,0
p0,1/(p1,0+p0,1)].
inser*on dele*on
91
Flash
Memory
Flash
is
a
non-vola*le
memory
which
is
fast,
power
ecient
and
has
no
moving
parts.
Electrically
programmed
and
erased.
Used
in:
Digital
cameras
Low
capacity
IPODS
Mobile
phones
Laptop
computers
Hybrid
drives
92
94
16.
In
order
to
reduce
a
cell
level,
all
the
cells
in
that
block
must
be
reset
to
level
0
before
rewri*ng.
A
VERY
EXPENSIVE
OPERATION
95
Wri*ng
Write
sequen*ally
to
the
next
available
page.
Erasing
Can
only
erase
an
en*re
block!
Page 63 Page 64
!
96
The
Goal:
Represen*ng
the
data
eciently
such
that
block
erasures
are
postponed
as
much
as
possible.
97
Experiment
Descrip*on
For
each
block
the
following
steps
were
repeated:
The
block
was
erased.
Pseudo-random
data
was
wri\en
to
the
block.
The
data
was
read
and
compared
to
nd
errors.
Remark:
The
experiment
was
done
under
lab
condi*ons.
Other
factors
such
as
temperature
change,
intervals
between
erasures
and
mul*ple
readings
before
erasures
were
not
considered.
98
99
!10
6
100
!10
5
An
Introduc*on
to
WOM-codes
WOM-codes
allow
us
to
write
several
*mes
to
the
same
block
of
memory
without
erasing.
Example:
In
1982,
Rivest
and
Shamir
found
a
way
to
write
2
bits
of
informa*on
twice
using
only
3
cells.
We
denote
a
WOM
code
that
writes
k
*mes
to
n
cells
as
a
<V1,
V2,
,
Vk>/n
code
where
Vi
is
the
number
of
messages
wri\en
on
the
ith
write.
Thus
the
Rivest
Shamir
code
is
a
<4,4>/3
code
with
k=2.
101
102
If we want to write the same data on the second write, we do not change what is wri\en on the rst write.
Note that when going from rst write to second write, no 1s are erased.
Rate
of
a
WOM-code
The
rate
of
the
ith
write
is:
Bits
of
informa*on
Ri
=
Total
number
of
bits
log2(Vi)
Ri
=
n
The
total
rate
of
a
WOM-code
is
R
=
(Ri).
The
Rivest
Shamir
code
has
R1
=
R2
=
2/3
and
R
=
4/3.
103
104
Mul*ply the stored vector e1 e2 by H: H(e1 e2) = He1 He2 = s1 (s1 s2) = s2
On the rst write, we program only vectors v such that rank(H (v)) = 3, V1 = { v {0,1}n | rank(H(v)) = 3}. For H as shown above, |V1| = 1 + 7 + 21 + 35 + (35-7) = 92. Thus, we can write one of 92 messages at the rst write. Encoding and decoding of the rst write are done with a lookup table. Say we write e1 = 0 1 0 1 1 0 0.
For a vector v {0,1}n, let H(v) be the matrix H with 0s in the columns that correspond to the posi*ons of the 1s in v.
111
Time
Sharing
If
we
know
two
codes
with
rates
(R1,
R2)
and
(R3,
R4),
we
can
achieve
any
rate
pair
(t*R1
+
(1-t)*R3,
t*R2
+
(1-t)*R4)
for
t
a
ra*onal
number
between
0
and
1.
112
Some
Achievable
Rate
Pairs
and
Capacity
for
WOM
With
Two
Writes
113
More Achievable Rate Pairs and Capacity for WOM With Two Writes
114
the ques*on: Can an Informa*on Theorist Be Happy in a Center for Informa*on Storage? is a resounding yes.
Paul Siegel
Aravind Iyengar
Eitan Yaakobi*