Rochester Institute of Technology
RIT Scholar Works
Theses
Thesis/Dissertation Collections
8-1-1996
M-ary run length limited coding
Jian Luo
Follow this and additional works at:
http://scholarworks.rit.edu/theses
This Thesis is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion
in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact
ritscholarworks@rit.edu.
Recommended Citation
M-ary
Run Length Limited
Coding
by
Jian Luo
A
thesis
submitted
in
Partial Fulfillment
of
the
Requirements for
the
Degree
of
MASTER
OF SCIENCE
in
Electrical
Engineering
Department
of
Electrical
Engineering
College
of
Engineering
Rochester
Institute
of
Technology
Rochester,
New York
M-ary
Run Length Limited
Coding
by
Jian Luo
A Thesis
submitted
in
partial
fulfillment
of
the
requirements
for
the
degree
of
Master
of
Science
in
Electrical
Engineering
Approved
by:
Professor
S.W.McLaughlin,
Thesis
Advisor
Professor
J. DeLorenzo
Professor
S.A.Dianat
Professor
R.Unnikrishnan,
Department Chair
Department
of
Electrical
Engineering
College
of
Engineering
Rochester
Institute
of
Technology
Rochester,
New York
August 1996
Thesis Release Permission
M-ary Run Length Limited Coding
I, Jian Luo, hereby permission to the Wallace Memorial Library of Rochester Institute of
Technology to reproduce my thesis in whole or in part. Any reproduction will not be for
commercial use or profit.
Signature:
Dedication
Dedicated
to
the
loving
memory
ofmy
motherTifei
Fang
andmy father Guangyu Luo.
Acknowledgments
I
wouldlike
to
expressmy
sincere appreciation and gratitudeto
my
advisor,
Dr.
S.W.McLaughlin,
for his
guidance, patience,
understanding,
encouragementandfinancial
support as well as
friendship
throughout
my
graduate study.I
wouldlike
to thank
my
thesis
committeefor
taking
the time to
participate andfor
their
valuable advice.
I
wouldlike
to thank
Dr. Joe
DeLorenzo,
Dr.Ed Salem
andDr.Soheil Dianat
for
their
time, helpful
advice andsuggestions,
andquality
education.I
wouldlike
to thank
Dr. Raman Unnikrishnan
for
encouragement,
andhelpful
advice.My
specialthanks to
my husband
Yingchong
Cheng,
for his love
andmy
newlife
he
Abstract
This
thesis
consists oftwo
parts:1)
a result ofthe
rationality
of channelcapacity for
the
M
-ary(d,
k)
constraint,
and2)
a software package whichhas been designed
to
aidin
the
development
ofefficient,
high
density
channel encodersfor
M
-ary(d,
k)
constrainedrecording
code.It
has been
writtenfor
the
platforms ofIBM PC
orPC
compatible andSun Workstations. The designed
softwarein
this thesis
involves
three
steps.The
first step
is
to
calculatethe
capacity
of aRLL
code afterthe
userhas
chosen(M,d,k)
.The
channelcapacity
specifiesthe
maximum ratio ofinformation
(
p
)
bits/codeword
(
q
)
bits
achievable
in
implementing
these
codes and gives options of code choicefor any
specifictask.
The
secondstep is
to
find
a codeefficiency
r\
=RI C
by
choosing
p
andq
, anddisplay
the
statetransition
matrix offinite
statetransition
diagram
andfinite
statetransition
diagram
ofthe
code.The last step is
to
aidthe
userin
the
generation of acode,
by
using
statesplitting
algorithm ofR.
Adler,
D. Coppersmith
andM. Hassner [10].
Table
of
Contents
List
ofFigures
ix
List
ofTables
xi1
Introduction
1
1
.1
Overview
1
1
.2Introduction
ofRLL
Code
4
1
.3
FSTD
ofM
-ray
(
d
,k
)
constraint6
1.4
Conclusion
9
2
On
the
capacity
ofM
-ary( d,
k)
codes10
2.1
Introduction
andBackgroud
10
2.2
The
Key
Lemma
14
2.3
Capacity
ofM-ary (d,k)
Codes
15
2.4
Code Constructions
19
2.5
Conclusion
23
3
Coding
Method
24
3.1.
Storage
Density
D
24
3.2
Three
Coding
Methods
26
3.3
Conclusion
35
4.1
State
Splitting
36
4.2
State
Merging
48
4.3
Examples
ofState
Splitting
andMerging
5 1
4.4
Conclusion
56
5
Softare Design
57
5.1
Architecture Design
57
5.2
Running
the
Software
62
5.3
Conclusion
69
List
of
Figures
Figure 1-1
Block diagram
ofrecording
system2
Figure 1-2
Finite State
Transition
Diagram for
aM
-ary( d, k)
constraint code
7
Figure 1-3
FSTD for
(4,1,3)
constraint code8
Figure
2-1
The
(M,d,k)
constraint1 2
Figure 2-2
The
(
M,
d,
)
constraint1 2
Figure 2-3
The
(3,1,
oo)
constraint19
Figure 2-4
The
(3,1,)
encoder20
Figure 2-5
The
(
1 3
,1
,)
constraint2 1
Figure
2-6
The
(
1 3
,1
,)
encoder2 1
Figure
2-7
The
(5,2,
oo)
constraint22
Figure 2-8
The
(5,2,
oo)
encoder23
Figure
3-1
The
(2,
1
,
)
constraint29
Figure
3-2
The
(2,1,3)
constraint31
Figure 3-3
Second
powerofFSTD in Figure 3-2
32
Figure 3-4
The
(4,1,3)
constraint34
Figure 4-1
(a)
FSTD
for RLL (0,1).
Figure 4-2
(a)
Reducible FSTD for
unconstrainedbinary
sequences(b)
Irreducible
components37
Figure 4-3
(a)
FSTD
for
a particular charge constraint(b)
FSTD for
second power ofFSTD
in
(a)
38
Figure 4-4
The
(2,0,1)
constraint39
Figure 4-5
The
(2,1,3)
constraint40
Figure 4-6
Second
power of(2,0,1)
constraint42
Figure
4-7
The basic
v-consistent statesplitting
of(4,1,3)
constraint45
Figure 4-8
After
mergethe
(4,1,3)
constraint50
Figure
4-9
The
(2,1,3)
encoder51
Figure
4-10
Merger
ofRLL
(2,
1
,3)encoder5 1
Figure 4-1 1
The finite FSTD
of(2,1,3)
encoder52
Figure 4-12
The
(4,3,4)
constraint52
Figure 4-13
Second
power ofFSTD in Figure 4-12
52
Figure 4-14
The splitting
graph of(4,3,4)
constraint54
Figure
4-1
5
The
finite FSTD
of(4,3,4)
encoder56
Figure
5-1
The Design
ofRLL Software
57
List
of
Tables
Table
1
-1
Basic coding
table
(2,11)
RLL
code6
Table
2-1
The
Capacity
ofBinary
(
d,
k)
constraints13
Table
2-2
The
Capacity
ofM
-ary(
d,
k)
constraints19
Table
3-1
Achievable densities (in bits I
T^n)
Dc
for
M-ary
(d,
k)
codes26
Table
3-2
Number
ofdistinct
(
d
)
sequences as afunction
ofthe
sequencelength
n andthe
minimumrunlengthJ
as a parameter27
Chapter
1
Introduction
1.1
Overview
Coding
techniques
arewidely
usedin digital recording
devices,
such asthe
digital
videoand audio compact
disk,
floppy
disk drives
and electrical and opticlinks between
variousdigital
machines.The coding
techniques
include coding
equipment andcoding
methods.The coding
methods are usedin
communication systemsto
improve
the
reliability
andefficiency
ofthe
communication channel.The reliability is commonly
expressedin
terms
of
the
probability
ofreceiving
the
wrong
information,
that
is,
the
probability
that the
information
differs from
what wasoriginally
transmitted.
Error
controlis
concerned withtechniques
ofdelivering
information
from
a source(the
sender)
to
adestination (the
receiver)
with a minimum of errors.As
suchit has its
rootsin information
theory
andgoes
back
to
the
pioneering
work ofShannon
[2].
The
fundamental
theorem
ofinformation
theory
statesthat
"it
is
possibleto
transmit
information
through
anoisy
channelat
any
speedless
than the
channelcapacity
with anarbitrarily
smallprobability
oferror".
On
the
otherhand,
if
the
transmission
speed exceedsthe
channelcapacity, the
error
probability
cannotbe
reducedto
zero.Information
theory
partitionsthe
coding
speaking,
atechnique to
reducethe
source symbol rateby
removing
the
redundancy
in
the
source signal.
A
convenientdefinition
of channelcoding is: The
technique
ofreliability
despite
shortcomings ofthe channel,
whilemaking
efficient use ofthe
channel capacity.In
essence,
the tenets
ofinformation
theory
showthat
astationary
channel canbe
madearbitrarily
reliable giventhat
afixed fraction
ofthe
channelis
usedfor
redundancy.In
digital recording
systems,
channelencoding is commonly
accomplishedin
two
successive steps:
(a)
error correction code and(b)
recording (or
modulation)
code.The
various
coding
steps are visualizedin Figure
1-1,
which shows ablock diagram
of apossible
recording
system ofthis
kind.
Error-correction
is
realizedby
systematically
adding
extra symbolsto the
conveyed message.These
extra symbols makeit
possiblefor
the
receiverto
detect
and correct some ofthe
errorsthat
may
occurin
the
receivedmessage.
A recording
code plays avariety
ofimportant
roles.Its
purposeis
to
improve
the
performance ofthe
systemby
matching
the
characteristics ofthe
recorded signalsto
those
ofthe
channel.In
otherwords,
it
acceptsthe
bit
stream(extra bits
addedby
the
error correction system
included)
asits input
and convertsthe
streamto
a waveformsuitable
for
the
specific recorder requirements.The
object ofthe
recording
codeis
to
bring
structureinto
adata
streamthat
is generally
not presentin
the
information
suppliedby
the
user.1 1
Isc-r
r i i i 1 1 Data Eiror Correction Code Recording Code Channel Recording Code Error Correction Code Data r i i ir
Ercoders D:coders
i i i
Most
magnetic and opticalrecording
system use saturation-type media wherethe
input
to
the
recording
channelsis
atwo-level
waveform(binary).
In
saturationrecording
a numberof
factors
and channel constraintsdetermine
the
recording
code,
such asdetection
method,
timing
generation andrecovery,
and channelbandwidth.
Many
magnetic and alloptical
recording
systems use peakdetection
[1]
during
the
readbackstep
to
recoverthe
information.
However,
the
recording
channelhas
imperfections,
such asthermal
noisegenerated
by
the
electronic circuits and noisearising from
magnetic properties ofthe
disk
medium can corrupt
the
readbacksignal,
leading
to
spurious peaks as well as shiftedpositions of
the
genuine peaks.To
compensatefor
these
effects,
atleast
in
part,
enhancements
to
peakdetection
the
addition of athreshold
(clip
level)
andthe
tracking
ofpeak polarities
have been
used.Another
major problemis
the
intersymbol
interference
(ISI)
ofneighboring
pulses.Magnetic transitions,
orrecording
symbols"1"
which are
written
too
closeto
one anotherhave
aninterfering
effectduring
readbackthat
canboth
reduce
the
amplitude ofthe
pulse as well as shiftthe
peak position.On
the
otherhand,
if
the transitions
are writtentoo
far
apart, the
clockrecovery
circuit will notbe receiving
adequate
information
to
maintain synchronization withthe
recordedbits.
These detection
error mechanisms are combatedby
the
use ofdata coding
techniques.
Two distinct kinds
ofcodes aretypically
used.A recording
codeis
usedto
guaranteethat
the
recordedbit
sequence codedoes
not sufferfrom
the
problemdescribed
above,
namely
runs of symbols
"0"
between
consecutive symbols "1"which are either
too
short ortoo
of channel
noise,
producing
an unacceptable error rate atthe
desired data density. The
second
kind
of codeis
anerror-correcting (also
called errorcontrol)
code(ECC),
useredundant
bits
whichhave been
addedto
the
original userdate
to
detect,
locate
andcorrectall
remaining
errors withvery high probability
of success.1.2 Introduction
ofRLL Code
A widely
used constraintfor
the
peakdetection
channelin recording
systemis
calledrun-length-limited
code(RLL)
or(d,k)
constraint.(d,k)
represent minimum andmaximum number of symbols "0"
between
consecutive symbols"1"
respectively in
the
allowable code strings.
The
parameterd
controlshigh
frequency
contentin
the
associatedwrite current
signals,
and so reducesthe
effect ofintersymbol interference. The
parameterk
controlslow
frequency
content,
and ensuresfrequent information for
the
clocksynchronization
loop. In
orderto
construct asimple,
efficientrecording
codes whichthe
peak
detector
can moreeasily
recoveredcorrectly,
the
code rateis
avery important
parameter
in RLL
codes.The
conversion ofarbitrary
stringsto
constrained(d,k)
stringscouldbe
accomplishedasfollows. Pick
a codewordlength q
.List
allthe
string
oflength q
whichsatisfy
the
(d,
k)
constraint.
If
there
areatleast
2P suchstrings,
assign an uniquecodewordto
each ofthe
2P
possible
binary
input
words oflength
p.This kind
of codemapping is
commonly
the
coderate,
here
we useR
= plq.Shannon
provedthat
the
rateR
of a constrainedcode cannot exceed a
quantity
[2],
here
referredto
asthe
Shannon capacity C
,that
depends only
onthe
constraintbut
canbe arbitrarily
closeto the
capacity.Since
there
areonly
2q unconstrainedbinary
strings oflength q
,there
willbe less
than
this
number ofconstrained codewords.
Therefore,
the
rate mustsatisfy
R<\.
In
fact,
there
is
amaximum achievable rate called
capacity C
.According
to
Shannon's information theory,
the
rateR
ofany
codemapping for
that
constraint mustsatisfy R
=p
I q
<C
.In
otherword,
if
a codehas
a rateR
=plq
is
closeto
C
,then that
codeis
calledefficient,
sothe
codeefficiency
y\
=RIC
Example 1-1:
(d,k)
=(2,11),
M
=2
andR
=1/2
RLL
code.This
code wasinvented
by
Jacoby[3],
is
also called3PM
code.The encoding is
explainedby
look up Table 1.1.
C
*0.545
>R
=0.50
,
r|
=R
/ C
91.74%
.Because
the
information data block length
p
=3,
the
number of constrained codewords must greater or equalthan
2Pwhichis 8 in
Data
Codeword
000
000010
001
000100
010
010000
Oil
010010
100
001000
101
100000
110
100010
111
100100
Table 1-1. Basic coding
table(2,
1
1)
RLL
code.As
wemay
noticefrom Table
1-1,
the
right-headboundary
positionis
occupiedby
"0"s
in
allcodewords.If
the
juxtaposition
of codewords would offendthe
d
=2
condition,
that
is,
if both
the
symbol at position5
ofthe
present codeword andthe
symbol at position1
of
the
next codewors are"1",
the
"merging" symbol at position6 is
setto
"1" andsymbols at position
1
and5
are setto
"0". The encoding logic
canbe very
simpleby
implementing
the
described merging
rule.The capacity C
we usedin
the
example willbe discussed in
next chapter.1.3 FSTD
ofM
-ray(d,k
)
constraintWhile
there
has been
a greatdeal
ofcoding
workin
(
d,k
)
codesfor
binary
recording
channelsfl]
there
is very little
workin nonbinary
runlength(
M
-ary,M
>3)
constrained
media
that
permitted multiamplitudeinput
waveforms(M
>3)
whilemaintaining
channel
linearity.
Now
that
such a channelis
being
used,
wecandesign
codesthat
exploitthe
multi-amplitudecapability
ofthe
media,
alarger information
density
onthe
channelcan
be
achieved.The finite
statetransition
diagram
(FSTD)
has been
usedto
representany M
-ray(d,
k)
(when M
=2
,
is
binary)
constraintby
using
afinite
number ofstates andedges,
and edgelabels containing
afinite
number of symbols.Figure 1-2
showsthe
FSTD.
M-l
Figure 1-2. Finite State Transition Diagram for
aM
-ary(d,k)
constraint code.The FSTD
consists a graph withk
+
1
nodes,
calledstates,
anddirected
edgesbetween
them,
called statetransitions
representedby
arrows,
it indicates
each edge canbe
traversed
in only
onedirection. The
edgeslabeled
with channelbits,
anddistinct
edgesthat
have
the
samestarting
andending
states will alwayshave distinct labels. Self-loops
are
the
edgesthat
start and endatthe
same state.The adjacency
matrixis defined
asstatetransition
matrixT
which givesthe
number of ways ofgoing (in
onestep) from
stateai
to
statecij
.The
matrixis
a(k
+
1)
x(k
+
1)
array T
withentries i\. whichis
the
numberof edges
from
statei
to
statej
in
FSTD,
wheref,0=l,2,...,M-l
i>d,
hj=o,
elsewhere.Then
the transition
matrixfor k
<o asillustrated:
T
=0
1
0
0
0
M-l
0
1
0
0
M-l
0
1
M-l
0
There
are(k
+1)
states.The
transmission
of a "0"takes the
sequencefrom
state0
to
statej
+1
, a "1,2,
...,M
-1
"
may be
transmitted
from
stated,...,k.
As
aexample,
we can writethe
FSTD
andthe
corresponding
transition
matrixfor
(M,d,k)
=(4,1,3)
constraints:the transition
matrixis:
Figure 1-3. FSTD for
(4,1,3)
constraint code.r=
0
1
0
3
0
1
0
3
0
0
1
3
0
0
0_
The finite-state
machinemodel allowsusto
computethe capacity,
andit
is
very
helpful
to
state-transition
matrix whoseij
-thentriesthat
givethe
number ofdistinct
sequencesfrom
statei
to
statej
that
areq
symbolslong. If
the
singlestep
state-transition
matrixis T
,
then
the
q
-step state-transition matrix equals Tq whichis
the
number of sequencesis
givenby
the
ij
-thentry
ofV
.1.4 Conclusion
In
this
chapter wehave
given anintroduction
to the
problem ofcoding for
an(
M,
d,k
)
constrained channel.
In
the
next chapter we will give a result onthe
rationality
ofthe
capacity for
all(
M,d,k
)
constraints.This
resultis
important,
because
as we will see withthe
statesplitting
algorithm presentedlater,
constraints with rationalcapacity have
the
potentialof
having
100%
efficient(i.e. capacity
Chapter
2
On the capacity
of
M-ary
(
d,k
)
codes
In
this
chapter a result onthe
capacity
ofM
-ary(d,k)
codesis
presented.This
chapterappeared
in
the
IEEE Transactions
onInformation
Theory
[23].
2.1 Introduction
andBackground
Traditional
magnetic and opticalrecording employ
saturationrecording,
wherethe
channel
input is
constrainedto
be
abinary
sequencesatisfying
run-lengthlimiting
(RLL)
or
(d,k)
constraints.A
binary
(d,k)
sequenceis
one wherethe
number of zeroesbetween
consecutive onesis
atleast d
and at mostk
.If
arecording
channel permitted anM
level
input
waveform(M>3),
alarge
information
density
onthe
channel mightbe
achieved.Until recently
there
were nooptical or magnetic media
that
permitted multi-amplitude waveforms whilemaintaining
channel
linearity. The recording
mediain
[9]
supportsunsaturated,
M-ary
(M>3)
signaling
whilerequiring
that
run-length-limiting
constraintsbe
satisfied.Assuming
anM-ary
symbolalphabet,
A
={0,1,...,
M-l], M<o,
an(M,d,k)
sequence[12, 13]
is
one where atleast d
and at mostk
zeroes occurbetween
nonzero symbols.
Binary
(d,
k)
code areM
-arycodes withM
-2.In
[7]
(and
the
applicable correctionsin
[8])
it
was shownthat
for
binary
(d,
k)
codes,
there
existno100-percent
efficient codes.Specifically
the
Shannon capacity
ofallbinary
RLL
(d,k)
constraintsis irrational for
all values of(d,k),
0<d
<k<,
andhence,
there
exist nofixed
rate codesthat
achieve capacity.In
this
chapter we provetwo
propositions on
(M,d,k)
codes.First
we showthat
for any integer
M,
100-percent
efficient
fixed-rate
codes areimpossible for
all values of(d,k),
0<d
<k<,
therefore
extending
[7]
to the
M-ary
case.Secondly,
unlike[7],
for k
=oo we showthat there
do
exist
(an
o numberof)
100-percent
efficientcodes,
and we constructthree
such codesusing
the
statesplitting
algorithm[10].
The RLL
(M,d,k)
constraintis
often representedby
afinite
statetransition
diagram
(FSTD)
G
given(for
<o)
in Figure 2-1. The FSTD for
the
(M,d,k
)
is
shownin
Figure 2-2.
Any
sequencesatisfying
the
(
M,d,k
)
constraintis conveniently described
by
reading
the
labels
off pathsthrough
G
.Associated
to the
FSTD
withk
+1
verticesis
astate-transition matrix
T,
a(k
+1)
x(k
+
1)
matrixdefined
by
T
=[tij]
wheretr
is
the
numberofedges
in G from
statei
to
statej
.M-l
Figure 2-1. The
(M,d,k)
constraint.M-l
Figure 2-2. The
(M,d,)
constraintShannon
[11]
showedthat
if
the
FSTD G has distinct labels
onthe
outgoing
edges ofeach
state, the
capacity
of systemS
constrainedby
sequencefrom G is
C(S)
=log2
X
(bits
persymbol)
(2-1)
where
X
is
the
largest
realeigenvalue ofthe
adjacency
matrixT
, associated withG
.An
equivalent statement
[12]
is
that
X
is
the
largest
real root ofthe
characteristic equationP(Z)
=zk^
-zk+l -
(M
-\)z*-*+l
+
M
-1
=0
(2-2)
We have
assumedbase-2 logarithms in
(2-1)
because it is
the
most common case.While
We
considerfixed
rateR
=plq
encodersthat
map p
userbits
to
q
channel symbolssatisfying
the
(
M,d,k
)
constraints.The capacity
C(S)
is
therefore
an upperbound
to
allachievablerates
R
, andthe
codeefficiency
r\
=R I C is
the
ratio ofthe
actualcoding
rateto
the
largest
rate achievable.The
codedesign
problemis,
for
afixed
M,
d,
andk,
to
construct a code
that
mapsp
userbits
to
q
channel symbolssatisfying
(M,d,k)
constraints.
It is desired
that the
codehave
rate closeto
C
(i.e. r\
closeto
1)
withreasonable complexity.
Marcus
[14]
andAdler
et. al.[10]
developed
a code constructionalgorithm
known
asthe
statesplitting
algorithmfor
abroad
class of constraintsincluding
(d,k)
and(M,d,k)
constraints.The
algorithm produces a code with afinite-state
machine encoder and
sliding block (state
independent)
decoder
atany
rateR
=p
I
q<C.Code
constructionfor
binary
codeshas been
addressedin
[15]
andfor
M-ary
codesin
[9,13,16].
Table 2-1 lists
some commonM
=2
,
(d,k)
code capacities.k
d=0
d=l
d=2
d=3
d=4
d=5
d=6
2
.8791 .40753
.9468 .5515 .28784
.9752 .6174 .4057 .22325
.9881 .6509 .4650 .3218 .18236
.9942 .6690 .4979 .3746 .2669 .15427
.9971 .6793 .5174 .4057 .3142 .2281 .13358
.9986 .6853 .5293 .4251 .3432 .2709 .19939
.9993 .6888 .5369 .4376 .3620 .2979 .238210
.9996 .6909 .5418 .4460 .3746 .3158 .263311
.9998 .6922 .5450 .4516 .3833 .3282 .280412
.9999 .6930 .5471 .4555 .3894 .3369 .292413
.9999 .6935 .5485 .4583 .3937 .3432 .301114
.9999 .6938 .5495 .4602 .3968 .3478 .307415
.9999 .6939 .5501 .4615 .3991 .3513 .3122oo
1.0000
.6942 .5515 .4650 .4057 .3620 .3282Table
2-1. The
Capacity
ofBinary (d,k)
constraints.In
the
next section we reviewthe
key
results of[7]
anddiscuss
their
applicability
to this
correspondence.
In Section 3
we provetwo
propositions onthe
Shannon
capacity
of(M,d,k)
codes andin Section 4
constructthree
100-percent
efficient codesthat
have
low complexity
encodersanddecoders.
2.2 The
Key
Lemma
In
this
section we use alemma from
[7]
(and
the
correctionsto
it in
[8])
to
determine
the
rationality
of(M,d,k)
capacities.Before
discussing
the
key
lemma
we givetwo
definitions
whichare alsodraw from [7].
Definition
1: A
nonnegative nxn real matrixT
is irreducible
if,
for any
\<i,j<n,
there
is
aninteger
m suchthat
(r%>0.
(2-3)
A directed
graphG
is irreducible if its
statetransition
matrixT
is irreducible.
Definition
2: The
period of anirreducible
nonnegative nxn real matrixT
is
the
greatestcommon
divisor
(gcd)
ofcyclelengths,
that
is,
gcd{fc:(T%
>0,
for
somei\
<i<n).
(2-4)
The
period of anirreducible directed
graphG is
the
period ofits
statetransition
matrix.The
matrixT (or
graphG
)
is
called aperiodic whenthe
periodis
one.We
now presentLemma
1
(Ashley
andSiegel)
:Let S be
a constrained system ofsequences,
representedby
anirreducible
FSTD G
withdistinct labels
onthe
edgesemanating from any
singlestate.
For base
b,
let I be
the
largest
nonnegative suchthat
b
=al
for
some positiveinteger
a.If
the
base b
capacity
ofS
satisfiesCb(S)
=plq
(2-5)
where
gcd( p I q
)
=1,
then
the
period ofG is
amultipleof qlf,
where
t
-gcd(l,q)
.Discussion. This lemma
establishes anecessary
condition onthe
periodicity
ofanFSTD
whose
capacity is
rational.Ashley
andSiegel
usethis
key
lemma
to
prove,
by
contradiction,
that the
capacity
ofbinary
(d,
k) (aperiodic)
constraintis irrational for
all(d,k),
0<d
<k<.
Thus
since practicalfixed
rate codes operate at rational rates
R-p
I
q
,there
exist no100-percent
efficientbinary
(d,k)
codes.Note
that
for
the
(
M,
d,k
)
constraint,
its FSTD G is irreducible
andaperiodic,
whichis
easily
shownby
noting
that
for any
statei
there
is
always a path(passing
through
state0)
of no
longer
than
d
+
k
+
2 from
statei
to
statei
.Aperiodicity
follows from
the
fact
that
G
containscycles of(d
+
1)
and(d
+
2)
, and gcd(d+
1,
d
+
2)
=1
.2.3
Capacity
ofM-ary (d,k)
Codes
In
this
section we presenttwo
propositionsthat
establish conditions onthe
rationality
ofcapacity
ofM
-ary(d,k)
codes.The
proofsrely
heavily
onLemma
1,
but
are quitedifferent
than
the
proofsin [7].
Denote
the
binary
capacity
ofthe
(
M
,d
,k
)
constraintasC(M,d,k).
Proposition 1:
C(M,
d,
k
)
is irrational for
all(M,d,
k
),
0<
d
<k
< ,M<.
Proof.
By
Lemma
1,
if
C(M,
d,k)
=piq
,then
the
periodicity
ofG is
a multiple ofq 1
1
.Since
t
=gcd(l,q)
andb
= 2'= 21
this
implies
t
=1
.Since G is
aperiodic,
q
=1
.Therefore
the
largest
eigenvalue ofT
mustbe
ofthe
form
X-2plq =2Pand
it
mustbe
asolution
to the
characteristic equation(2-2). We
shall showthat
X
=2P cannotbe
a rootof
(2-2)
andhence,
the
capacity
cannotbe
rational.JfX
=2pis
a root ofzk+2 _
zm _
(M
_l)zk-d+l
+
M
_ 1=0
(2_6)
then
the
product ofthe
remaining
roots-Y\=
xt
IS sucntnat
2" =
(M
-1).
Substituting
(|>into
(2)
yields(2p)k+1 +2P
=0
2P(2pk+p
+(]))
=0
(2-7)
Since
2P*
0
,(after collecting
terms)
2pk(2p-l)-ty(2p(k-d+1)-l)
=0
(2-8)
2**(2" -1) (h =
z
^
~v>(2-9)
Simplifying
this expression,
weget2P*(2'-1)
<!> =p(t-d+l) OP i TP
2p+2"-l
)pk/r,p2"'C(2;,-1)
2p(2p(k~a) -\)+
2p -I(2-10)
Since
2p(*"d)-1
=(2P-l)(X,-"0
2"').
2pk(2"-\)
<l> =...^k-d-i.
2p[(2p-l)(Xi.;o"2'")]
+(2''-l)
\pk
2'[(5^2")]+l
(2-11)
Using
this
form
of <|> we recallthat
if
2Pis
a root of(2-2),
then
<j)2p=
M
-1
, which
implies
that
2p2pk
<t>2"= .
d_, ,
(2-12)
must
be
aninteger. This is
not possible sincethe
denominator is
odd,
andthere
are noodd
factors
ofthe
numerator.Therefore
2Pis
not an eigenvalue ofT
andthe
capacity
is
not rational.
This
completesthe
proof ofthe
proposition.Discussion of
Proposition
1. The
proof ofthis
propositionfor
non-base-2 capacitiesproceeds similarly.
When b
=al
and a
is
eitherodd,
or even with no oddfactors,
the
proofis
identical.
When
ais
even andhas
oddfactors,
one more simplificationstep is
needed after(2-12)
before
the
proofis
complete.Next
we considerthe
case wherek
=o,namely
the
maximum run-length of zeroesis
unconstrained.
An FSTD H for
(
M,d,k
)
constraintis
shownin Figure 2-2. It is easy
to
show
that
H is irreducible
and aperiodic.As
before,
the
capacity is
computedusing
the
largest
eigenvalue ofthe
adjacency
matrixT
ofH
, whichis
equivalentto
finding
the
largest
positive root ofthe
characteristic equation[6]
pM)
=zd+l-zd-(M-l)
=0.
(2-13)
Proposition
2: For any
0<d<o;
andk
=the set ofM's
for
whichC(M,d,)
is
rational
is
{
M:M
=2*(2P-1)
+1,
integer
p>\}.Proof:
Since
the
FSTD for
an(M,d,)
constraintis irreducible
andaperiodic,
Lemma 1
applies and
the
largest
real root of(2-13)
has
the
form
X
=2P .Substituting
into
(2-13)
yields
2pd(2p-\)
=M-\
(2-14)
which
implies (for any d
>0
andany integer p
>1
)
there
exists aninteger M
satisfying
(2-13).
Discussion
of
Proposition
2. The
proof ofthis
propositionfor
non-base-2 capacitieseasily found for any d
andbase b
.Table 2-2 listed
the
capacity
ofM
-ary( d,
k
)
constraint.(d,k)
M=2
M=3
M=4
M=5
M=6
M=7
M=8
(0,2)
.87911.5458
1.9820
2.3125
2.579
2.8037
2.9975
(0,3)
.94701.5726
1.9957
2.3200
2.584
2.8068
2.9997
(1.3)
.5515 .52551.1540
1.3210
1.4528
1.5622
1.6557
(1,7)
.6793 .99621.2018
1.3563
1.4805
1.5847
1.6557
(2,7)
.5174 .7470 .8895 .99421.0774
1.1465
1.2059
Table 2-2. The
Capacity
ofM-ary
(d,k)
constraints.In
the
nextsection weinvoke Proposition 2
and constructthree
new100-percent
efficientcodes.
2.4 Code Constructions
In
this
section we constructthree
100-percent
efficient,
low complexity
codessatisfying
the
(3,1,
o),(13,1,
oo)
and(5,2,
oo)
constraintsusing
the
statesplitting
algorithm whichwill
be discussed in
the
next chapter.For details
onthe
statesplitting
algorithm see[10,15].
A
.(3,1,
o)
code:The FSTD for
the
(3,1,
oo)
is
givenin Figure 2-3
Figure
2-3. The
(3,1,
oo)
constraintand
the
corresponding adjacency
matrixis
T
=0
1
2
1
It is easy
to
showthat
the
largest
positive eigenvalue ofT
is
X
=2
withcorresponding
capacity C
=log2
2
=1.
Choosing
p
=q
-1
, one can
design
a rateR
=p I q
=
l
=C
code via
the
statesplitting
algorithm.It
is
alsoeasily
showthat
an approximateeigenvector v
satisfying Tv
>2v
is
v= (l,2)r.
By
splitting
state1,
athree-state
encoderresults and
is
givenin Figure 2-4. A sliding block decoder
withsliding
windowfour
symbols wide
(corresponding
to
memory
p
=2
and anticipations =1)
is
sufficient.Figure 2-4. The
(3,1,
)
encoder.12
Figure 2-5. The
(13,1,
oo)
constraintand
corresponding adjacency
matrixis
T
=0
1
12
1
with
capacity C
=2
.
Choosing
p
=2,
q
=\
one candesign
a rateR
=
p
I
q
=2
=C
code.
An
approximate eigenvector vsatisfying Tv
>4v is
v=(l,4)r
.
By
splitting
state1
into 4
states,
afive-state
encoder results andis
givenin Figure 2-6. It is easy
to
showthat
a
sliding block decoder
withsliding
windowfour
symbols wide(corresponding
to
memory
p
=2
and anticipation a -1
)
is
sufficient.1,4,7,10
3,6,9,12
Figure 2-6. The
(13,l,oo)
encoder.C
.(5,2,
oo)
code:The FSTD for
the
(5,2,
>)
is
givenin Figure 2-7
Figure 2-7. The
(5,2,
oo)
constraintand
the
corresponding adjacency
matrixis
T
=0
1
0
0
1
A
0
1_
with
capacity C
=1
.Choosing
p
=q
=1
one candesign
a rateR
=piq
=\
=C
code.An
approximate eigenvector vsatisfying Tv
>2v is
v=(1,2,4)
T.After
two
rounds ofsplitting
and some simplemerging
the
five
state encoderin Figure 2-8
results.A sliding
block decoder
with asliding
window six symbols wide(corresponding
to
memory
p-3and anticipation a=
2
0/1,
1/3
0/2,
1/4
Figure 2-8. The
(5,2,
oo)
encoder.2.5 Conclusion
In
this
chapter wehave
shownthat the
results onthe
irrationality
ofcapacity
ofbinary
RLL
constraints of[7]
canbe
extendedto
M-ary (d,k)
constraintsonly
whenk
<.When k
=oothere
existmany
(M,d,k)
constraints.Three
such codes were constructedand shown
to
have
simple encoders andsliding block decoders.
Chapter 3
Coding
Method
3.1 Storage
Density
D
The coding efficiency
can alsobe
measuredby
its
storagedensity
D
(in bits/unit
area).For M
-ary runlength constrainedrecording
channel, the
peak/edgedetection may be
usedin
orderto
increase
the
rate at which symbols are stored onthe
disk
and use a codethat
ensures minimum and maximum mark
width/spacing
constraints are satisfied.The
marksare constrained
in
minimum and maximum widthdenoted
Tmin
andTmax
.The
minimummark
length
constraint arisesfrom
the
read and writeprocesses,
where we can neitherread nor write marks smaller
than
TmiB
.In
systemsthat
use peakdetection,
the
maximummark constraint
Tmax
is
requiredto
maintain systemtiming
sincetiming
is derived from
transitions
in
the
storeddata.
Specifically,
channelbits
are writtenevery T
<Tmin
that
satisfy
the
minimum and maximum mark widths eachconveying
R
information
bits/channel
bit.
Signaling
faster
onthe
disk improves
the
symboldensity
onthe
disk,
but
imposing
constraints onthe
symbol sequences reducesthe
possibleencoding
rateR
.It
is
commonto
define
the
symbolinterval T
suchthat
7\n
=(d
+
Y)T for
somefinite
integer
d
andrmax
=(k
+
\)T
for
somefinite integer
k
.
To satisfy
constraints, the
encoder must produce a run of atleast
d
+
\
and at mostk
+
\
O's
between
consecutive nonzero symbols.The
bit
density
D
in bit/area is
computed asD
=bl
vTW, where
b is
the
number ofbits
storedduring
time
T
on adisk rotating
atspeed v and
track
widthW
.Assuming
a normalized product vW-1
,
the
lineal
density
(often
calleddensity
ratio)
D-bIT (in
bits/sec)
willbe
our measure ofperformance.Since
the
code produces(d
+
1)
symbols/Tmin
, eachconveying
R
bits/symbol,
it follows
that
D=(fW
(into/7^)
(3-1)
min
Since
the
constrainedcapacity C is
afundamental
upperbound
to
R
given constraintsM
,d
, andk
.The
largest bit
density
Dc
achievableby
any
codesatisfying
the
M
?try{d,k)
constraintis
T.
T
nun nun
We
madeTable 4
to
demonstrate
the
resulting
achievabledensities
Dc
for
selected valuesof
M
,d
andk
.The
table
entriesfor
M
=2
correspondto
conventionalbinary
recording
channels.fd.k,
M=2
3
4
5
6
7
8
9
10
11
12
(1,3)
1.10
1.85
2.30
2.64
2.90
3.12
3.31
3.47
3.62
3.87
4.26
(1,7)
1.35
1.99
2.40
2.71
2.96
3.16
3.34
3.50
3.64
3.89
4.27
(2,7)
1.55
2.24
2.66
2.98
3.23
3.43
3.61
3.77
3.91
4.15
4.52
(2,15)
1.65
2.28
2.69
2.99
3.24
3.44
3.62
3.77
3.91
4.15
4.53
(3,8)
1.70
2.42
2.86
3.18
3.43
3.64
3.82
3.98
4.12
4.35
4.73
(3,10)
1.78
2.47
2.89
3.21
3.45
3.66
3.84
3.99
4.13
4.37
4.74
(4,9)
1.81
2.56
3.01
3.33
3.59
3.80
3.98
4.14
4.28
4.52
4.90
(4,15)
1.99
2.67
3.09
3.40
3.65
3.88
4.03
4.18
4.31
4.55
4.92
(5,10)
1.89
2.66
3.13
3.46
3.72
3.93
4.12
4.28
4.42
4.66
5.04
(5,15)
2.10
2.80
3.24
3.55
3.82
4.01
4.18
4.33
4.47
4.70
5.07
(6,13)
2.10
2.86
3.31
3.64
3.89
4.10
4.28
4.44
4.58
4.82
5.19
(7,14)
2.17
2.94
3.40
3.74
4.00
4.21
4.39
4.55
4.69
4.93
5.30
(7,22)
2.37
3.07
3.50
3.82
4.06
4.27
4.44
4.60
4.73
4.97
5.33
Table 3-1. Achievable
densities
(in
bits I
Tmin
)
Dc
for
M
-ray(d,
k)
codesIt is
evidentthat
aM-ary (d,k)
constrained code canincrease
the
storagedensity
significantly.Such
like (d,k
)
=(1,3)
constrainedcode,
whenM
=2
,
the
density
valueis
1.10,
whenM
=4
,
the
density
value changeto
2.30. But
the
increase in bit
density
due
to
M
-arysignaling may be
offsetby
anincreased
error probability.Namely,
while one often assumesthe
M
=2
saturation channelis
noiseless ornearly
noiseless, the
generalM
-arychannel
is
noisy,
especially
whenM
is large.
3.2 Three
Coding
Methods
There
are avariety
techniques
arevery
usefulto
producerunlength-limited sequencesin
the
practicalway.Here
weintroduce
some methods:1.
Fixed-length
(d,k
)
codesintroduce
block
codes,
first let
us address a method ofcounting
the
numberof sequencesof acertain
length
whichcomply
with given( d,
k
)
constraints[17].
Let
Nd
(n)
denote
the
numberofdistinct
(
d
)
sequences oflength
n anddefine
Nd(n)
=0,
NAO)
=U
n<0,
The
number of(d
)
sequences oflength
n>0 is
fund
withthe
recursive relations[12]
Nd(n)
=n+
l,
\<n<d
+l,
Nd(n)
=Nd{n-\)
+Nd(n-d-\),
n>d+\.
By
using
abovefunctions,
we getatable
ofNd
(n)
whichis
the
number of(
d
)
sequencelength
n:n=
1
2
3
4
5
6
7
8
9
10
11
12
13
14
d=
1
2
3
5
8
13
21
34
55
89
144
233
377
610
987
2
2
3
4
6
9
13
19
28
41
60
88
129
189
277
3
2
3
4
5
7
10
14
19
26
36
50
69
95
131
4
2
3
4
5
6
8
11
15
20
26
34
45
60
80
5
2
3
4
5
6
7
9
12
16
21
27
34
43
55
Table
3-2. Number
ofdistinct
( d )
sequences as afunction
ofthesequencelength
n andtheminimumrunlength
d
as a parameter.The
number of(d,k)
sequences oflength
n canbe found in
a similar way.Let
N(n)
denote
the
numberof sequences oflength
n.Define
N(n)
=0,
/Y(0)
=1
.
n<0.