4
1973 lEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 39, NO. 6, NOVEMBER 1993 Increased Information Rate by Oversampling E. N. Gilbert Abstract-A noiseless ideal low-pass filter, followed by a limiter, is normally used as a binary data channel by sampling the output once per Nyquist interval. Detectors that sample more often encounter intersym- bo1 interference, but can be used in ways that increase the information rate. Index Terms-Intersymbol interference, band-limited signals, Nyquist rate, oversampling, limiter. 1. INTRODUCTION Fig. 1 shows a noiseless channel, consisting of an ideal low-pass filter followed by a limiter. If the limiter input function is f(t), the output is y(t) = sgn(f(t)). Channels of this sort are com- monly used with binary data serving directly as the Nyquist samples of a low-pass input function x(t). The detector then samples thc limiter output at the Nyquist times to recover the data and achieve a signaling rate of one bit per Nyquist interval. A detector that merely samples the channel output at Nyquist times receives information no faster than 1 b/Nyquist interval, regardless of the source. One may reasonable ask if, with a suitable source, the detector can achieve a higher information rate by sampling more often (ouersampling). The answer is not obvious, as examples in Section I1 show. Section I11 will exhibit a signaling system that achieves an information rate 1.072 b/Nyquist interval by allowing the detector to sample at times a half Nyquist interval apart. This rate increase is small, but it shows that, in principle, oversampling does permit faster rates. Section IV gives another signaling system that more closely resembles a digital data system; it has rate 1.050. By allowing a more irregular pattern of sampling times, Section V increases the rate to 1.090 (or 1.062 for the “digital” system). It may be that only a complicated oversampling system can make a large improvement in rate, but that will not be proved. Mazo suggested the present problem as a simplification of a more realistic one in Kalet et al. [l]. The simplification omits details like multiple slicing levels (256 versus 2) and a less convenient sample rate (8/7 versus 2) which seem unrelated to the main problem of showing that extra samples can supply new information. Several authors [2]-[4] have derived high bit-error ratcs for moderately oversampled digital systems. 11. PRELIMINARIES Time will be measured in units of the Nyquist interval. Then, the cutoff frequency of the filter in Fig. 1 will be W= 1/2 and the response of the filter to a unit impulse at time 0 will be The sampling theorem (Shannon [5], [6]) represents an output function f(t) of the filter in terms of its Nyquist samples f(i) at integer times (2) If the input x(t) is itself a low-pass signal, f(t) = x(t) and the f(t) = Cf(i)F(t - i). Manuscript received September 8, 1992; revised February 22, 1993. The author is with AT & T Bell Laboratories, Murray Hill, NJ 07974. IEEE Log Number 9213026. Fig. 1. Channel. output is y(t) = sgn (x(t)). Then, taking binary data x(n) = + 1 or - 1 for the input Nyquist samples will make y(n) = x(n). Now, suppose the detector samples twice as often, say at times i/2 with integer i. Times with even i, and so integer i/2, are again the Nyquist times. If i is odd, i/2 has fractional part 1/2 and will be called a half-Nyquist time. If the detector could sample f(t) directly, sampling at the half-Nyquist times would be wasted effort; the Nyquist samples in (2) already determine f(t). But, since the detector can only observe y(t) = sgn(f(t)), the half-Nyquist samples of y(t) may indeed contain extra informa- tion. How can a detector use these samples? A naive approach might transmit two message sequences, say a(j) at the Nyquist times and b(i) at the half-Nyquist times, making the filter output Although a(j) and b(i) are the coefficients of the biggest con- tributors to (3) at times t = j and t = i - 0.5, the detection is now complicated. For example, f(n) = a(n) + xb(i)F(n - i + 0.5). (4) The series in (4) represents a kind of noise (intersymbol interfer- ence) obscuring the sample a(n). It is not obvious that intersym- bo1 interferences will not always reduce the information rate of y(t) to at most 1 b/Nyquist interval. As a special case, let the a(j> and b(i) be independent Gaussian variates, all with zero mean and the same variance. The two series in (3) become independent Gaussian noise functions with the same power. Shannon’s formula for the capacity of a Gaussian channel then shows that f(f) gives information about the message sequence {a(j)) at rate 0.5 b/Nyquist interval. The rate of information about both messages in only 1 b/Nyquist interval, even if the detector observes f(t) instead of y(t). A detector that samples y(t) many times per Nyquist interval can accurately locate the zeros of f(t), and perhaps thereby receive information at a rate faster than 1. Zero crossings do not determine a low-pass function, even to within a scale factor. Consider the simple example: f(t) = sin (27rft/3) + c sin (27rft) = (1 + c + 2c cos 4rft/3} sin (2.rrft/3) where the parameter c can have any value in - 1/3 < c < 1. With c in that range, the bracketed term remains positive and the zeros of f(t) are just those of sin (2.rrft/3). However, at least one low-pass function with prescribed simple zeros does exist if each Nyquist interval contains exactly one of the zeros (see the infinite product for such a function in Logan [7] and Titchmarsh [8]). Using that result, Shamai [9] proposed the following over- sampling system, with N samples per Nyquist interval, to achieve an information rate log(N + 1). Each Nyquist interval (n, n + 1) will contain N sample points, say at n + (2k - 1)/(2N), k = l;.., N, for an evenly spaced 0018-9448/93$03.00 0 1993 IEEE

Increased information rate by oversampling

  • Upload
    en

  • View
    213

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Increased information rate by oversampling

1973 lEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 39, NO. 6, NOVEMBER 1993

Increased Information Rate by Oversampling

E. N. Gilbert

Abstract-A noiseless ideal low-pass filter, followed by a limiter, is normally used as a binary data channel by sampling the output once per Nyquist interval. Detectors that sample more often encounter intersym- bo1 interference, but can be used in ways that increase the information rate.

Index Terms-Intersymbol interference, band-limited signals, Nyquist rate, oversampling, limiter.

1. INTRODUCTION Fig. 1 shows a noiseless channel, consisting of an ideal low-pass

filter followed by a limiter. If the limiter input function is f ( t ) , the output is y ( t ) = sgn(f(t)). Channels of this sort are com- monly used with binary data serving directly as the Nyquist samples of a low-pass input function x( t ) . The detector then samples thc limiter output at the Nyquist times to recover the data and achieve a signaling rate of one bit per Nyquist interval.

A detector that merely samples the channel output at Nyquist times receives information no faster than 1 b/Nyquist interval, regardless of the source. One may reasonable ask if, with a suitable source, the detector can achieve a higher information rate by sampling more often (ouersampling). The answer is not obvious, as examples in Section I1 show. Section I11 will exhibit a signaling system that achieves an information rate 1.072 b/Nyquist interval by allowing the detector to sample at times a half Nyquist interval apart. This rate increase is small, but it shows that, in principle, oversampling does permit faster rates. Section IV gives another signaling system that more closely resembles a digital data system; it has rate 1.050. By allowing a more irregular pattern of sampling times, Section V increases the rate to 1.090 (or 1.062 for the “digital” system). It may be that only a complicated oversampling system can make a large improvement in rate, but that will not be proved.

Mazo suggested the present problem as a simplification of a more realistic one in Kalet et al. [l]. The simplification omits details like multiple slicing levels (256 versus 2) and a less convenient sample rate (8/7 versus 2) which seem unrelated to the main problem of showing that extra samples can supply new information. Several authors [2]-[4] have derived high bit-error ratcs for moderately oversampled digital systems.

11. PRELIMINARIES

Time will be measured in units of the Nyquist interval. Then, the cutoff frequency of the filter in Fig. 1 will be W = 1/2 and the response of the filter to a unit impulse at time 0 will be

The sampling theorem (Shannon [5], [6]) represents an output function f ( t ) of the filter in terms of its Nyquist samples f ( i ) at integer times

(2)

If the input x ( t ) is itself a low-pass signal, f ( t ) = x ( t ) and the

f ( t ) = C f ( i ) F ( t - i ) .

Manuscript received September 8, 1992; revised February 22, 1993. The author is with AT & T Bell Laboratories, Murray Hill, NJ 07974. IEEE Log Number 9213026.

Fig. 1. Channel.

output is y ( t ) = sgn ( x ( t ) ) . Then, taking binary data x ( n ) = + 1 or - 1 for the input Nyquist samples will make y ( n ) = x ( n ) .

Now, suppose the detector samples twice as often, say at times i / 2 with integer i. Times with even i, and so integer i / 2 , are again the Nyquist times. If i is odd, i/2 has fractional part 1/2 and will be called a half-Nyquist time. If the detector could sample f ( t ) directly, sampling at the half-Nyquist times would be wasted effort; the Nyquist samples in (2) already determine f ( t ) . But, since the detector can only observe y ( t ) = sgn(f(t)), the half-Nyquist samples of y ( t ) may indeed contain extra informa- tion. How can a detector use these samples?

A naive approach might transmit two message sequences, say a ( j ) at the Nyquist times and b ( i ) at the half-Nyquist times, making the filter output

Although a ( j ) and b(i) are the coefficients of the biggest con- tributors to (3) at times t = j and t = i - 0.5, the detection is now complicated. For example,

f ( n ) = a ( n ) + x b ( i ) F ( n - i + 0.5). (4)

The series in (4) represents a kind of noise (intersymbol interfer- ence) obscuring the sample a(n). It is not obvious that intersym- bo1 interferences will not always reduce the information rate of y ( t ) to at most 1 b/Nyquist interval. As a special case, let the a( j> and b(i) be independent Gaussian variates, all with zero mean and the same variance. The two series in (3) become independent Gaussian noise functions with the same power. Shannon’s formula for the capacity of a Gaussian channel then shows that f ( f ) gives information about the message sequence {a( j ) ) at rate 0.5 b/Nyquist interval. The rate of information about both messages in only 1 b/Nyquist interval, even if the detector observes f ( t ) instead of y( t ) .

A detector that samples y ( t ) many times per Nyquist interval can accurately locate the zeros of f ( t ) , and perhaps thereby receive information at a rate faster than 1. Zero crossings do not determine a low-pass function, even to within a scale factor. Consider the simple example:

f ( t ) = sin (27r f t /3) + c sin (27rf t )

= (1 + c + 2c cos 4 r f t / 3 } sin (2.rrft/3)

where the parameter c can have any value in - 1/3 < c < 1. With c in that range, the bracketed term remains positive and the zeros of f ( t ) are just those of sin (2.rrft/3). However, at least one low-pass function with prescribed simple zeros does exist if each Nyquist interval contains exactly one of the zeros (see the infinite product for such a function in Logan [7] and Titchmarsh [8]). Using that result, Shamai [9] proposed the following over- sampling system, with N samples per Nyquist interval, to achieve an information rate log(N + 1).

Each Nyquist interval ( n , n + 1) will contain N sample points, say at n + (2k - 1)/(2N), k = l;.., N , for an evenly spaced

0018-9448/93$03.00 0 1993 IEEE

Page 2: Increased information rate by oversampling

1974 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 39, NO. 6, NOVEMBER 1993

sampling sequence. Sample points cut each Nyquist interval into N + 1 segments, two of length 1/(2N) and the rest of length 1,". From each Nyquist interval, the source chooses one seg- ment at random, all N + 1 equally likely. The midpoint of that segment will become a simple zero. The transmitter sends a low-pass signal f ( t ) with the chosen zeros. Because a segment has at most one simple zero, a segment with a zero is one having endpoints with y ( t ) of opposite sign. Then, zeros in segments between sample points are easily detected. An endpoint of a segment may also be an integer point n, not a sample point, but y ( n ) can be deduced from nearby samples because each Nyquist interval contains exactly one zero ( y ( n ) = -y(n - 1)). Generat- ing the required signal f ( t ) presents practical problems; f ( t ) depends on locations of zeros far in the future, and so one must expect long coding delays. The systems to follow are much slower, but use transmitted signals constructed only from given Nyquist samples.

111. GAUSSIAN SOURCE

In this section, x ( t ) will be a low-pass Gaussian noise signal; the input Nyquist samples x(n) will be independent Gaussian variates with zero mean and unit variance. The receiver will try to deduce x(n) from just two output samples, the Nyquist sample y ( n ) and half-Nyquist sample y (n - 0.5). Write X = x(n) and Y = ( y ( n - OS), y(n)). The detector reveals information about the input at a rate

in bits per Nyquist interval. Here, H ( X ) , H ( Y ) , H ( X I Y ) , H( Y (X) are information and conditional information measures as defined by Shannon [ 5 ] . Of course, the detector cannot reconstruct the input x(n) exactly from a few received binary samples. However, Shannon's coding theorem will guarantee that other sources with rate less than R can be coded to signal over this channel at low error rate.

Gaussian input samples were chosen purely for analytical convenience. Restricting input samples x ( n ) to a few discrete levels would have simplified the practical problem of coding discrete sources for the channel. There is also no reason to expect Gaussian input to maximize the rate R. Using more than two output samples might give a larger rate R , but the objective is to obtain R > 1 as simply as possible.

It may be intuitively clear that the entire output sequence gives information about the entire input sequence at a rate at least R , the information that two output samples give about one input sample. A formal proof follows. Write Xk and Yk for the transmitted and received samples in the first k Nyquist intervals, i.e.,

R = H ( X ) - H ( X I Y ) = H ( Y ) - H ( Y J X ) (5)

Xk = ( x ( l ) , ( 2 ) ; . . , x ( k ) ) and Yk = ( ~ ( 0 3 , y ( I ) ; . . , Y ( k - 0.51, y ( k ) ) ,

and write kR, = H(Xk) - H(X,(Y,) for the information that Yk gives about X, . Since the input samples are independent,

k

H ( X k ) = H ( x ( i ) ) . r = l

Also, H(Xk(yk) = H ( x ( l ) / Y k ) f ff(x(2)1x(l), y k )

k

+ '.. = H ( x ( i ) l X L - , , Y , )

H(x( i ) l y ( i - 0.51, y ( i > ) ,

I = 1 k

- < r = 1

the inequality following because X r - l and Yk determine y ( i -

0.5) and y ( i ) exactly. But then,

k

kRk 2 { H ( x ( i ) ) - H ( ~ ( i ) ( y ( i - O S ) , ~ ( i ) } = kR. i = 1

To calculate R , begin with the joint probability distribution of X and Y. Define two Gaussian variates f = f ( n ) and g = f ( n - O S ) , so that X = f and Y = (sgn (g ) , sgn (f)) with E ( f ) = E ( g ) = 0 and E(f2) = E ( g 2 ) = 1. The correlation coefficient of f and g is

p = E ( f g ) = F(0.5) = 2 / r r . (6)

Using (6), one can easily verify that the new Gaussian variates

u = (f + g>/ Jm, U = ( f - g ) / J m ( 7 )

have zero mean, unit variance, and are independent. Then, probabilities of events determined by ( f , g), or by ( U , U ) , are integrals of the elements

e - r 2 12du du e - r 2 12df dg (8) - -

2 T 2 4 7

in which r 2 is the quadratic form

r 2 = u 2 + u 2 = ( f 2 - 2pfg + g2)/(I - p 2 )

= f 2 + ( g - pfI2/(1 - p2) , (9)

i.e., r is the radial coordinate in the (U, U ) plane. To evaluate H(Y) , find probabilities of the four possible sign

pairs + +, + -, - f , - - for Y . Y = + + if (f,g) is in the first quadrant, i.e., if (U, U ) is in an angular sector A between the lines U / U = f J(1 + p ) / ( l - p ) . The probability that Y =

+ + is easily evaluated as an integral in the (U, U ) plane. Express U and U in polar coordinates to obtain

p ( + +) = L m e - r 2 / 2 r d r j d0/2n A /z = 0.35983396. = - arctan -

1

rr

The other outcomes for Y have probabilities

P ( - - > = P ( + +),

P( - +) = P( + -) = 0.5 - P( + +) = 0.14016604.

Then, H ( Y ) = 1.855903 b/Nyquist interval. Because X determines the second coordinate of Y exactly,

H ( Y I X ) is just the uncertainty of sgn(g) when f is known. For a given f, the conditional probability density for g is obtained by dividing the joint density in (8) by the marginal density exp ( - f 2 / 2 ) / 6 for f. Then, (9) shows the conditional distri- bution for g to be Gaussian with mean p f and variance 1 - p2. An integration over 0 < g <

P{sgn (8) = + I l f } = a( p f / \ / l - s z )

gives

(10)

where

is the normal distribution function. Using (lo), one can calculate the uncertainty about sgn(g) when f has a known value. That

Page 3: Increased information rate by oversampling

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 39, NO. 6, NOVEMBER 1993 1975

uncertainly must be multiplied by the Gaussian density function at f and integrated to get H(YIX) . The integration was done numerically, with the result H ( Y I X ) = 0.7838. The rate in (5) is then R = 1.072.

IV. DIGITAL SOURCE

The idea in Section I11 might be used with digital signals by encoding letters into discrete amplitude levels that approximate a Gaussian distribution. A large alphabet should give a good approximation to a Gaussian input and a rate near 1.072. This section modifies that idea to obtain a rate greater than 1 for digital sources having alphabets of only four letters.

Instead of assigning a specific level to each of the four letters of the digital signal alphabet, the encoding will choose levels at random. The random choices will be designed to make the channel input samples Gaussian so that the methods of Section 111 again apply. First, classify the values f of inputs to the channel of Fig. 1 into four ranges:

-B = (-CO, - L ) , -S = ( -L ,O) ,

S = (0, L ) , B = ( L , x ) (11) with L any convenient positive number ( B and S stand for “big” and “small”). Also, use - B , - S , S, and B as names for the four letters of the digital alphabet. Whenever letter Z is to be transmitted (2 = -B, -S, S, or B), the input Nyquist sample to the channel will be chose at random from the range Z in (11). To make the input Gaussian, the letter probabilities must be

P ( B ) = P ( - B ) = 1 - @ ( L ) , P(S) = P ( - S ) = @ ( L ) - 0.5. (12)

Thus, a digital source with four equally likely letters will require L = 0.67449. Also, for each letter Z = -B, -S, S, or B, the random Nyquist sample f must be chosen from range Z with a conditional probability density

otherwise.

The signaling rate is now

R = H ( Z ) - H ( Z I Y ) = H ( Y ) - H ( Y I Z ) . (14) Input letters Z will be assumed to arrive independently with the probabilities (12). The outputs Y are binary pairs (sgn (g ) , sgn (f)) of one half-Nyquist sample and one Nyquist sample, as in Section 111. The rate R is then what would be achieved by a memoryless source signaling over a noisy memory- less discrete channel with four inputs and four outputs. The transition probabilities of this channel can be found from (8) and (13) in order to find R.

The second equation of (14) will be convenient; Section 111 has already found H ( Y ) = 1.855903. Since the output sample y ( n ) always has the same sign as Z, there are only eight nonzero transition probabilities P(Y 12). These are simply related to one joint probability Q = P(S, + + ) = P( -S, - - ) and letter probabilities (12) by

P ( + + IS) = P( - - I - S) = Q / P ( S )

P ( + + IB) =P(- - I - B ) = ( P ( + + ) - Q ) / P ( B )

P ( - + IS) = P ( + - I - S) = 1 - P ( + + IS)

P ( - + IB) = P ( + - I - B ) = 1 - P ( + + IB). (15) The value P(+ + ) = 0.35983396, from Section 111, appears in

(15). Q is found by integrating (8) over the region 0 < g < x, 0 < f < L. Use the last equation in (9) to obtain

and compute Q numerically. The letter probabilities (12) and transition probabilities (15) now determine the conditional infor- mation H(YIZ) and the information rate (14). These numbers all depend on the value L. At L = 0.67449, which makes the four letters equally likely, Q = 0.1511, P ( + + IS) = 0.6044. P ( + + IB) = 0.8349, H(YIZ) = 0.8073, and R = 1.049.

The choice L = 0.67449 maximizes H ( Z ) , but not R(H(ZIY) also depends on L). As a function of L , R has a broad maximum, attaining R = 1.0502 at L = 0.83.

This signaling system might be used to multiplex two binary sources. Let a sign source have alphabet (+, - } and rate 1. Let an amplitude source have alphabet {S, B} and rate less than R - 1. The letter pairs -B, -S, +S, +B of the combined source are then suitable channel inputs. The receiver simply decodes - -, + -, - +, + + into -B, -S, +S, +B, respec- tively. That reconstructs the sign message exactly. Errors in the decoded letters S, B may be corrected if the amplitude message was suitably encoded.

V. OTHER SPACING

An alternative to sampling at evenly spaced times is to dis- place the half-Nyquist times slightly. Let y ( n - T ) and y (n ) , for some T different from 0.5, be used for g and f in Sections 111 and IV. The equations of Sections 111 and IV remain unchanged, with p in (6) replaced by a new value p = F(T) . In fact, other values of T can improve the rate R.

This section will take T = 0.44294647, which makes p =

1/ a. Then, in (10) and (16), p / = 1. That simplifica- tion allows all the integrals of Sections 111 and IV to be done analytically. P( + + ) = P( - - ) = 3/8, P( + - ) = P( - + ) =

1/8, so that H ( Y ) = 1.811278. In (lo), P(sgn(g) = + l l f ) =

@(f), and now

H ( Y I X ) = 2/I{-@log@ - (1 - @ ) l o g ( l - @ ) } d @ , (I

= 1/(2In2) = 0.72135. The rate in Section 111 increases to R = 1.08993.

In Section IV, (16) simplifies to

which, together with P ( + + ) = 3/8, reduces the transition probabilities (15) to expressions involving @ ( L ) only. With L =

0.67449 for equally likely letters, the rate is R = 1.06228, Chang- ing L to 0.8 increases the rate to R = 1.06344.

REFERENCES I. Kalet, J . E. Mazo, and B. R. Satzberg, “The capacity of PCM voiceband channels,” submitted to ZEEE Trans. Commun. J. E. Mazo, “Faster-than-Nyquist signaling,” Bell Syst. Tech. J., vol. 54, pp. 1451-1462, 1975. B. R. Saltzberg, “Intersymbol interference error bounds with ap- plication to ideal bandlimited signaling,” ZEEE Trans. Znform. Theoq, vol. IT-14, pp. 563-568, 1968. G. D. Forney, “Lower bounds on error probability in the presence of large intersymbol interference,” ZEEE Trans. Commun., vol.

C. E. Shannon, “A mathematical theory of communication,” Bell Syst. Tech. J . , vol. 27, pp. 379-423, 623-656, 1948.

COM-20, pp. 76-77, 1972.

Page 4: Increased information rate by oversampling

I976 IEEE TRANSACTIONS O N INFORMATION THEORY, VOL. 39, NO. 6, NOVEMBER 1993

[6] -, “Communication in the presence of noise,” Proc. I R E , vol. 37, pp. 10-21, 1949; reprinted in Proc. IEEE, vol. 72, pp. 1192-1201, 1984. B. F. Logan, Jr., “Signals designed for recovery after clipping-I, 11, 111,” ATT Bell Lab. Tech. J . , vol. 63, pp. 261-286, 287-306,

[8] E. C. Titchmarsh, “The zeros of certain integral functions,” Proc. London Math. Soc., vol. 25, pp. 283-302, 1926.

[9] S. Shamai, private communication.

[7]

379-399, 1984.

Cascading Runlength-Limited Sequences

Jos H. Weber, Member, IEEE, and Khaled A. S. Abdel-Ghaffar

Abstract-In magnetic or optical storage devices, it is often required to map the data into runlength-limited sequences. To ensure that cascad- ing such sequences does not violate the runlength constraints, a number of merging hits are inserted between two successive sequences. A theory is developed, in which the minimum number of merging bits is deter- mined and the efficiency of a runlength-limited fixed-length coding scheme is considered.

between two consecutive ones. Such sequences are called dklr- sequences. The cardinality of is denoted by Mdkll(n). In case there is no upper runlength constraint, we say k = x. Note that the assumption 1 I r is not a real restriction, since

d d k l r ( n ) with 1 > r could be obtained by simply reversing each sequence in J d k r l ( n ) . Let y i denote the sequence y repeated i times. For example, 0310(100)21 = 000101001001. Let Rev[y] denote the reversed sequence of y . The number of “1’s” in a sequence will be called the weight of the sequence. Let 1x1 and 1x1 denote the floor and the ceiling of the real number x, respectively.

A way to guarantee that cascading dklr-sequences gives se- quences still satisfying the (d , k) constraint is to insert b appro- priate merging bits between any two consecutive sequences. If

k l r (n ) as a runlength-limited code where each codeword corresponds to a message from a message set d of size A , then the actual number of bits needed to represent a message is thus N = n + b. The code rate R can then be defined by

As an example, we mention the Eight-to-Fourteen Modulation (EFM) code for the compact disk [4], where eight data bits are mapped to a runlength-limited sequence of length 14 with ( d , k )

Index Terms-Fixed-length codes, modulation systems, runlength- limited sequences.

I. INTRODUCTION = ( 2 , l O ) . In order to maintain the runlength conditions when

Many modulation systems used in magnetic and optical recording are based on binary runlength-limited codes. A string of bits is said to be runlength-limited if the number of zeros between two consecutive ones is bounded between a certain minimum value d and a certain maximum value k. Such se- quences are also called ( d , k ) constrained. The lower runlength constraint is imposed to reduce intersymbol interference, while the upper runlength constraint is imposed to maintain synchro- nization. For a general introduction to the theory of runlength- limited sequences and its applications, we refer to the excellent overview papers by Siege1 [7] and Immink 131.

In order to prevent the cascading of runlength-limited se- quences of fixed length from violating the runlength constraints, some merging techniques have been given by Tang and Bahl [SI and Beenker and Immink 111. In this paper, we will develop a general theory on cascading runlength-limited sequences, in which the above-mentioned methods appear as special cases.

The following notation will be used. Let d , k , I , r , and n be integers satisfying 0 I d < k, 0 I I I r I k , and n 2 1, unless explicitly stated otherwise. Let J d J n ) denote the set of binary sequences of length n , starting with at most 1 zeros, ending with at most r zeros, and having at least d and at most k zeros

cascading these sequences, merging bits are inserted between adjacent sequences. The runlength-limited sequences are judi- ciously chosen such that three merging bits suffice to maintain the runlength constraints. If the runlength is in danger of becoming too short, all of them are set to “0.” If the runlength is in danger of becoming too long, one of the merging bits is set to “1.” The code rate of this system is (log, 28)/(14 + 3 ) = 8/17.

Our first goal will be to minimize the number of merging bits needed to cascade any two sequences from Mdklr(n) . Our final goal will then be to maximize the code rate R for given values of d, k, and A by a proper choice of 1 and r. To this end, note the following with regard to this choice. On one hand, it seems a good idea to take large values for 1 and r , since Mcikl,.(n) is nondecreasing in 1 and r. On the other hand, small values of 1 and r may require fewer merging bits than large values. This dilemma plays a central part in optimizing the efficiency of the coding scheme.

The rest of this paper will be organized as follows. First, in Section 11, we derive recursion relations for determining Mdklr (n ) . Further, in Section 111, we give lower bounds on the number of merging bits needed for cascading any two sequences from Adkcr (n ) . Next, in Section IV, we give optimal merging rules for such a cascading scheme. Then, in Section V, we try to

I

find out, for any choice of d and k , which values of 1 and r give the best overall efficiency for a fixed-length coding scheme’ Finally, conc~usions are drawn in Section VI.

Manuscript received May 23, 1992; revised March 3 , 1993. This work was supported by the National Science Foundation under Grants NCR 89-08105 and NCR 91-15423. The work of K. A. S. Abdel-Ghaffar was supported by an IBM Faculty Development Award. This paper was presented in part at the 12th Symposium on Information Theory in the Benelux, Veldhoven, The Netherlands, May 1991, and at the 3991 lEEE International Symposium on Information Theory, Budapest, Hungary, June 1991.

J. H. Weber is with the Information Theory Group, Department of Electrical Engineering, Delft University of Technology, 2600 GA Delft, The Netherlands.

ing and Computer Science, University of California, Davis, CA 95616.

11. NUMBER OF dkb-SEQUENCES

The number Of dklr-sequences Of length has been derived by Tang and Bahl [8] for the special case I = r = k . In this section, we will give recursion relations to determine Mdklr(n) for arbitrary choices of 1 and y . Let /,(n) denote the set of

the sizes M T k / r ( n ) Of K, A, S, Abdel-Ghaffar is with the Department of Electrical Engineer- dkfr-sequences Of length and weight ”‘ We first

IEEE Log Number 9213023. Since the all-zero sequence of length n is a dklr-sequence if

0018-9448/93$03.00 0 1993 IEEE