+ All Categories
Home > Documents > A Min Dist of Turbo Like Codes

A Min Dist of Turbo Like Codes

Date post: 14-Apr-2018
Category:
Upload: jayant5253
View: 218 times
Download: 0 times
Share this document with a friend
10
6 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, J ANUARY 2009 The Minimum Distance of Turbo-Like Codes Louay Bazzi, Mohammad Mahdi an, and Danie l A. Spielman  Abstract—Worst -case upper bounds are derive d on the min- imum dis tance of par allel con cat enated tur bo cod es, ser ial ly concatenated convolutional codes, repeat-accumulate codes, re- peat-convolute codes, and generalizations of these codes obtained by allowing nonlinear and large-memory constituent codes. It is shown that parallel-concatenated turbo codes and repeat-convo- lute codes with sub-linear memory are asymptotically bad. It is also shown that depth-two serially concatenated codes with con- stant-memory outer codes and sublinear-memory inner codes are asymptotically bad. Most of these upper bounds hold even when the convolutional enco ders are rep laced by gener al nite -sta te automata enco ders . In contr ast, it is prov en that depth-th ree serially concatenated codes obtained by concatenating a repetition code with two accumulator codes through random permutations can be asymptotically good.  Index Terms—Asymptoti c gro wth, conc atena ted codes , min- imum dista nce, rep eat-accumulate-accumula te (RAA) codes, turbo codes. I. INTRODUCTION T HE low-compl exit y and near- capac ity perfor mance of turbo codes [3], [9] has led to a revolution in coding theory. The most famous casualty of the revolution has been the idea that good codes should have high minimum distance: the most useful turbo codes have been observed to have low minimum distance. In this work, we provide general conditions under which many constructions of turbo-like codes, including families of serially concatenated convolutional codes [2] and Repeat-Ac- cumulate (RA) codes [5]–[7], must be asymptotically bad 1 . We also pres ent a simple family of depth- seria lly conca tenated convolutional codes that are asymptotically good. Our work is motivated by the analyses of randomly con- structed parallel and serially concatenated convolutional codes by Kahale and Urbanke [7] and of parallel concatenated turbo codes with two branches by Breiling [4]. Manuscrip t recei ved October 15, 2007 ; revi sed Augu st 20, 2008. Curre nt ver- sion published December 24, 2008. This work was supported by ARO under Grant DAA L03-92-G-0115, by MURI under Grant DAAD19-00-1-0466, by NSF CCR: 9701304, and by NSF CCR: 9701304. L. Bazzi is with the Department of Electrical and Computer Engineering, American University of Beirut, Beirut, Lebanon (e-mail: [email protected]). M. Mahdian is with Yahoo! Research, Santa Clara, CA 95054 USA (e-mail: [email protected]). D. A. Spielman is with the Department of Computer Science and Program in Appl ied Math emat ics, Y ale Univ ersit y , New Hav en, CT 0652 0 USA ([email protected]). Communicated by T. J. Richardson, Associate Editor for Coding Theory. Digital Object Identier 10.1109/TIT.2008.200 8114 1 A sequence of codes of increasing block length is called an asymptotically good code if the message length and the minimum distance of the codes grows linea rly with the block length. Codes for whic h eithe r the message length or minimum distance grow sub-linearly with the block length are called asymptot- ically bad . Kahale and Urbanke [7] provided probabilistic estimates on the minimum distance of randomly generated parallel concate- nated turbo codes with a constant number of branches. They also provided similar estimates for the minimum distance of the random concat ena tion of two con vol uti ona l codes wit h bounded memory. In particular, Kahale and Urbanke proved that if one b uilds a para llel co ncate nated code with branc hes from ran dom per mut ations and con vo lut ion al encoders of memory a t mos t , the n the r esu lti ng co de ha s min imum dis - tance at most and at least with h igh pro babil ity , where is the number of mes sage bits. For ra te seri al ly c oncatena ted co nvo luti onal codes wit h random interleavers, they proved that the resulting code has minimum distance at most and at least with high probability , where is the free dis tan ce of the out er code a nd is the inn er cod e memory. Breiling [4] proved that the parallel concatenation of two convolutional codes with bounded memory has at most loga- rithmic minimum distance, regardless of the choice of inter- leaver. In particular, for parallel concatenated turbo codes with two branches, Breiling proved that no construction could be muc h bet ter tha n a ran dom constr uct ion : if the constitue nt cod es have memor y , the n the mini mum dist anc e of the res ult ing code is . These bounds naturally lead to the following ve questions. (Bett er than rand om?) Do there exist asymp totica lly good parallel concatenated turbo codes with more than two branches or do there exist asymptotically good re- peat-convolute or repeat-accumulate codes? Note that the result of Breiling only applies to turbo codes with two branches and the results of Kahale and Urbanke do not preclude the existence of codes that are better than the randomly generated codes. (Larger memory?) What happens if we allow the mem- ories of the constituent convolutional codes to grow with the block length? All the previous bounds become vacuous if the memory even grows logarithmically with the block length. (Nonlinearity?) Can the minimum distance of turbo-like codes be improved by the use of non-linear constituent en- coders, such as automata encoders? (Concatenation depth?) Can one obtain asymptotically goo d cod es by serial ly concat ena tin g a rep eti tio n code wit h two levels of convolutional codes? We will give essentially negative answers to the rst three questions and a positive answer to the last one. For parallel conca tenat ions and depth- serial concaten ation s of con volu - tional codes and non-linear automata codes, we prove upper bounds on the minimum distance of the resulting codes in terms of the memories of the constituent codes. In Section II-A, we show that parall el conca tenat ed codes and repea t-con volu te 0018-9448/$25.00 © 2009 IEEE
Transcript
Page 1: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 1/10

6 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, JANUARY 2009

The Minimum Distance of Turbo-Like CodesLouay Bazzi, Mohammad Mahdian, and Daniel A. Spielman

 Abstract—Worst-case upper bounds are derived on the min-imum distance of parallel concatenated turbo codes, seriallyconcatenated convolutional codes, repeat-accumulate codes, re-peat-convolute codes, and generalizations of these codes obtainedby allowing nonlinear and large-memory constituent codes. It isshown that parallel-concatenated turbo codes and repeat-convo-lute codes with sub-linear memory are asymptotically bad. It isalso shown that depth-two serially concatenated codes with con-stant-memory outer codes and sublinear-memory inner codes areasymptotically bad. Most of these upper bounds hold even whenthe convolutional encoders are replaced by general finite-stateautomata encoders. In contrast, it is proven that depth-threeserially concatenated codes obtained by concatenating a repetitioncode with two accumulator codes through random permutationscan be asymptotically good.

 Index Terms—Asymptotic growth, concatenated codes, min-imum distance, repeat-accumulate-accumulate (RAA) codes,turbo codes.

I. INTRODUCTION

THE low-complexity and near-capacity performance of 

turbo codes [3], [9] has led to a revolution in coding

theory. The most famous casualty of the revolution has been

the idea that good codes should have high minimum distance:

the most useful turbo codes have been observed to have low

minimum distance.In this work, we provide general conditions under which

many constructions of turbo-like codes, including families of 

serially concatenated convolutional codes [2] and Repeat-Ac-

cumulate (RA) codes [5]–[7], must be asymptotically bad 1. We

also present a simple family of depth- serially concatenated

convolutional codes that are asymptotically good.

Our work is motivated by the analyses of randomly con-

structed parallel and serially concatenated convolutional codes

by Kahale and Urbanke [7] and of parallel concatenated turbo

codes with two branches by Breiling [4].

Manuscript received October 15, 2007; revised August 20, 2008. Current ver-sion published December 24, 2008. This work was supported by ARO under

Grant DAA L03-92-G-0115, by MURI under Grant DAAD19-00-1-0466, byNSF CCR: 9701304, and by NSF CCR: 9701304.

L. Bazzi is with the Department of Electrical and Computer Engineering,American University of Beirut, Beirut, Lebanon (e-mail: [email protected]).

M. Mahdian is with Yahoo! Research, Santa Clara, CA 95054 USA (e-mail:[email protected]).

D. A. Spielman is with the Department of Computer Science and Programin Applied Mathematics, Yale University, New Haven, CT 06520 USA([email protected]).

Communicated by T. J. Richardson, Associate Editor for Coding Theory.

Digital Object Identifier 10.1109/TIT.2008.2008114

1A sequence of codes of increasing block length is called an asymptoticallygood code if the message length and the minimum distance of the codes growslinearly with the block length. Codes for which either the message length orminimum distance grow sub-linearly with the block length are called asymptot-

ically bad .

Kahale and Urbanke [7] provided probabilistic estimates on

the minimum distance of randomly generated parallel concate-

nated turbo codes with a constant number of branches. They

also provided similar estimates for the minimum distance of 

the random concatenation of two convolutional codes with

bounded memory. In particular, Kahale and Urbanke proved

that if one builds a parallel concatenated code with branches

from random permutations and convolutional encoders of 

memory at most , then the resulting code has minimum dis-

tance at most and at least

with high probability, where is the number of message bits.

For rate serially concatenated convolutional codes with

random interleavers, they proved that the resulting code has

minimum distance at most and atleast with high probability, where is the free

distance of the outer code and is the inner code memory.

Breiling [4] proved that the parallel concatenation of two

convolutional codes with bounded memory has at most loga-

rithmic minimum distance, regardless of the choice of inter-

leaver. In particular, for parallel concatenated turbo codes with

two branches, Breiling proved that no construction could be

much better than a random construction: if the constituent codes

have memory , then the minimum distance of the resulting

code is .

These bounds naturally lead to the following five questions.

(Better than random?) Do there exist asymptoticallygood parallel concatenated turbo codes with more than

two branches or do there exist asymptotically good re-

peat-convolute or repeat-accumulate codes?

Note that the result of Breiling only applies to turbo codes

with two branches and the results of Kahale and Urbanke

do not preclude the existence of codes that are better than

the randomly generated codes.

(Larger memory?) What happens if we allow the mem-

ories of the constituent convolutional codes to grow with

the block length?

All the previous bounds become vacuous if the memory

even grows logarithmically with the block length.

(Nonlinearity?) Can the minimum distance of turbo-likecodes be improved by the use of non-linear constituent en-

coders, such as automata encoders?

(Concatenation depth?) Can one obtain asymptotically

good codes by serially concatenating a repetition code with

two levels of convolutional codes?

We will give essentially negative answers to the first three

questions and a positive answer to the last one. For parallel

concatenations and depth- serial concatenations of convolu-

tional codes and non-linear automata codes, we prove upper

bounds on the minimum distance of the resulting codes in terms

of the memories of the constituent codes. In Section II-A, we

show that parallel concatenated codes and repeat-convolute

0018-9448/$25.00 © 2009 IEEE

Page 2: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 2/10

BAZZI et al.: THE MINIMUM DISTANCE OF TURBO-LIKE CODES 7

codes are asymptotically bad if their constituent codes have

sublinear memory. These bounds hold even when the codes are

generalized by replacing the constituent convolutional codes

by automata codes. In Section II-B, we restrict our attention

to concatenations of ordinary convolutional codes, and obtain

absolute upper bounds that almost match the high-probability

upper bounds for random permutations obtained by Kahale andUrbanke.

In Section III-A, we show that depth-two serially concate-

nated codes are asymptotically bad if their inner code has sub-

linear memory and their outer code has constant memory. This

bound also applies to the generalized case of constituent au-

tomata codes.

In contrast, we show in Section III-B that depth-three con-

catenations of constant-memory codes can be asymptotically

good. In particular, we prove this for the random concatenation

of a repetition code with two accumulator codes.

 A. Turbo-Like Codes

The fundamental components of the codes we consider in

this paper are convolutional codes and their non-linear gen-

eralizations, which we call automata codes. The fundamental

parameter of a convolutional code that we will measure is its

memory—the number of registers in its encoder. The amount

of memory can also be defined to be the binary logarithm of 

the number of states in the encoder’s state diagram. A general

automata encoder is obtained by considering an encoder with

any deterministic state diagram. We will consider automata en-

coders that read one bit at each time step, and output a constant

number of bits at each time step. These can also be described as

deterministic automata or transducers with one input bit and a

constant number of output bits on each transition. We will againdefine the memory of an automata encoder to be the binary log-

arithm of its number of states.

Given a convolutional encoder and permutations

, each of length , we can define the parallel

concatenated turbo code with branches [3], [9] to be

the code whose encoder maps an input to

, where denotes the per-

mutation of the bits in according to and denotes the

output of the convolutional code on input .

Given an integer , we define the repeat- -times encoder, ,

to be the encoder that just repeats each of its input bits times.

Given a convolutional encoder , a message length , and a per-mutation of length , we define the repeat-convolute code

[5] to be the code whose encoder maps an input

to . That is, each bit of the input is repeated

times, the resulting bits are permuted, and then fed through

the convolutional encoder. We also assume that the input is

output as well. While some implementations do not include in

the output, its exclusion cannot improve the minimum distance

so we assume it appears. The number is called the repetition

 factor of the code. When the convolutional encoder is the ac-

cumulator (i.e., the map ), this code is called

a repeat-accumulate (RA) code [5].

Given two convolutional encoders and that output

and bits per time step respectively, an integer , anda permutation of length , we define the depth-two

serially concatenated convolutional code [2], [9] to be the

rate code whose encoder maps an input

to the codeword . The codes and are

called outer  and inner  codes, respectively. A classical ex-

ample of serially concatenated convolutional codes, and

that considered in [7], is a rate code given by the

map , where andare rate- convolutional codes. This fits into our framework 

with and .

One can allow greater depth in serial concatenation. The only

codes of greater depth that we consider will be repeat-accumu-

late-accumulate (RAA) codes. These are specified by a repeti-

tion factor , an integer , and two permutations and of 

length . Setting and to be accumulators, the resulting

code maps an input to .

We can generalize each of these constructions by allowing

the component codes to be automata codes. In this case, we will

refer to the resulting codes as generalized parallel concatenated 

turbo codes, generalized repeat convolute codes, and general-

ized serially concatenated codes.In practice, some extra bits are often appended to the input

so as to guarantee that some of the encoders return to the

zero state. As this addition does not substantially increase the

minimum distance of the resulting code, we will not consider

this technicality in this paper.

 B. Previous Results

Kahale and Urbanke [7] proved that if one builds a parallel

concatenated turbo code from a random interleaver and convo-

lutional encoders of memory at most , then the resulting code

has minimum distance at most and at least

with high probability. For rate serially concate-nated convolutional codes of the form mentioned in the previous

section with a random interleaver, they proved that the resulting

code has minimum distance at most

and at least with high probability, where is the

freedistance of the outer codeand is the inner codememory.

For parallel concatenated turbo codes with two branches,

Breiling [4] proved that no construction could be much better

than a random code: if the constituent codes have memory

, then the minimum distance of the resulting code is

.

Serially concatenated codes of depth greater than were

studied by Pfister and Siegel [8], who performed experimental

analyses of the serial concatenation of repetition codes with

levels of accumulators connected by random interleavers,

and theoretical analyses of concatenations of a repetition code

with certain rate-1 codes for large . Their experimental results

indicate that the average minimum distance of the ensemble

starts becoming good for , which is consistent with our

theorem. For certain rate- codes and going to infinity, they

proved their codes could become asymptotically good.

C. Our Results

In Section II-A, we upper bound the minimum distance

of generalized repeat-convolute codes and generalized par-

allel concatenated turbo codes. We prove that generalizedrepeat-convolute codes of message length , memory ,

Page 3: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 3/10

8 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, JANUARY 2009

and repetition factor have minimum distance at most

. The same bound holds for generalized par-

allel concatenated turbo codes with branches, memory ,

and message length . Therefore such codes are asymptotically

bad when is constant and is sublinear in . Note that

sublinear in corresponds to the case when the size of 

the corresponding trellis is subexponential, and so it includesthe cases in which the codes have natural subexponential

time iterative decoding algorithms. This proof uses techniques

introduced by Ajtai [1] for obtaining time-space tradeoffs for

branching programs. Comparing our upper bound with the

high-probability upper bound of Ka-

hale and Urbanke for parallel concatenated codes, we see that

our bound has a much better dependence on and a slightly

worse dependence on . A similar relation holds between our

bound and the upper bound of Breiling [4] for

parallel concatenated codes with branches.

In Section II-B, we restrict our attention to linear re-

peat-convolute codes, and prove that every repeat-convolute

code with repetition factor in which the convolutionalencoder has memory has minimum distance at most

. For even , this bound is very close to

the high-probability bound of Kahale and Urbanke.

In Section III-A, we study serially concatenated codes with

two levels, and prove that if the outer code has memory and

the inner code has memory , then the resulting code has min-

imum distance at most . Ac-

cordingly, we see that such codes are asymptotically bad when

and are constants and is sub-linear in . The

proof uses similar techniques to those used in Section II-A.

When specialized to the classical rate construction of se-

rially concatenated convolutional codes considered by Kahaleand Urbanke [7], our bound on the minimum distance becomes

. Comparing this with the high-

probability upper bound of Ka-

hale and Urbanke, we see that our bound is better in terms of 

, comparable in terms of , and close to their existential

bound of .

Finally, in Section III-B, we show that serially concatenated

codes of depth greater than two can be asymptotically good,

even if the constituent codes are repetition codes and accumu-

lators. In particular, we prove that randomly constructed RAA

codes are asymptotically good with constant probability.

Throughout this paper, our goal is to obtain asymptotic

bounds. We make no claim about the suitability of our bounds

for any particular finite .

II. REPEAT-CONVOLUTE-LIKE AND

PARALLEL TURBO-LIKE CODES

In this section we consider codes that are obtained by serially

concatenating a repeat- -times code with any code that

can be encoded by an automata (transducer) with at most

states and one output bit per transition. More precisely, if is

such an encoder, is a permutation of length , and is the

repeat- -times map, we define the generalized repeat-convolute

code to be the code whose encoder maps a stringto .

We consider also the parallel concatenated variations of these

codes. Given an automata with at most states and one

output bit per transition and permutations each of 

length , we define the generalized parallel concatenated turbo

code [3], [9] to be the code whose encoder maps

an input to

.

 A. An Upper Bound on the Minimum Distance

Theorem 1: Let be a constant integer, an automata

encoder with at most states, and an integer.

Let be a permutation of length . If , then

the minimum distance of the generalized repeat-convolute code

encoded by is at most

The same bound holds for the parallel concatenated turbo-like

code encoded by , where are permuta-

tions each of length .Proof: We first consider the case of repeat-convolute-like

codes. We explain at the end of the proof how to modify the

proof to handle parallel concatenated turbo-like codes.

To prove this theorem, we make use of a technique introduced

by Ajtai [1] for proving time-space tradeoffs for branching pro-

grams. In particular, for an input of length , the encoding

action of is naturally divided into time steps in which the

automata reads a bit of , outputs a bit, and changes state.

For convenience, we will let denote the set of  

time steps, and we will let denote the state of on input

at the end of the th time step.

Let denote the encoder . Following Ajtai [1], wewill prove prove the existence of two input strings, and , a

set of size at most , and

of size at most such that and may only

differ on bits with indices in and and may only

differ on time steps with indices in . The claimed bound on the

minimum distance of code encoded by will follow from the

existance of these two strings.

To construct the set , we first divide the set of time steps

into consecutive intervals, where is a parameter we will

specify later. We choose these intervals so that each has size

or . For example, if , and

we can divide into the intervals [1, 3], [4, 6],and [7, 8]. For each index of an input bit , we let

denote the multiset of time intervals in which reads input

bit (this is a multiset as a bit can appear multiple times in the

same interval). As each bit appears times, the multisets each

havesize . Asthere are intervals, there are at most possible

-multisets of intervals. So, there exists a set

of size at least and a multiset of intervals, , such that for

all , . Let be such a set with and

let be the corresponding set of intervals. Let . The set

will be the union of the intervals in .

Let be the last times in the time intervals in (e.g.,

in the above example the last time of the interval is ). For

each , that is zero outside , we consider the vectorof states of at times on input : .

Page 4: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 4/10

BAZZI et al.: THE MINIMUM DISTANCE OF TURBO-LIKE CODES 9

As the number of such possible sequences is at most and

the number of that are zero outside is , if 

(1)

then there should exist two different strings and that are both

zero outside of and such that for .To make sure that (1) is satisfied, we set

Our assumption that ensures that . Now, since

1) and agree outside ,

2) the bits in only appear in time intervals in , and

3) traverses the same states at the ends of time intervals in

on inputs and

must traverse the same states at all times in intervals outside

on inputs and . Thus, the bits output by

in time steps outside intervals in must be the same on inputsand . So and can

only disagree on bits output during times in the intervals in ,

and hence on at most bits. This means that the distance

between and is at most

as and

as .

Now, we explain how to apply the proof to the general-

ized parallel concatenated turbo codes. Let be

permutations each of length , be an automata encoder

with at most states and consider the parallel concatenated

turbo-like code encoded by . Let be the length-

permutation constructed from and the repetition

map in such a way that

for all . Let be the time-varying automata thatworks exactly like except that it is goes back to the start

state at the time steps . Thus

for all . In

other words, we can realize as a time-varying re-

peat-convolute-like code whose encoder maps a string

to . To extend the

minimum distance bound to generalized parallel concatenated

turbo codes, it is sufficient to note that the Proof of Theorem1 works without any changes for time-varying generalized

repeat-convolute codes, which the are natural generalizations of 

repeat-convolute-like codes obtained by allowing the automata

to be time-varying .2

Corollary 1: Let be a constant. Then, every generalized re-

peat-convolute code with input length and memory and

repetition factor and every generalized parallel concatenated

turbo code with input length , convolutional encoder memory

and branches has minimum distance .

Thus, such codes cannot be asymptotically good for sub-

linear in .

This means that if we allow to grow like , or even likefor some , the minimum relative distance of the code

will still go to zero. Moreover, sublinear in corresponds to

the case in which the size of the corresponding trellis is subex-

ponential, and therefore it includes all the cases in which such

codes have natural subexponential-time iterative decoding algo-

rithms.

It is interesting to compare our bound with that obtained by

Kahale and Urbanke [7], who proved that a randomly chosen

parallel concatenated code with branches has minimum dis-

tance with high probability. Theorem 1

has a much better dependence on and a slightly worse de-

pendence on . A similar comparison can be made with thebound of Breiling [4], who proved that every parallel concate-

nated code with branches has minimum distance at

most . In the next section, we prove an upper bound

whose dependence on and is asymptotically similar to that

obtained by these authors. B. Improving the Bound in the Linear, Low-Memory Case

We now prove that every repeat-convolute code with

repetition factor , memory , and input length , and

every parallel concatenated turbo code with branches,

memory , and input length has minimum distance at most

.

Theorem 2: Let and be integers and let be aconvolutional encoder with memory .

Let be a permutation of length . Assuming

, the minimum distance of the repeat-convolute

code encoded by is at most

Thus, when is constant, the minimum distance of the repeat-

convolute code encoded by is

2

A time-varying automata is specified by a state transition map  : S 2 

f  0  ; 1  g 2 T !   S  and an output map   : S 2 f  0  ; 1  g 2 T ! f  0  ; 1  g  , where S 

is the set of states and T  =  f  1  ; 2  ; 3  ; . . . g  is the set of time steps

Page 5: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 5/10

10 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, JANUARY 2009

If , the same bounds hold for the

parallel concatenated turbo code encoded by , where

are permutations each of length .

If we ignore constant factor, we see that our bound asymptot-

ically matches the bound of Breiling for parallel concatenated

turbo codes with two branches [4] (i.e., when ). The con-

stant factor in our bound is however larger. We have not at-tempted to optimize the constants in our proof. Our main objec-

tive is to establish, when , an asymptotic bound in terms

of the growth of and .

When is even, our bound asymptotically matches also the

bound for randomly constructed parallel-concatenated turbo

codes proved by Kahale and Urbanke [7]. As Kahale and Ur-

banke proved similar lower bounds for , we learn that the

minimum distances of randomly constructed turbo codes is not

too different from that of optimally constructed turbo codes.

Proof of Theorem 2: First we note that if the convolutional

code is non-recursive, it is trivial to show that on input

(i.e., a 1 followed by zeros) the output codeword will

have weight at most . Thus, without loss of generality, weassume that the convolutional code is recursive.

Our Proof of Theorem 2 will make use of the following fact

about linear convolutional codes mentioned in Kahale–Urbanke

[7]:

 Lemma 1 [7]: For any recursive convolutional encoder of 

memory , there is a number such that, after pro-

cessing any input of the form for any positive integer

, comes back to the zero state after processing the second .

In particular, the weight of the output of after processing any

such input is at most .

We consider first the case of repeat-convolute codes. We ex-plain at the end of the proof how to customize the proof to the

setting of parallel concatenated turbo codes.

Let be the number shown to exist in Lemma 1 for convolu-

tional code . As in [7] and [4], we will construct a low-weight

input on which also has low-weight by taking the

exclusive-or of a small number of weight inputs each of whose

two ’s are separated by a low multiple of ’s. As the code en-

coded by is a linear code, its minimum distance equals

the minimum weight of its codewords.

To construct this low-weight input, we first note that

every bit of the input appears exactly times in the string

. For every and every ,let denote the position of the th appearance of the

bit in . For each bit , consider the sequence

. Since there are

at most such possible sequences and input bits, there exists

a set of size at least such that all of its

elements induce the same sequence. That is, for all and in ,

is divisible by for all . From now on,

we will focus on the input bits with indices in , and construct

a low-weight codeword by setting some of these bits to .

Asin the Proof of Theorem 1, we now partition the set of time

steps into consecutive intervals, ,

each of length or , where is a parameter we

will specify later. For every index , we let the signature

of be the -tuple whose ’th component is the index of the in-

terval to which belongs.

Now, we construct a hypergraph as follows: has parts,

each part consisting of vertices which are identified with the

intervals . There are hyperedges in , one corre-

sponding to each input bit with index in . The vertices con-

tained in the hyperedge are determined by the signature of thecorresponding bit: if input bit has signature ,

then the th hyperedge contains the ’th vertex of the th part,

for every . Thus, is a -partite -uniform hy-

pergraph (i.e., each hyperedge contains exactly vertices, each

from a different part) with vertices in each part and

edges.

We now define a family of subgraphs such that if con-

tains one of these subgraphs, then the code encoded by

must have a low-weight codeword. We define an -forbidden

subgraph of to be a set of at least one and at most hyper-

edges in such that each vertex in is contained in an even

number of the hyperedges of . (One can think of an -forbidden

subgraph as a generalization of a cycle of length to hyper-graphs). In Lemma 2, we prove that if contains an -forbidden

subgraph then the code encoded by has a codeword of 

weight at most . In Lemma 3, we prove that if  

contains at least edges, then it contains a for-

bidden subgraph. As has at least edges, if we set

then ; so, Lemma 3 will imply that has a

-forbidden subgraph and Lemma 2 will imply that the

code encoded by has a codeword of weight at most

. Plugging in the value we havechosen for , we find that the minimum distance of the code

encoded by is at most

as

where the last inequality follows from , and the

second-to-last inequality follows from combining this in-

equality with the assumption in the theorem that to

show , and applying the bound

for all .

Now, we explain how to modify the proof to handle par-

allel concatenated turbo codes. Let be permutations

Page 6: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 6/10

BAZZI et al.: THE MINIMUM DISTANCE OF TURBO-LIKE CODES 11

each of length be a recursive convolutional encoder with

memory , and consider the parallel concatenated turbo code

encoded by . We can associate with the re-

peat-convolute encoder , where is the length-

permutation constructed from and the repetitionmap

in such a way that for all

. To extend the bound to parallel-concatenated turbocodes, we will force and have the same input-output be-

havior on the special low-weight inputs considered in this proof.

To do this, we set

and require that be the first

times in the intervals in which they appear. We then guarantee

that, on the special low-weight inputs considered in the proof,

the convolutional encoder will be in the zero state at steps

, and so it will have the same output

as . The rest of the analysis is similar, except that we use the

slightly stronger assumption .

 Lemma 2: If contains an -forbidden subhypergraph, then

there is an input sequence of weight at most whose corre-

sponding codeword in has weight at most .

Proof: Let denote the set of hyperedges of the -for-

bidden subhypergraph in , and consider the set of bits of 

the input that correspond to the hyperedges of . By definition,

and . We construct an input of weight at most

by setting the bits in to and other bits to , and consider

the codeword corresponding to . As each

vertex of is contained in an even number of the hyperedges

in , each interval in contains an even number of bits that

are in . Thus, by the definition of and Lemma 1,

is zero everywhere except inside those intervals of 

that contain a bit that is in . Since there are at most

such intervals, the weight of is at most .

Therefore, the weight of the codeword corresponding to is at

most .

 Lemma 3: Every -partite -uniform hypergraph with

vertices in each part and at least hyper-edges contains a

-forbidden subhypergraph.

Proof: We construct a bipartite graph from as follows.

For every -tuple where is a vertex

in the th part of , we put a vertex in the first part of , andfor every -tuple where is a vertex in

the th part of , we put a vertex in the second part of . If 

there is a hyperedge in , where is a vertex

of the th part, we connect the vertices and

in .

By the above construction, each edge in corresponds to a

hyperedge in . There are at least edges and at most

vertices in . Thus, by Lemma 4 below, has a cycle

of length at most . It is easy to see

that the hyperedges corresponding to the edges of this cycle con-

stitute a -forbidden subhypergraph in .

 Lemma 4: Let be a graph on vertices with at leastedges. Then, has a cycle of length at most .

Proof: We first prove the theorem in the case that every

vertex of has degree at least 3. In this case, if the shortest cycle

in the graph had length , then a breadth-first search tree

of depth from any vertex of the graph would contain at least

distinct vertices. As ,

this would be a contradiction for . So, the graph must

contain a cycle of length at most .We may prove the lemma in general by induction on . As-

sume the lemma has been proved for all graphs with fewer than

vertices, and let be a graph on vertices with at least

edges. If the degree of every node in is at least , then

has a cycle of length at most by the preceding argu-

ment. On the other hand, if has a vertex of degree , we con-

sider the graph obtained by deleting this vertex and its two

adjacent edges. The graph has vertices and at least

edges, and so by induction has a cycle of length at

most . As is a subgraph of also has a

cycle of length at most , which proves

the lemma.

III. SERIALLY CONCATENATED CODES

In this section, we consider codes that are obtained by seri-

ally concatenating convolutional codes and, more generally, au-

tomata codes. In Section III-A, we prove an upper bound on the

minimum distance of the concatenation of a low-memory outer

automata encoder with an arbitrary inner automata encoder. In

particular, we prove that if the memory of the outer code is

constant and the memory of the inner code is sub-linear, then

the code is asymptotically bad. In contrast, in Section III-B, we

prove that if the input is first passed through a repetition code

and a random permutation, then the code is asymptotically good

with constant probability, even if both convolutional encodersare accumulators.

 A. Upper Bound on the Minimum Distance When the Outer 

Code is Weak 

In this section, we consider the serial concatenation of au-

tomata codes. We assume that each automata outputs a constant

number of bits per transition. This class of codes includes

the serially concatenated convolutional codes introduced by

Benedetto, Divsalar, Montorsi, and Pollara [2] and studied by

Kahale and Urbanke [7]. If the outer code has constant memory

and the inner code has sub-linear memory, then our bound

implies that the code cannot be asymptotically good.Formally, we assume that ( , respectively) is an au-

tomata encoder with at most ( , respectively) states and

( , respectively) output bits per transition. For an integer

and a permutation of length , we define to be

the encoder that maps an input to the codeword

. We will assume

without loss of generality that , and are such that this

mapping is an injective mapping. The encoders and are

called the outer and inner encoders, respectively.

Theorem 3: Let be an automata encoder with at most

states that outputs bits at each time step, and let be an au-

tomata encoder with at most states that outputs bits ateach time step. For any positive integer and any permutation

Page 7: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 7/10

12 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, JANUARY 2009

of length , the minimum distance of the serially concate-

nated code encoded by is at most

In particular, if is constant (and and are constants),

the minimum distance of the serially-concatenated code en-coded by is

and consequently any such family of codes is asymptotically

bad as long as is sub-linear in .

Proof: The proof follows the same outline as the Proof of 

Theorem 1. We begin by setting to be the set

of times steps in the computation of on input ,

and setting to be the set of times steps in the

computation of on input . We simi-

larly, let denote the sequence of states traversed

by on input and denote the sequence of states

traversed by on input .

To prove the claimed bound on the minimum distance of the

code encoded by , we will prove the existence of two

distinct input strings and , a set , a set

, and a set such that and are both on bits not

in and only differ for , and and

only differ for . The minimum distance bound will

then follow from an upper bound on the size of .

To construct these sets, we make use of parameters and

to be determined later. We first partition the set intointervals each of size or , and we partition the

set into intervals each of size or .

As outputs at most bits during the time steps

in an interval in , the bits output by during an interval in

are r ead b y d uring a t most intervals in . A s there

are fewer than sets of at most intervals

in , there exists a set of at least intervals in

such that all the bits output by during these intervals are

read by during a single set of at most intervals in

. Let denote the set of at least intervals in

and let denote the corresponding set of at most

intervals in . We then let denote the set of input bits read by

during the intervals in . As all the intervals in have size

at least , we have . The set will be the union

of the intervals in and will be the union of the intervals in

.

Let and denote the last time steps in the

intervals in and , respectively. For each that

is zero outside , we consider , the sequence

of states traversed by on at times , and,

, the sequence of states traversed by on input

at times . There are at most

such pairs of sequences. So, if 

(2)

then there are two distinct and in that are both 0

outside and a pair of sequences and

such that for all

and for all . This

means that the bits output and states traversed by on inputs

and are the same at time steps outside the time intervals in

, and therefore the bits output and states traversed by oninputs and are the same outside time steps

in intervals in . Thus

(3)

As this bound assumes (2), we will now show that for

and

this assumption is true.

Our setting of reduces (2) to

which would be implied by

(4)

To derive this inequality, we first note that since for

Rearranging terms, we find this implies

Again rearranging terms, we obtain

which implies

By now dividing both sides by and recalling

, we derive (4).

Finally, the bound on the minimum distance of the code now

followsby substitutingthechosen values for and into(3).

We now compare this with the high-probability upper bound

of on the minimum distance of rate

random serially concatenated convolutional codes obtained

by Kahale and Urbanke [7]. In their case, we have ,and our upper bound becomes .

Page 8: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 8/10

BAZZI et al.: THE MINIMUM DISTANCE OF TURBO-LIKE CODES 13

Fig. 1. An RAA code.

We note that the dependence of our bound on is comparable,and the dependence of our bound on is much better.

 B. A Strong Outer Code: When Serially Concatenated Codes

 Become Asymptotically Good 

The proof technique used in Theorem 3 fails if the outer code

is not a convolutional code or encodable by a small finite au-

tomata. This suggests that by strengthening the outer code one

might be able to construct asymptotically good codes. In fact,

we will prove that the serial concatenation of an outer repeat-ac-

cumulate code with an inner accumulator yields an asymptoti-

cally good code with some positive probability.

Let be an integer, be the repeat- -times map,

and be accumulators 3, be an integer, and and be

permutations of length . We define to be the en-

coder that maps input strings to the codeword

. We call the code en-

coded by an RAA (Repeat, Accumulate, and Accumu-

late) code (See Fig. 1). We note that this code has rate .

In contrast with the codes analyzed in Theorem 3,

these RAA codes have a repeat-accumulate encoder,

where those analyzed in The-

orem 3 merely have an automata encoder.

Theorem 4: Let and be integers, and let and

be permutations of length chosen uniformly at random. Then

for each constant , there exists a constant and an

integer , such that the RAA code encoded by has

minimum distance at least with probability at least for

all .

So specifically, there exists an infinite family of asymptoti-

cally good RAA codes.

Proof: Conditions bounding the size of will be appear

throughout the proof.

Let denote the expected number of non-zero codewords

in the code encoded by of weight less than or equal to

. Taking a union bound over inputs and applying linearity of 

expectation, we see that the probability the minimum distance

of the code encoded by is less than is at most .

Thus, we will bound this probability by bounding .

To bound , we use techniques introduced by Divsalar,

Jin and McEliece [5] for computing the expected input-output

weight enumerator of random turbo-like codes. For an accumu-

lator code of message length , let denote the number of 

inputs of weight on which the output of the accumulator has

weight . Divsalar, Jin and McEliece [5] prove that

(5)

3 While Q  and Q  are identical as codes, we give them different names toindicate their different roles in the construction.

where is defined to be zero if . Therefore, if the input tois a random string of length and weight , the probability

that the output has weight is

(6)

Now consider a fixed input . If has weight

and is a random permutation of length , then

is a random string of length and weight . This random

string is the input to the accumulator . Therefore, by (6), for

any the probability that the output of has weight is. If this happens, the input to will be a random

string of weight , and therefore, again by (6), the probability

that the output of has weight will be equal to .

Thus, for any fixed input string of weight , and any fixed

and , the probability over the choice of and that the

output of has weight and the output of (which is also

the output of ) has weight is equal to

Thus, by the linearity of expectation, the expected number of 

non-zero codewords of the code encoded by of weight

at most equals

as the terms with or are zero.

Using the inequalities ,

and , for positive integers and , we

bound this sum by the equation at the top of the following page.

The summand in that expression is at maximum when .

Therefore

Page 9: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 9/10

14 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 1, JANUARY 2009

(7)

since for all . To bound (7), note that the sum

has the form

where , and

. If we can guarantee that

(8)

for all , we can use the bound

(9)

We can express (8) as . Thus (8) holds

for all the desired values of if , or

equivalently

which can be guaranteed when

(10)

via the bounds

where we need in the second inequality. Going back to(7), we get via (9) and (10) that

(11)

when

and

or, equivalently, when

and

(12)

It follows from (11) and (12), that for each , and for each

constant , there is constant such when is

sufficiently large.

While the constants we obtain are not particularly sharp, theyare sufficient to prove the existence of asymptotically good fam-

ilies of depth-three serially-concatenated codes based on accu-

mulators.

This result should be compared with the work of Pfister and

Siegel [8], who performed experimental analyses of the serial

concatenation of repetition codes with levels of accumulators

connected by random interleavers, and theoretical analyses of 

concatenations of a repetition code with certain rate- codes

for large . Their experimental results indicate that the average

minimum distance of the ensemble starts becoming good for

, which is consistent with our theorem. For certain rate- codes

and going to infinity, they proved their codes could become

asymptotically good. In contrast, we prove this for andaccumulator codes.

Page 10: A Min Dist of Turbo Like Codes

7/30/2019 A Min Dist of Turbo Like Codes

http://slidepdf.com/reader/full/a-min-dist-of-turbo-like-codes 10/10

BAZZI et al.: THE MINIMUM DISTANCE OF TURBO-LIKE CODES 15

IV. CONCLUSION AND OPEN QUESTIONS

We derived in Section II-A a worst-case upper bound on the

minimum distance of parallel concatenated convolutional codes,

repeat convolute codes, and generalizations of these codes ob-

tained by allowing nonlinear and large-memory automata-based

constituent codes.The bound implies that such codes are asymp-totically bad when the underlying automata codes have sub-

linear memory. In the setting of convolutional constituent codes,

a sublinear memory corresponds to the case when the size of the

corresponding trellis is sub-exponential, and so it includes the

cases in which the codes have natural subexponential time it-

erative decoding algorithm. In Section II-B, we improved the

bound in the setting of low-memory convolutional constituent

codes. We leave the problem of interpolating between the two

bounds open:

• Is it possible to interpolate between the bounds of Theo-

rems 1 and 2?

Then, we derived in Section III-A a worst-case upper bound

on the minimum distance of depth- serially concatenated

automata-based codes. Our bound implies that such codes are

asymptotically bad when the outer code has a constant memory

and the inner code has a sublinear memory. This suggests the

following question:

• If one allows the memory of the outer code in a depth-

serially concatenated code to grow logarithmically with the

block length, can one obtain an asymptotically good code?

In contrast, we provedin Section III-B that RAAcodes, which

are depth- serially concatenated codes obtained by concate-

nating a repetition code with two accumulator codes through

random permutations, can be asymptotically good. This result

naturally leads to the following open questions:• Can one obtain depth- serially concatenated codes with

better minimum distance by replacing the accumulators

in the RAA codes with convolutional codes of larger

memory? Also, can one improve the minimum distance

bounds on the RAA codes?

• Can the RAA codes be efficiently decoded by iterative de-

coding, or any other algorithm?

ACKNOWLEDGMENT

The authors would like to thank Sanjoy Mitter for many stim-

ulating conversations and the anonymous reviewers for many

useful comments on the paper.

REFERENCES

[1] M. Ajtai, “Determinism verus nondeterminism for linear time RAMswith memory restrictions,” J. Comput. Syst. Sci., vol. 65, no. 1, pp.2–37, 2002.

[2] S. Benedetto, D. Divsalar, G. Montorsi, and F. Pollara, “Serial concate-nation of interleaved codes: Performance analysis, design, and iterativedecoding,” IEEE Trans. Inf. Theory, vol. 44, no. 3, pp. 909–926, 1998.

[3] C. Berrou, A. Glavieux, and P. Thitimajshima, “Near Shannon limit

error-correctingcoding and decoding: Turbo codes,” in Proc. IEEE Int.Conf. Commun. (ICC-1993), 1993, pp. 1064–1070.

[4] M. Breiling, “A logarithmic upper bound on the minimum distance of turbo codes,” IEEE Trans. Inf. Theory, vol. 50, no. 8, pp. 1692–1710,2004.

[5] D. Divsalar, H. Jin, and R. J. McEliece, “Coding theorems for “turbo-like” codes,” in Proc. 36th Annu. Allerton Conf. Commun., Contr.,Comput., Monticello, IL, USA, Sep. 1998, pp. 201–210.

[6] H. Jinand R. J. McEliece,“RA codes achieveAWGN channel capacity.In Marc,” in AAECC , P. C. Fossorier, H. Imai, S. Lin, and A. Poli,Eds. Berlin, Germany: Springer, 1999, vol. 1719 , Lecture Notes inComputer Science, pp. 10–18.

[7] N. Kahale and R. Urbanke, “On the minimum distance of parallel andserially concatenated codes,” in Proc. IEEE Int. Symp. Inf. Theory,Cambridge, MA, Aug. 1998, p. 31.

[8] H. D. Pfister andP. H. Siegel, “The serialconcatenation of rate-1codesthrough uniform random interleavers,” IEEE Trans. Inf. Theory, vol.49, no. 6, pp. 1425–1438, 2003.

[9] B. Vuceticand J. S.Yuan , TurboCodes: Principles and Applications.New York: Kluwer Academic, 2000.

Louay Bazzi received the Ph.D. degree from the Department of Electrical Engi-neering and Computer Science at Massachusetts Institute of Technology (MIT),Cambridge, in 2003.

He is currently an Assistant Professor in the Electrical and Computer Engi-neering Department at the American University of Beirut. His research interestsinclude coding theory, psuedorandomness, complexity theory, and algorithms.

Mohammad Mahdian received the B.S. degree in computer engineering fromSharif University of Technology in 1997, the M.S. degree in computer sciencefrom University of Toronto in 2000, and the Ph.D. degree in mathematics from

Massachusetts Institute of Technology, Cambridge, in 2004.He has worked as an intern and a Postdoctoral Researcher at IBM Research

Labs and Microsoft Research, and is currently a Research Scientist at Yahoo!Research Lab in Santa Clara, CA. His current research interests include algo-rithm design, algorithmic game theory, and applications in online advertisingand social networks.

Daniel A. Spielman (A’97–M’03) was born in Philadelphia, PA, in 1970. Hereceived the B.A. degree in mathematics and computer science from Yale Uni-versity, New Haven, CT, in 1992, and the Ph.D. degree in applied mathematicsfrom Massachusetts Institute of Technology, Cambridge, in 1995.

From 1996 to 2005, he was an Assistant and then an Associate Professor of Applied Mathematics at M.I.T. Since 2005, he has been a Professor of AppliedMathematics and Computer Science at Yale University.

Dr. Spielman is the recipient of the 1995 Association for Computing Ma-chinery Doctoral Dissertation Award, the 2002 IEEE Information Theory So-ciety Paper Prize, and the 2008 Gödel Prize, awarded by the European Asso-ciation for Theoretical Computer Science and the Association for ComputingMachinery.


Recommended