This preview shows pages 1–2. Sign up to view the full content.
This preview has intentionally blurred sections. Sign up to view the full version.
View Full Document
Unformatted text preview: Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 30: Achieving the BSC p capacity (II) Tuesday, November 6, 2007 Lecturer: Atri Rudra Scribe: Nathan Russell & Atri Rudra In the last lecture, we started with the description of our BSC p capacity achieving code C * , which is a concatenated code C out C in , where C out and C in satisfying the following properties: (i) C out : The outer code with block length N and rate 1 2 over F 2 k , with k = O (log N ) . Further, the outer code has a unique decoding algorithm D out that can correct at most fraction of worstcase errors in time T out ( N ) . (ii) C in : The inner code has dimension k , dimension n and a rate of 1 H ( p ) / 2 . Further, there is a decoding algorithm D in that runs in T in ( k ) time and has decoding error probability no more than 2 over BSC p . In todays lecture, we will analyze the properties of C * and also see how to get our hands on C out and C in with the desired properties. For the rest of the lecture, we will assume that p is an absolute constant. Note that this implies that k = ( n ) and thus, we will use k and n interchangeably in our asymptotic bounds. Finally, we will use N = nN to denote the block length of C * . 1 Decoding Error Probability We begin this section by analyzing the natural decoding algorithm that we saw in the last lecture: Step 1 : Let y i = D in ( y i ) , 1 i N . Step 2 : Run D out on y = ( y 1 ,...,y N ) . By the properties of D in , for any fixed i , there is an error at y i with probability 2 . Each such error is independent, since errors in BSC p itself are independent by definition. Because of this, and by linearity of expectation, the expected number of errors in y is N 2 . Taken together, those two facts allow us to conclude that, by the Chernoff bound, the probability that the total number of errors will be more than N is at most e N 6 . Since the decoder D out fails only when there are more than N errors, this is also the decoding error probability. Expressed in asymptotic terms, the error probability is 2 ( N n ) ....
View Full
Document
 Spring '11
 RUDRA
 Algorithms

Click to edit the document details