UC BERKELEY
EECS technical reports
TECHNICAL REPORTS


EECS-2007-168.pdf
Conditions of Use

Archive Home Page

Rethinking the Minimum Distance: Channels With Varying Sampling Rate and Iterative Decoding of LDPC Codes

Authors:
Dolecek, Lara
Technical Report Identifier: EECS-2007-168
December 19, 2007
EECS-2007-168.pdf

Abstract: In this dissertation we develop novel coding theoretic approaches for two problems relevant to modern communication systems. In the first part we address the issue of reliable communication under varying sampling rate, while in the second part we focus on the analytic understanding of the performance of low density parity check (LDPC) codes in the low bit error rate (BER) region. The underlying theme in both of these somewhat non-standard yet relevant problems is that the notion of a fundamental performance metric, typically taken to be the minimum distance of an additive error correcting code, needs to be rethought when the standard assumptions on the communication no longer hold.

We first investigate the problem of overcoming inadequate timing recovery from a coding theoretic perspective. We show how to systematically thin first order Reed-Muller codes, as a representative example of highly structured additive error correcting codes with good minimum distance properties, such that the resulting thinned code is immune to additive errors and a synchronization error. The method heavily relies on the novel run-length properties also developed here.

We also propose and study number theoretic constructions of sets of strings immune to multiple repetitions, also shown to have good cardinalities. These constructions are used to develop a prefixing-based method to improve the immunity of an arbitrary code to repetition errors. This judiciously chosen prefix carries asymptotically negligible redundancy. We also provide a low complexity message passing decoding algorithm that can handle both repetitions and additive errors.

In the second part, we study the performance of iteratively decoded LDPC codes in the low BER regime. Due to the prior limited analytical tools available to address and guarantee their performance in this regime, deployment of these powerful codes has still not met the original promise.

In order to gain a better understanding of the low BER performance of LDPC codes, we introduce the notion of a combinatorial object that we call an absorbing set. This object is viewed as a stable point of the bit-flipping algorithm, an algorithm that can be viewed as an asymptotic 1-bit approximation to many message passing decoding algorithms. Since absorbing sets are fixed points of the message passing algorithms, the decoder can get stuck in an absorbing set that is not a codeword. In particular, if there are absorbing sets smaller that the minimum distance of the code, the decoder is likely to converge to these objects. As a result, under iterative decoding, the low BER performance will be dominated by the number and the size of dominant absorbing sets, rather that the number of minimum distance codewords and the minimum distance itself, which is considered to be the performance metric under the maximum likelihood decoding and the key property of a code.

As a case study, we analyze the minimal absorbing sets of high-rate array-based LDPC codes. We provide a comprehensive analytic description of the minimal absorbing sets for this family of codes. In this study, we demonstrate the existence of absorbing sets whose weight is strictly smaller than the minimum distance of the code. These minimal absorbing sets, rather than minimum distance codewords, are also experimentally verified to dominate the low BER performance, further confirming the importance of our theoretical contributions.