encoders and methods for designing encoders for Low density parity check (ldpc) and other block codes are presented. An efficient and systematic method for designing partially parallel encoders is presented. A parallelism factor is selected such that the end result for the encoder is similar to the partially parallel G matrix multiplication method. In addition to the method an initial circuit is given for the G matrix multiplication encoder and the RU encoder. A circuit for the hybrid encoder is presented which achieves less power consumption and smaller area than an equivalent encoder based on the G matrix multiplication with a smaller critical path than previous encoders.

Patent
   7657816
Priority
Jul 13 2005
Filed
Jul 13 2006
Issued
Feb 02 2010
Expiry
Jun 09 2028
Extension
697 days
Assg.orig
Entity
Small
16
9
EXPIRED
1. A ldpc block code encoder, comprising:
a first matrix multiplier coupled to the input information bits and a first memory unit used to generate first set of parity values,
a second matrix multiplier coupled to the output of the said first matrix multiplier and the input information bits,
a third matrix multiplier coupled to the output of the second matrix multiplier and a second memory unit used to generate the second set of parity values.
9. An integrated circuit having a substrate and a low density parity check block code encoder comprising:
a first matrix multiplier coupled to the input information bits and a first memory unit used to generate first set of parity values,
a second matrix multiplier coupled to the output of the first matrix multiplier and the input information bits,
a third matrix multiplier coupled to the output of the second matrix multiplier and a second memory unit used to generate the second set of parity values.
17. A method to implement a ldpc block code encoder, comprising the steps:
a) converting the H matrix to an upper triangular form,
b) computing the first set of parity values by a first matrix multiplication,
c) Computing an intermediate result by a second matrix multiplication,
d) computing the second set of parity values by using the result of the matrix multiplication in part (c), wherein this step is implemented comprising the sub-steps:
(i) selecting a method to implement this computation based on either matrix multiplication or back substitution or other elimination methods,
(ii) selecting a level of parallelism if needed, and
(iii) applying look-ahead to the method in (i) using the level of parallelism in step (ii) if needed.
2. The block code encoder of claim 1, wherein said encoder is part of a communications transceiver.
3. The ldpc code encoder of claim 1, wherein the ldpc code is a regular code.
4. The ldpc code encoder of claim 1, wherein the ldpc code is an irregular code.
5. The ldpc block code encoder of claim 1, wherein said matrix multipliers are implemented by a plurality of AND gates and EX-OR (exclusive-or) gates.
6. The ldpc block code encoder of claim 1, wherein the said third matrix multiplier is implemented using back substitution.
7. The ldpc block code encoder of claim 1, where the said third matrix multiplier is implemented using a plurality of binary weighted sums, wherein said binary weighted sums are formed by a plurality of AND gates and EX-OR gates.
8. The ldpc block code encoder of claim 1, wherein the said third matrix multiplier is implemented using an elimination method.
10. The integrated circuit of claim 9, wherein the said ldpc block code encoder is part of a communications transceiver.
11. The integrated circuit of claim 9 comprising the block code encoder, wherein the said ldpc code is a regular code.
12. The integrated circuit of claim 9 comprising the ldpc block code encoder, wherein the said ldpc code is an irregular code.
13. The integrated circuit of claim 9 comprising the ldpc block code encoder, wherein the said matrix multipliers are implemented by a plurality of AND gates and EX-OR (exclusive-or) gates.
14. The integrated circuit of claim 9 comprising the ldpc block code encoder, wherein the said third matrix multiplier is implemented using back substitution.
15. The integrated circuit claim 9 comprising the ldpc block code encoder, wherein the said third matrix multiplier is implemented using a plurality of binary weighted sums, wherein said binary weighted sums are formed by a plurality of AND gates and EX-OR gates.
16. The integrated circuit of claim 9 comprising the ldpc block code encoder, wherein the said third matrix multiplier is implemented using an elimination method.

This application claims the benefit of the U.S. Provisional Application No. 60/699,171 filed Jul. 13, 2005, which is incorporated herein by reference in its entirety.

This invention was made with Government support from the National Science Foundation (NSF) under Grant No. 0441632, SBIR Phase I: Design of a 10-Gigabit Ethernet Transceiver Over Copper.

The present invention relates to data processing, transmission, and digital communication. More specifically, it is related to encoder design for error correcting codes.

Error correcting codes are a protection mechanism to ensure reliable transmission through noisy communication channels. The main principle of error correcting codes is to add redundancy to the information at the transmitter. This redundancy allows error detection and error correction at the receiver. Error correcting codes fall into one of several categories: block codes and tree codes. Block codes are memoryless codes whereas tree codes require memory (George C. Clark and J. Bibb Cain, Error-Correction Coding for Digital Communications, Plenum Press, 1981). There are several important block codes in use today such as Low Density Parity Check (LDPC), BCH, and Reed-Solomon codes. The most common tree codes are Convolutional codes. Block codes can be found in everyday products and services, where digital communication is used, like digital video, networks, hard disk drives, and satellites.

Block codes can be represented by two matrices (George C. Clark and J. Bibb Cain, Error-Correction Coding for Digital Communications, Plenum Press, 1981). One matrix defines the parity checks such that HcT=0 for a codeword (c) that is free of errors. This matrix is referred to as the parity check matrix (H). The other matrix is referred to as the Generator matrix (G). Together the generator matrix and the parity check matrix have the following relationship: (HGT=0). This relationship means that all codewords defined in (G) are valid codewords. Furthermore if an error occurs such that the received information is r=c+e. Then the error will generate a syndrome HrT=HcT+HeT=HeT which can be used to easily detect the error locations. In a systematic code, where c=[s p] and s contains the k information values and p contains the parity vector of n-k values. Then the Generator and parity check matrices are defined as G=[Ik P] and H=[−PT In-k], where P represents the parity check matrix, Ix represents the identity matrix of size (x by x).

The primary challenges with error correcting codes (ECC) is achieving near optimal use of available bandwidth and minimizing encoding and decoding complexity. ECCs that achieve near optimal use of available bandwidth are said to be near the Shannon limit. Unfortunately, these codes suffer from high decoding and encoding complexity. LDPC codes fall into this category of requiring high decoding and encoding complexity to achieve near Shannon limit performance (D. J. C. MacKay and R. M. Neal, “Near Shannon limit performance of low density parity check codes,” Electronic Letters, Vol. 32, pp 1645-1646, August 1996).

Recently, the near Shannon limit LDPC codes have become important to Industry. They have found their way into consumer standards such as Digital Video Broadcasting (DVB), 10 Gigabit Ethernet over copper (10GBase-T), and high speed wireless (IEEE 802.11n). The main disadvantages to using LDPC codes in these standards are 1) implementing LDPC decoders requires a significant amount of power, area, and latency and 2) implementing efficient partially parallel encoders requires a significant amount of power and area (and/or storage) overhead. Therefore, efficient implementations of LDPC codes are important.

The present invention focuses on solving problem 2) from above. LDPC encoders can be implemented in several different ways depending on the specific LDPC code. In all cases the Generator (G) matrix multiplication method is valid. Due to the higher complexity of the LDPC decoder it is often more efficient to fold the matrix multiplication operation. Unfortunately, folding the matrix multiplication looses the advantage of the sparseness of the G matrix. The Richardson-Urbanke (RU) method (T. J Richardson and R. L. Urbanke, “Efficient Encoding of Low Density Parity Check Codes”, IEEE Transactions on Information Theory, Vol. 47 No. 2, February 2001) was proposed as an alternative encoder which takes advantage of the sparseness of the H matrix. Although the RU method achieves savings on implementation costs, it is primarily intended as a parallel design and suffers from high complexity when implemented in hardware. Designing partially parallel RU LDPC encoders is a challenging problem.

Low complexity block code encoders are important in minimizing the overall power consumption and area costs for digital transmitter systems. What is needed is a systematic method for designing partially parallel low complexity block code encoders (and circuits) that achieve minimal power consumption and area costs.

The present invention provides an efficient partially-parallel implementation of block code encoders and describes a method for designing efficient partially parallel block code encoders.

In accordance with the present invention, the H matrix is transformed into a more efficient form referred to as approximate lower triangular form with a lower triangular T matrix for encoding. This form in conjunction with the G matrix forms the mathematical equations for the Hybrid encoder. The main challenge is to convert the back substitution operation into a partially parallel operation without a long critical path. The inversion and multiplication operation, with the T matrix, is modified to generate a partially parallel inversion and multiplication operation without a long critical path or significant cycle latency. There are several options for inversion and multiplication which are discussed in the present invention. Then a method is described for generating an inversion and multiplication method for partially parallel systems without sacrificing critical path or substantial storage or area overhead. This method is combined with the G matrix multiplication method such that the initial parity values that cannot be computed with the T matrix are generated in parallel or in partially parallel by the G matrix. The final encoder design yields a partially parallel encoder which requires less power and area than a partially parallel G matrix multiplication encoder.

Further embodiments, features, and advantages of the present invention, along with structure and operation of various embodiments of the present invention, are discussed in detail below with reference to the accompanying figures.

The present invention is described with reference to the accompanying figures. In the figures, like reference numbers indicate identical or functionally similar elements. Additionally, the left-most digit or digits of a reference number identify the figure in which the reference number first appears. The accompanying figures, which are incorporated herein and form part of the specification, illustrate the present invention and, together with the description, further serve to explain the principles of the invention and to enable a person skilled in the relevant art to make and use the invention.

FIG. 1. illustrates a partially parallel encoder based on the Generator (G) matrix multiplication method for block codes.

FIG. 2. illustrates approximate lower triangular form for the H matrix.

FIG. 3. illustrates the Richardson-Urbanke (RU) encoder method for block codes.

FIG. 4. illustrates the low complexity partially parallel Hybrid encoder for block codes.

FIG. 5. illustrates an example of 2-Parallel back substitution.

FIG. 6. illustrates the low complexity partially parallel Hybrid encoder with parallel back substitution for block codes.

Modern digital communication systems require partially parallel error correcting code encoders for efficient power utilization and minimal area consumption. The encoding process for block codes can be described by the mathematical matrix multiplication formula c=sG=[s p1 p2], where s is the input information vector, G is the generator matrix, and c is the resulting codeword vector consisting of s and the parity values vector [p1 p2]. Although fully parallel matrix multiplication encoders are possible to implement it is more area and power efficient to implement the matrix multiplication with the circuit in FIG. 1. FIG. 1 illustrates circuit 100 the partially parallel G matrix multiplication encoder.

Circuit 100 consists of blocks 101, 102 and 103. Block 101 is the storage memory for the G matrix. In an embodiment, block 101 is a read only memory (ROM). In another embodiment, block 101 is random access memory (RAM). In other embodiments, other devices may be used. Block 102 is the field multiplication logic which multiplies the input information vector (s) by the appropriate entries in the G matrix memory. In an embodiment, block 102 consists of several logical AND gates. In other embodiments, other devices may be used such as multiplexers. Block 103 is the field summation (addition) trees which perform a summing operation over a field to generate the parity bits (pi's). In an embodiment, block 103 consists of logical exclusive or (XOR) trees. In other embodiments, other devices such as adders may be used.

Circuit 100 operates in a partially parallel manner such that if L parity values are generated per cycle then it must read L*rowsize(P) values per cycle. In a systematic code the parity values are generated in ceiling of (colsize(P)/L) cycles. The primary advantage of the G matrix multiplication encoder is that it is easily folded to generate partially parallel designs. The primary disadvantage with this encoder is it requires large storage overhead for the coefficients.

Although the G matrix multiplication method leads to simple encoder designs, it is not necessarily the most efficient in terms of computation complexity, power consumption, or area requirements. The Richardson-Urbanke (RU) encoder method uses the relationship of the information values and the parity values with the H matrix to define efficient encoding methods. The first step in the RU method is to convert the ordinary H matrix into an approximate lower triangular form as in FIG. 2.

FIG. 2 illustrates block 200 the approximate lower triangular form as defined in (T. J Richardson and R. L. Urbanke, “Efficient Encoding of Low Density Parity Check Codes”, IEEE Transactions on Information Theory, Vol. 47 No. 2, February 2001). Block 200 consists of blocks 201 (A), 202 (B), 203 (T), 204 (C), 205 (D), and 206 (E). The dimensions of blocks 201 to 206 are set by maximizing the size of block 203 (T) through the process of swapping entries in the original H matrix such that block 203 is lower triangular in form. In an embodiment, row swapping is used to determine the size of block 203. In another embodiment column swapping is used to determine the size. In another embodiment column additions may be used. In other embodiments, other swapping methods or addition methods or combinations may be used.

The new H matrix will now be in the following approximate lower triangular form:

H = ( A B T C D E ) ,
where T is lower triangular and Φ=−ET−1B+D is non-singular (in GF(2) for LDPC codes). The next step is Gaussian elimination (which works whether T is singular or not) to yield the new H′ matrix of the form:

H = ( A B T - ET - 1 A + C - ET - 1 B + D 0 ) .

Note, if Φ=−ET−1B+D is singular in GF(2) then it is necessary to swap columns to generate a non-singular matrix. (Note: swapping columns at this stage means one will need to swap columns back prior to decoding which basically makes this a non-systematic encoding approach. Another method is to perform row swapping but this may suffer from a larger Φ matrix which significantly degrades encoding performance. Richardson and Urbanke suggested using the permuted H matrix as the decoding matrix but this may lead to a less efficient decoder if the original H matrix had special properties which simplify the decoder.)

Now the new H′ matrix gives us two equations to use to solve for the parity checks. Assuming codeword c=[s p1 p2] where s contains the systematic information vector and p1, p2 are parity vectors with HcT=H′cT=0 for all valid codewords. Then the two equations are given by:
AsT+Bp1T+Tp2T=0
(−ET−1A+C)sT+(−ET−1B+D)p1T+0=0.
The two equations above allow one to solve for the parity vectors (p1, p2) using only the supplied information vector (s) such that:
p1T=−Φ−1(−ET−1A+C)sT
p2T=−T−1(AsT+Bp1T).
Assuming the use of pre-computation, one can perform the p1 calculation with one dense matrix multiplication or several smaller sparse multiplications and one smaller dense multiplication. Similarly the p2 calculation can be computed by two small sparse matrix multiplications and one back substitution or two larger and denser matrix multiplications. FIG. 3 includes an example implementation of the parallel RU encoder.

FIG. 3 illustrates circuit 300 the Richardson-Urbanke (RU) parallel encoder. Circuit 300 consists of blocks 301, 302, 303, 304, 305, 306, 307, and 308. Block 301 consists of the matrix multiplication of A (block 201) with input S. In an embodiment, block 301 is implemented as a matrix multiplication. In other embodiments, other methods may be used. Block 302 consists of the matrix multiplication of C (block 204) with input S. In an embodiment, block 302 is implemented as a matrix multiplication. In other embodiments, other methods may be used. Block 303 consists of the matrix multiplication of inverse of T (block 203) with output from block 301. In an embodiment, block 303 is implemented as matrix multiplication. In another embodiment, block 303 is implemented as back substitution. In another embodiment, block 303 is implemented using an iterative matrix inversion technique. In other embodiments, other methods may be used. Block 304 consists of the matrix multiplication—E (block 206) with the output from block 303. In an embodiment, block 304 is implemented as a matrix multiplication. In other embodiments, other methods may be used. Block 305 consists of an adding function. In one instance it adds the output of block 304 with the output of block 302. In another instance it adds the output of block 301 with the output of block 307. In an embodiment, block 305 is implemented as an array of adders. In other embodiments, other devices may be used such as logical XOR gates. Block 306 consists of the matrix multiplication of the negative of the inverse of Φ=(−ET−1B+D), where E is block 206, T is block 203, B is block 202, and D is block 205. In an embodiment, block 306 is implemented as a matrix multiplication. In another embodiment, block 306 is implemented with iterative matrix inversion. In other embodiments, other methods may be used. Block 307 is matrix multiplication of block 202 with the output of block 306. In an embodiment, block 307 is implemented as a matrix multiplication. In other embodiments, other methods may be used. Block 308 is a matrix multiplication by the negative of the inverse of the T (block 203) matrix with the output of block 307. In an embodiment, block 308 is implemented with back substitution. In another embodiment, block 308 is implemented with matrix multiplication. In other embodiments, other methods can be used such as iterative matrix inversion.

Circuit 300, the parallel RU encoder, is able to compute the parity values in V cycles where V is the number of pipelining cutsets (K. K. Parhi, VLSI Digital Signal Processing Systems Design and Implementation, John Wiley & Sons, 1999). The parallel RU encoder computes the parity values with seven matrix multiplications of varying sizes, which are significantly faster than the G matrix multiplication, and two additions. The primary advantage of the RU encoder is that it takes advantage of the sparseness of the H matrix to minimize area overhead. Unfortunately, there is no simple relationship between columns and folding or time-multiplexing technique (K. K. Parhi, VLSI Digital Signal Processing Systems: Design and Implementation, John Wiley & Sons, 1999) is unable to take advantage of the sparseness. The RU encoder has several disadvantages: 1) it has long routing overhead, 2) long critical path, 3) multiple steps which cannot be folded, and 4) dependencies which stall computations. Although pipelining can solve problem 2) the other problems cannot be easily fixed.

The last two encoder designs suffered from several problems. First the partially parallel G matrix encoder required significant storage overhead and power requirements due to large number of memory values which need to be read per cycle which makes it less ideal for implementation. The RU Method suffered from a long critical path, inability to be easily implemented as a partially parallel design, and odd constraints which lead to a non-systematic encoder. Therefore the next method, which is the proposed invention, is an improvement over these two methods. This method is the low complexity Hybrid encoder which requires significantly less storage and maintains a systematic encoder form as in FIG. 4.

FIG. 4 illustrates circuit 400 the partially parallel low complexity Hybrid encoder. Circuit 400 consists of blocks 401, 402, 403, 404, and 405. Block 401 computes the first parity values based on the G matrix. In an embodiment, block 401 is implemented as a fully parallel matrix multiplication. In another embodiment, block 401 is implemented as a partially parallel matrix multiplication. In other embodiments, other methods are used. Block 402 stores a select portion of the G matrix coefficients. Block 402 is optional. In an embodiment, block 402 is implemented as read only memory (ROM). In another embodiment, block 402 is implemented as random access memory (RAM). In other embodiments, other devices are used. Block 403 is a matrix multiplication of [A B] (blocks 201 and 202) with the input S and the output of block 401 (the first set of parity values (P)). In an embodiment, block 403 is implemented as a fully parallel matrix multiplication. In other embodiments, other methods are used. Block 404 is a matrix multiplication by the inverse of block 203 (T). In an embodiment, block 404 is implemented as a fully parallel matrix multiplication. In another embodiment, block 404 is implemented as a partially parallel matrix multiplication. In another embodiment block 404 is implemented as a fully parallel back substitution. In another embodiment block 404 is implemented as a partially parallel back substitution. In other embodiments, other methods are used. Block 405 is the storage element for the coefficients of (block 203) the T matrix. Block 405 is optional. In an embodiment, block 405 is implemented as a read only memory (ROM). In another embodiment, block 405 is implemented as a random access memory (RAM). In another embodiment, block 405 is implemented as shift registers. In other embodiments, other devices may be used.

The main idea is to compute the p1 values by using the G Matrix. Therefore, one does not need the inverse of the Φ matrix as in the RU method. This method proposes to compute the p1 values sequentially as in the partially parallel G matrix encoder. However, instead of using the G matrix to compute p2 values this method uses the special property (p2T=−T−1(AsT+Bp1T)) from the RU method to compute p2 using a parallel sparse matrix multiplication ([A B]*[s p1]T=AsT+Bp1T) and a sequential back substitution (or matrix multiplication) computation. This has two advantages. First this method has a shorter critical path when the back substitution is implemented in an L-parallel fashion with small L and second the memory storage requirements can be significantly reduced because T is lower triangular and much smaller than the parity check equations (P) from the G matrix. Therefore this effectively cuts the memory storage in half when using back substitution versus matrix multiplication.

The following is the mathematical representation of the encoder.
p1T=s*G(:, k+1:n−(m−g)), where columns k+1 to n−(m−g) are used
p2T=−T−1(AsT+Bp1T)
where, k is the number of information values, n is the number of values in the codeword, g is the column and row size of Φ, and m is the number of rows in the H matrix.

The following steps are required to design the Hybrid encoder:

The previous method did not discuss details about efficient partially parallel implementations of the T−1 matrix multiplication. Several different methods exist for the T−1 matrix multiplication such as the pre-computed T−1 folded matrix multiplication which is similar to the partially parallel G matrix encoder. However, this method suffers from increased complexity in storage overhead because the T−1 matrix is not lower triangular like the T matrix. Another method is referred to as back substitution which relies on the T matrix. However, it suffers from a long critical path (when implemented in parallel) and intra-cycle dependencies (when folded). Finding a solution which overcomes these problems is important.

Next is an example of back substitution for a 4 bit example. Back substitution is defined as the problem given the equation Ty=x, with values for T and x, find y. The full matrix formula is described by

[ 1 0 0 0 t 2 , 1 1 0 0 t 3 , 1 t 3 , 2 1 0 t 4 , 1 t 4 , 2 t 4 , 3 1 ] [ y 1 y 2 y 3 y 4 ] = [ x 1 x 2 x 3 x 4 ] .

Solving for the yi's becomes a sequential operation shown below:
y1=x1
y2=x2⊕t2,1y1
y3=x3⊕t3,1y1⊕t3,2y2
y4=x4⊕t4,1y1⊕t4,2y2⊕t4,3y3

In FIG. 5, the look-ahead technique (K. K. Parhi, VLSI Digital Signal Processing Systems: Design and Implementation, John Wiley & Sons, 1999) was applied to provide an example for a 2-parallel back substitution method.

FIG. 5 illustrates circuit 500 the example 2-parallel back substitution method for block code encoders. Circuit 500 consists of blocks 501, 502, and 503. Block 501 performs a multiplication. In an embodiment, block 501 is implemented as logical AND gates. In other embodiments, other devices may be used. Block 502 performs a summing operation. In an embodiment, block 502 is implemented with logical exclusive or gates. In other embodiments other devices may be used. Block 503 performs a shifting operation. In an embodiment, block 503 is implemented by a shift register. In other embodiments, other devices may be used.

Circuit 500 is an efficient implementation of an L-parallel back substitution which reduces the critical path. It produces L parity values per cycle and operates in the ceiling of (rowsize(T)/L) cycles. The primary advantage of this method lies in balanced critical path between the field summation of the yi values and the field summation of the xi values. In essence, this method can be considered a partial Gaussian elimination method. This method converts the T matrix into the T′ and M matrices. The new matrix M represents the xi summation tree. The main disadvantage of this method is the additional memory overhead for the M matrix.

An example of the 2-parallel look-ahead technique for back substitution is presented next. Given the following matrix multiplication for Ty=x

[ 1 0 0 0 ( 1 ) 1 0 0 1 0 1 0 1 0 ( 1 ) 1 ] [ y 1 y 2 y 3 y 4 ] = [ x 1 x 2 x 3 x 4 ] ,
where the dependencies are shown with parentheses applying the look-ahead technique amounts to adding the odd row with the even row to create a new matrix multiplication where additional work has been moved to the right hand side of the equation as seen below

[ 1 0 0 0 0 1 0 0 1 0 1 0 0 0 0 1 ] [ y 1 y 2 y 3 y 4 ] = [ x 1 x 1 + x 2 x 3 x 3 + x 4 ] = [ 1 0 { 1 } 1 ] [ 1 0 { 1 } 1 ] [ x 1 x 2 ] [ x 3 x 4 ] .

This improvement has come at a cost and now extra coefficient storage is required for the right hand side of the equation. The entries with braces around them correspond to the values which must be stored. Total additional storage overhead to remove the dependencies is defined as the number of lower triangular values times the number of cycles for the encoder.

The following steps are required to design the L-parallel back substitution:

FIG. 6 illustrates Circuit 600 the partially parallel low complexity Hybrid encoder. Circuit 600 is composed of blocks 401, 402, 403, 601, 602, 603, 604, 605, and 606. Blocks 401 to 403 are identical in functionality to circuit 400. Block 601 is the field summation tree for the yi's. In an embodiment, block 601 is implemented as adder trees. In another embodiment, block 601 is implemented as a logical exclusive or (XOR) tree. In other embodiments, other devices may be used. Block 602 holds the T′ matrix coefficients. In an embodiment, block 602 is implemented as read only memory (ROM). In another embodiment, block 602 is implemented as random access memory (RAM). In another embodiment, block 602 is implemented as shift registers. In other embodiments, other devices may be used. Block 603 is the field summation tree for the xi's. In an embodiment, block 603 is implemented as adder trees. In another embodiment, block 603 is implemented as a logical exclusive or (XOR) tree. In other embodiments, other devices may be used. Block 604 is holds the M matrix coefficients. In an embodiment, block 604 is implemented as read only memory (ROM). In another embodiment, block 604 is implemented as random access memory (RAM). In another embodiment, block 604 is implemented as shift registers. In other embodiments, other devices may be used. Block 605 is a field addition which combines the results from the yi's summation and the xi's summation. In an embodiment, block 605 is implemented with an adder. In another embodiment, block 605 is implemented with a logical exclusive or gate. In other embodiments, other devices may be used. Block 606 is a memory element to store the previous yi results for the feedback portion of the back substitution. In an embodiment, block 606 is implemented with an L-parallel shift register. In other embodiments, other devices may be used.

Parhi, Keshab K., Cohen, Aaron E.

Patent Priority Assignee Title
10216574, Oct 24 2012 Western Digital Technologies, Inc. Adaptive error correction codes for data storage systems
7913149, Dec 20 2006 AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE LIMITED Low complexity LDPC encoding algorithm
8473824, Sep 08 2008 CAVIUM INTERNATIONAL; MARVELL ASIA PTE, LTD Quasi-cyclic low-density parity-check (QC-LDPC) encoder
8605383, May 21 2012 Western Digital Technologies, Inc. Methods, devices and systems for characterizing polarities of piezoelectric (PZT) elements of a two PZT element microactuator
8797664, Dec 22 2012 Western Digital Technologies, Inc. Polarity detection of piezoelectric actuator in disk drive
8966339, Dec 18 2012 Western Digital Technologies, INC Decoder supporting multiple code rates and code lengths for data storage systems
8972826, Oct 24 2012 Western Digital Technologies, Inc.; Western Digital Technologies, INC Adaptive error correction codes for data storage systems
9021339, Nov 29 2012 Western Digital Technologies, INC Data reliability schemes for data storage systems
9059736, Dec 03 2012 Western Digital Technologies, INC Methods, solid state drive controllers and data storage devices having a runtime variable raid protection scheme
9122625, Dec 18 2012 Western Digital Technologies, INC Error correcting code encoder supporting multiple code rates and throughput speeds for data storage systems
9153283, Sep 30 2014 Western Digital Technologies, INC Data storage device compensating for hysteretic response of microactuator
9203434, Mar 09 2012 Western Digital Technologies, Inc.; Western Digital Technologies, INC Systems and methods for improved encoding of data in data storage devices
9214963, Dec 21 2012 Western Digital Technologies, INC Method and system for monitoring data channel to enable use of dynamically adjustable LDPC coding parameters in a data storage system
9294130, Sep 08 2008 CAVIUM INTERNATIONAL; MARVELL ASIA PTE, LTD Quasi-cyclic low-density parity-check (QC-LDPC) encoder
9495243, Dec 18 2012 Western Digital Technologies, Inc. Error correcting code encoder supporting multiple code rates and throughput speeds for data storage systems
9619317, Dec 18 2012 Western Digital Technologies, INC Decoder having early decoding termination detection
Patent Priority Assignee Title
6757122, Jan 29 2002 Seagate Technology LLC Method and decoding apparatus using linear code with parity check matrices composed from circulants
6928602, Jul 18 2001 Sony Semiconductor Solutions Corporation Encoding method and encoder
7313752, Aug 26 2003 Samsung Electronics Co., Ltd.; SAMSUNG ELECTRONICS CO , LTD Apparatus and method for coding/decoding block low density parity check code in a mobile communication system
7447235, Oct 08 2003 Qualcomm Incorporated FEC-based reliability control protocols
7484159, Mar 05 2004 Sony Corporation Encoding method and encoding apparatus
7493551, Sep 22 2004 STMICROELECTRONICS INTERNATIONAL N V Method and device for delivering punctured code words encoded with a LDPC code
7499490, Jun 24 2005 The Regents of the University of California Encoders for block-circulant LDPC codes
7502987, May 12 2004 Postech Academy Industry Foundation Apparatus and method for encoding and decoding block low density parity check codes with a variable coding rate
7559008, Oct 03 2005 Seagate Technology LLC Nested LDPC encoders and decoder
////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Jul 13 2006Leanics Corporation(assignment on the face of the patent)
Jul 13 2006COHEN, AARON E Leanics CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0180650551 pdf
Jul 13 2006PARHI, KESHAB KLeanics CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0180650551 pdf
Jan 18 2016Leanics CorporationPARHI, KESHAB KNUNC PRO TUNC ASSIGNMENT SEE DOCUMENT FOR DETAILS 0375120055 pdf
Date Maintenance Fee Events
Apr 17 2013M2551: Payment of Maintenance Fee, 4th Yr, Small Entity.
Sep 18 2017REM: Maintenance Fee Reminder Mailed.
Mar 05 2018EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
Feb 02 20134 years fee payment window open
Aug 02 20136 months grace period start (w surcharge)
Feb 02 2014patent expiry (for year 4)
Feb 02 20162 years to revive unintentionally abandoned end. (for year 4)
Feb 02 20178 years fee payment window open
Aug 02 20176 months grace period start (w surcharge)
Feb 02 2018patent expiry (for year 8)
Feb 02 20202 years to revive unintentionally abandoned end. (for year 8)
Feb 02 202112 years fee payment window open
Aug 02 20216 months grace period start (w surcharge)
Feb 02 2022patent expiry (for year 12)
Feb 02 20242 years to revive unintentionally abandoned end. (for year 12)