On the Construction of the Kernel Matrix by Primitive BCH Codes for Polar Codes

Abstract

The polar codes defined by the kernel matrix are a class of codes with low coding-decoding complexity and can achieve the Shannon limit. In this paper, a novel method to construct the 2n-dimensional kernel matrix is proposed, that is based on primitive BCH codes that make use of the interception, the direct sum and adding a row and a column. For ensuring polarization of the kernel matrix, a solution is also put forward when the partial distances of the constructed kernel matrix exceed their upper bound. And the lower bound of exponent of the 2n-dimensional kernel matrix is obtained. The lower bound of exponent of our constructed kernel matrix is tighter than Gilbert-Varshamov (G-V) type, and the scaling exponent is better in the case of 16-dimensional.

Share and Cite:

Lin, L. (2022) On the Construction of the Kernel Matrix by Primitive BCH Codes for Polar Codes. Communications and Network, 14, 23-35. doi: 10.4236/cn.2022.141003.

1. Introduction

Polar codes can achieve the Shannon limit for binary-input discrete memoryless channels (BI-DMC) in theory and with low encoding and decoding complexity [1]. Polar codes employ the n-th Kronecker power of the matrix G 2 = ( 1 0 1 1 ) to encode 2 n channels, where G 2 is called the polarizing matrix or kernel matrix. As the number of channels grows, sub-channel becomes either a noiseless channel or a pure-noise channel, the proportion of the noiseless channels is close to the original channel capacity, and the noiseless channels transmit the information bits without error.

The kernel matrix is an important part of polar codes, which determines the polarization property of polar codes. The kernel matrix G 2 is also called 2-dimensional kernel matrix, there are other kernel matrices, and different kernel matrices will have different polarization effects. In order to improve the polarization of the polarization code, many researchers have done a lot of research on the kernel matrix.

There are two extended research directions for the construction of the kernel matrix, one is to expand the field of the kernel matrix, the other is to increase the dimension of the kernel matrix. Generally speaking, the larger the field or the higher the dimension of the kernel matrix, the better the polarization effect of the polar codes. Şaşoğlu et al. [2] [3] [4] generalized the polar code to any discrete memoryless channel and q-ary field, which expanded the field of the kernel matrix. Korada [5] proposed to construct a ( 2 n 1 ) -dimensional kernel matrix with BCH codes, which increased the dimension of kernel matrix. E. Moskovskaya et al. [6], based on [5] [7], put forward a method to construct a 2 n -dimensional kernel matrix with extended BCH codes, which further increased the dimension of kernel matrix. E. Moskovskaya et al. stacked the matrix blocks to construct a 2 n × ( 2 n 1 ) matrix, and then added a column of parity bits to construct a 2 n -dimensional kernel matrix. Although the 2 n -dimensional kernel matrix had a higher dimension and exponent, it did not consider the problem of partial distance’s upper bound.

Therefore, based on [5] [6], we take advantage of the primitive BCH codes to design a higher-dimensional kernel matrix which meets the upper bound of partial distance. Compared with the work in [5], the proposed construction enjoys two advantages, one is that the kernel matrix has a higher dimension, and the other is that the obtained kernel matrix is naturally lower triangular, guaranteeing the polarization property of kernel matrix. And compared with the work in [6], a solution is given to adjust these rows whose partial distances are beyond their upper bounds. The comparison result shows that the proposed 2 n -dimensional kernel matrix has a tighter lower bound of the exponent than the Gilbert-Varshamov (G-V) type construction in [5], the scaling exponent is not very different from [6], and 16-dimensional kernel matrix is even slightly better than [6].

2. Preliminaries

This section reviews the knowledge of primitive BCH codes and background of polar codes, including the polarization condition, the scaling exponent, the partial distance, and the exponent.

2.1. Primitive BCH Codes

Definition 1 (Primitive BCH Codes [8] ). Let α be the primitive element, and g ( x ) be the lowest degree polynomial with α , α 2 , , α 2 t as the roots in finite field G F ( 2 ) . The code generated by g ( x ) is called a primitive BCH code, denoted as B C H ( N , k , t ) . It has a code length N = 2 n 1 , encodes k symbols and corrects at most t errors.

For example, if α is the primitive element of polynomial g ( x ) = x 4 + x + 1 , the finite field G F ( 2 4 ) can be generated by g ( x ) , the minimum polynomial of all elements can be obtained as follows:

α , α 2 , α 4 , α 8 x 4 + x + 1 = ϕ 1 ( x ) . (1)

α 3 , α 6 , α 9 , α 12 x 4 + x 3 + x 2 + x + 1 = ϕ 3 ( x ) . (2)

α 5 , α 10 x 2 + x + 1 = ϕ 5 ( x ) . (3)

α 7 , α 11 , α 13 , α 14 x 4 + x 3 + 1 . (4)

If t = 1 and n = 15 , then the generator polynomial of B C H ( 15 , 11 , 1 ) is

g ( x ) = ϕ 1 ( x ) = x 4 + x + 1 . (5)

If t = 2 and n = 15 , then the generator polynomial of B C H ( 15 , 7 , 2 ) is

g ( x ) = L C M ( ϕ 1 ( x ) , ϕ 3 ( x ) ) = ( x 4 + x + 1 ) ( x 4 + x 3 + x 2 + x + 1 ) = x 8 + x 7 + x 6 + x 4 + 1. (6)

where L C M ( ) denotes the Least Common Multiple of its inputs.

If t = 3 and n = 15 , then the generator polynomial of B C H ( 15 , 5 , 3 ) is

g ( x ) = L C M ( ϕ 1 ( x ) , ϕ 3 ( x ) , ϕ 5 ( x ) ) = x 10 + x 8 + x 5 + x 4 + x 2 + x + 1 . (7)

Lemma 1 (Minimum Distance [8] ). For any positive integer n , t ( n 3 , t 2 n 1 ) , if there is a cyclic code of length 2 n 1 , the number of parity bits is N k n t , and the minimum distance d min 2 t + 1 .

2.2. The Performance and Property of Polar Code

Polar code can produce polarization phenomena, which can polarize channel into pure noise channel or noise-free channel. These matrices that can produce polarization phenomena are polarization matrices, and the kernel matrix actually is a polarization matrix. Given a l × l matrix G l , the necessary and sufficient condition for judging whether it is a polarization matrix is that the matrix is non-singular, and the upper triangular matrix is not formed after any column transformation [3]. Therefore, kernel matrix G l is non-singular, and G l also can be called l-dimensional kernel matrix.

The polar code polarizes the sub-channels capacity I ( W ) to 0 or 1, but there are some sub-channels whose channel capacity is between 0 and 1, these sub-channels are called unpolarized channels. The speed of unpolarized channel’s channel capacity tends to 1 is called the polarization speed. Fazeli et al. [9] [10] applied the scaling exponent to measure the polarization speed of polar code. The scaling exponent was proposed on the basis of the scaling assumption. The definition of the scaling assumption is as follows:

Definition 2 (scaling assumption [11] ). Given a kernel matrix G l and binary discrete memoryless channel W, there exists

lim n number of unpolarized channels number of total channels N 1 μ ( 0 , + ) . (8)

where μ denotes the scaling exponent, and N = 2 l ( l 2 ) denotes the length of polar code.

The optimal value of scaling exponent is 2, the closer the scaling exponent of kernel matrix is to 2, the better the polarization effect. Most of the scaling exponents of kernel matrices are between 3 and 5, but Yao et al. [12] proved that when the dimension of kernel matrix is 64, the scaling exponent can be less than 2.9603. The scaling exponent can be calculated by

N 1 / μ ( C R ) = β , (9)

where β = ( 1 + 2 P e 0.01 ) 3 , p e , R and C denote block error rate, ratio and channel capacity [9], respectively.

The exponent is also an index to measure the speed of polarization. The value of the exponent is between 0 and 1. The larger the exponent, the better the polarization of polar code. The exponent is determined by the partial distance sequence, which is defined as follows:

Definition 3 (partial distance [5] ). For the kernel matrix G l = [ g 1 T , g 2 T , , g l T ] T , its i-th partial distance D i is

D i = d H ( g i , g i + 1 , , g l ) , i = 1 , 2 , , l 1 , (10)

D l = d H ( g l , 0 ) . (11)

where g i ( i = 1 , 2 , , l ) denotes the row vector of length l, d H ( a , b ) denotes the Hamming distance of the vector a and b, d H ( a , C ) = min c C d H ( a , c ) , g 1 , , g k denotes the linear space generated by g 1 , , g k , 0 denotes the zero vector of length l.

According to the partial distance sequence, the definition of the exponent is as follows:

Definition 4 (exponent [5] ). The exponent of kernel matrix G l is defined as

E ( G l ) = 1 l i = 1 l log l D i . (12)

In this paper, all operations on matrix elements are XOR on G F ( 2 ) , i.e., 1 + 0 = 1 mod ( 2 ) = 1 , 1 + 1 = 2 mod ( 2 ) = 0 .

The size of the square matrix in this paper is described by the concept of order and dimension. For example, a 3*3 matrix can be expressed as a 3-dimensional matrix or a 3-order matrix in this paper. The 3-dimensional matrix indicates that the matrix is non-singular, and the 3-order matrix only indicates the size of the matrix.

3. The Construction of 2n-Dimensional Kernel Matrix

In this section, we employ the classification formula to construct a ( 2 n 1 ) -order matrix, adding a column vector and a row vector of all 1s before the first column and below the last row of the ( 2 n 1 ) -order matrix, and the 2 n -dimensional kernel matrix can be obtained.

3.1. The Construction of the ( 2 n 1 ) -Order Matrix

Definition 5 (Classification Formula [13] ). According to the classification formula of cyclotomic coset, the integer set { 0 , 1 , , 2 n 1 } is classified as

S = i = 0 2 n 2 { 2 k i mod ( 2 n 1 ) : k N } = j = 1 c v ( j ) . (13)

where v ( i ) denotes the elements in the i-th partitioned set with i ( 1 i c ) , and c denotes the number of cyclotomic cosets in S.

According to the Definition 5, we can obtain the following theorem 1:

Theorem 1. Let l ( i ) denotes the number of elements in v ( i ) , and l ( 1 ) = 1 , l ( 2 ) = l ( 3 ) = n can be obtained.

Proof. v ( 1 ) = { 0 2 k mod ( 2 n 1 ) , k N } = { 0 } , v ( 2 ) = { 1 2 k mod ( 2 n 1 ) , k N } , v ( 3 ) = { 3 2 k mod ( 2 n 1 ) , k N } can be known from Equation (13), therefore l ( 1 ) = 1 . for v ( 2 ) , when k = n , ( 2 n 1 ) = 1 = 1 2 0 mod ( 2 n 1 ) , therefore v ( 2 ) = { 1 , 2 , 4 , , 2 n 1 } and l ( 2 ) = n . Similarly, v ( 3 ) = { 3 , 6 , 12 , , 3 2 n 1 } and l ( 3 ) = n can be obtained.£

For example, n = 4 and S = { 0 , 1 , , 14 } = { { 0 } , { 1 , 2 , 4 , 8 } , { 3 , 6 , 9 , 12 } } , therefore, c = 5 , l ( 1 ) = 1 , l ( 2 ) = l ( 3 ) = 4 , l ( 4 ) = 2 , l ( 5 ) = 4 , v ( 1 ) = { 0 } , v ( 2 ) = { 1 , 2 , 4 , 8 } , v ( 3 ) = { 3 , 6 , 9 , 12 } , v ( 4 ) = { 5 , 10 } , v ( 5 ) = { 7 , 11 , 13 , 14 } .

According to the classification formula (13), we use c different sub-matrices to construct a ( 2 n 1 ) -order matrix by stacking the matrix blocks from top to bottom. The specific construction process is as follows:

1) The first layer sub-matrix is a zero matrix of dimension l ( 1 ) .

2) The second layer sub-matrix is a unit matrix with dimension l ( 2 ) .

3) The third layer sub-matrix is intercepting from the generator matrix of the primitive BCH code with dimension l ( 3 ) and 1 error correction capability. The interception criterion is that the rows of intercepted matrix are continuous, and the intercepted matrix can form a sub-lower triangular matrix with the previous two layers.

4) If c > 3 , looking for other generating matrices of the primitive BCH codes with length ( 2 n 1 ) and error correction capability t ( t = 2 , 3 , , c 2 ) in turn, then intercepting the matrix according to the l ( i ) ( i = 4 , 5 , , c ) and interception criterion.

The above construction of the ( 2 n 1 ) -order matrix is summarized in Figure 1.

3.2. A Row and Column Addition

Through adding a column and a row vector of all 1s before the first column and below the last row of the ( 2 n 1 ) -order matrix, the 2 n -dimensional kernel matrix is obtained. The 2 n -dimensional kernel matrix is a lower triangular matrix, which guarantees the polarization.

For example, in order to construct the 16-dimensional matrix, the 15-order matrix should be constructed firstly. When n = 4 , t = 1 , the generator polynomial of the primitive BCH code is Equation(5), and c = 5 , l ( 1 ) = 1 , l ( 2 ) = l ( 3 ) = 4 , l ( 4 ) = 2 , l ( 5 ) = 4 .

15-order matrix can be constructed with 5 layers sub-matrices, and the dimension of theses sub-matrices are 1, 4, 4, 2 and 4, and the construction of the

Figure 1. The construction of ( 2 n 1 ) -order matrix.

16-dimensional kernel matrix is as follows:

( 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 ) t = 1 ( 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 )

t = 2 ( 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 0 0 0 0 1 1 0 0 1 0 1 0 0 0 1 0 1 1 1 0 0 0 1 0 0 0 1 0 1 1 1 0 )

t = 3 ( 0 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 1 1 0 0 1 0 0 0 0 1 1 0 0 1 0 0 0 0 1 1 0 0 1 0 0 0 0 1 1 0 0 1 1 0 0 0 1 0 1 1 1 0 0 1 0 0 0 1 1 1 0 0 0 1 1 1 0 0 0 1 1 1 0 1 0 1 1 1 1 1 0 1 0 0 1 1 0 1 0 0 1 1 1 1 0 0 1 1 0 1 0 0 1 0 1 0 0 1 0 1 0 0 1 0 0 0 0 1 0 0 0 0 )

a row and column addition ( 1 0 1 1 0 0 0 1 0 1 0 0 1 0 0 1 0 1 0 0 0 1 1 1 1 0 0 1 0 0 0 1 0 1 1 0 0 1 0 0 1 0 0 1 1 0 0 1 0 1 0 0 0 1 1 0 0 1 1 1 0 0 0 1 0 1 1 1 0 1 0 1 0 0 0 1 0 1 1 1 1 1 1 1 0 1 1 0 0 1 0 1 0 0 0 0 1 0 1 1 1 0 1 1 0 0 1 0 1 0 0 0 1 0 0 1 1 1 0 1 1 0 0 1 0 1 0 0 1 0 0 0 1 1 1 0 1 1 0 0 1 0 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 )

Filling 0 in the remaining blank positions, and 16-dimension kernel matrix can be obtained.

( 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 1 1 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 1 1 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 1 1 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 1 1 0 0 1 0 0 0 0 0 0 0 1 1 0 0 0 1 0 1 1 1 0 0 0 0 0 0 1 0 1 0 0 0 1 0 1 1 1 0 0 0 0 0 1 1 1 1 0 1 1 0 0 1 0 1 0 0 0 0 1 0 1 1 1 0 1 1 0 0 1 0 1 0 0 0 1 0 0 1 1 1 0 1 1 0 0 1 0 1 0 0 1 0 0 0 1 1 1 0 1 1 0 0 1 0 1 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 ) . (14)

3.3. The Analysis of Construction

From Lemma 1, ( N k ) min = n and the lowest degree of the generator polynomial ( g ( x ) ) min = n can be obtained. If α is the primitive of g ( x ) , then G F ( 2 n ) is the n-th extension of G F ( 2 ) and β G F ( 2 n ) , β = a 0 + a 1 α + a 2 α 2 + + a n 1 α n 1 , ( a 0 , , a n 1 F 2 ) [14]. Because G F ( 2 n 1 ) = G F ( 2 n ) / { 0 } , the element in G F ( 2 n 1 ) also can be expressed linearly by 1 , α , α 2 , , α n 1 . If the above process is extended to the matrix, a similar conclusion can be obtained.

Marking 0 , 1 , α , α 2 , , α 2 n 1 above each column of the 2 n -dimensional kernel matrix, and the construction analysis of the ( 2 n 1 ) -order matrix is as follows:

1) For the first layer sub-matrix, l ( 1 ) = 1 and the field is G F ( 2 ) , therefore the first sub-matrix is zero matrix.

2) For the second layer sub-matrix, l ( 2 ) = n can be obtained from Theorem 2, therefore, the second sub-matrix is a unit matrix, and row vectors correspond to 1 , α , α 2 , , α n 1 .

3) For the i-th layer sub-matrix, 3 i c . Because the i-th sub-matrix is composed of primitive BCH code, it is connected by the generator polynomial g ( x ) , and g ( α ) = 0 , the column indicators corresponding to non-zero elements add up to 0.

4) For the 2 n -dimensional kernel matrix. Because the 2 n -dimensional kernel matrix is a lower triangular matrix and the first column’s indicator is marked as 0, the column indicators corresponding to the diagonal elements are the sum of the non-zero element before the diagonal elements

Therefore, the diagonal elements of 2 n -dimensional kernel matrix can be expressed linearly by 1 , α , α 2 , , α n 1 , The sub-diagonal elements of ( 2 n 1 ) -order matrix are equivalent to G F ( 2 n 1 ) , and the diagonal elements of 2 n -dimensional matrix are equivalent to G F ( 2 n ) .

For example, the 7-th row vector in Equation (14), Since the 16-dimensional kernel matrix is a lower triangular matrix, only the elements before the diagonal are listed, it is 0 1 α α 2 α 3 α 4 α 5 ( 1 0 1 1 0 0 1 ) , g ( α ) = α 4 + α + 1 = 0 , and α 4 = α + 1 , therefore, α 5 = α 2 + α or α + α 2 + α 5 = 0 can be obtained.

4. The Partial Distance and Exponent

4.1. The Upper of Partial Distance

Equation (10) can infer

D i = d H ( g i , g i + 1 , , g l ) = min j i D j = d min ( g i , , g l ) d max [ l , l i + 1 ] . (15)

where d min ( a , b ) denotes the minimum distance for generating codeword from vector a and b, d max [ l , k ] , 1 k l denotes the maximum minimum Hamming distance attained by a code of length l and size k [15].

Therefore, the partial distance of kernel matrix cannot exceed its upper bound. Because [5] constructed a ( 2 n 1 ) -dimensional kernel matrix from bottom to top according to the definition of partial distance, the ( 2 n 1 ) -dimensional kernel matrix in [5] met the upper bound of partial distance. [6] also used a similar method to construct a matrix, but [6] obtained a 2 n × ( 2 n 1 ) matrix, not a ( 2 n 1 ) -dimensional kernel matrix. Through adding a column of parity bits, [6] obtained a 2 n -dimensional kernel matrix, but the added column vector would affect the partial distance, and which may cause the partial distance exceeded its upper bound. When l = 32 , D 25 = 14 , for example, D 25 exceeds its upper bound 13 [15].

We take the opposite approach, through constructing ( 2 n 1 ) -order matrix from top to bottom, and then adds a column vector and a row vector of all 1 s before the first column and below the last row of the ( 2 n 1 ) -order matrix, and the 2 n -dimensional kernel matrix can be obtained. Therefore, we only need to analyze the partial distance of ( 2 n 1 ) -order matrix. The Definition of partial distance is proposed for the kernel matrix, but it can be applied to other square matrices, In order to analyze the partial distance of the 2 n -dimensional kernel matrix, We use the definition of partial distance to analyze the ( 2 n 1 ) -order matrix. because according to the Definition 3, the first column of 2 n -dimensional matrix is all 1, and it will not affect the partial distance. The 2 n -dimensional kernel matrix can meet its partial distance by appropriately adjusting the ( 2 n 1 ) -order matrix, the specific adjustment process is as follows:

1) If D i exceeds its upper bound, adjusting the codeword in the i-th row to the ( i 1 ) -th row by shifting one bit to the right.

2) If D i still exceeds its upper bound after a right cyclic shift., both thei-th row and the ( i 1 ) -th row are adjusted to the right cyclic shift of the ( i 2 ) -th row. this cycle continues until D i meets its upper bound.

This method can not only ensure that the ( 2 n 1 ) -order matrix is composed of the primitive BCH codes, but also can ensure that it is still a sub-lower triangular matrix, thereby ensuring the polarization of the 2 n -dimensional kernel matrix. The process is demonstrated in Figure 2.

Figure 2. Adjust the row exceeds the upper bound of the partial distance.

Through the above adjustment method, the 32-dimensional kernel matrix can be obtained as follows:

(16)

4.2. The Lower Bound of Exponent

Definition 6 (Direct-Sum). Let C 1 be a [ n , k 1 , d 1 ] q linear code, C 2 be a [ n , k 2 , d 2 ] q linear code, and C 1 C 2 = { 0 } , if C 1 C 2 = { x + y | x C 1 , y C 2 } , then C 1 C 2 is called the direct-sum of C 1 and C 2 .

The construction of the ( 2 n 1 ) -order matrix is direct-sum construction. Therefore, the relevant properties of the direct-sum construction can be used to analyze the partial distance of the 2 n -dimensional kernel matrix. The minimum distance of direct-sum has the following relationship:

Lemma 2 (Minimum distance [16] ). The minimum distance of C 1 C 2 is d. where

d min { d 1 , d 2 } .

Theorem 2. If the i-th row vector belongs to the j-th layer, then D i 2 j 3 .

Proof. From Lemma 2 and Definition 3, D i is only related to the layer where it belongs with and has nothing to do with the subsequent layers. Because thej-th layer is capable of correcting j 2 errors, D i 2 ( j 2 ) + 1 = 2 j 3 can be obtained.£

Remark. The primitive BCH codes with different error correction capabilities may have the same generator polynomial. For example, when constructing a 31-order matrix, the generator polynomials of t = 5 and t = 6 are same, so the 6-th and 7-th layers matrix are intercepted from different rows of the same generator matrix. However, in order to construct a kernel matrix with larger exponent, the sub-matrices of each layer should be intercepted from different generator matrices, because as the number of primitive BCH code error corrections increases, the minimum distance of these sub-matrices increases, the partial distance of the 2 n -dimensional kernel matrix will increases, therefore the exponent will also increases.

The partial distance lower bound sequence of 2 n -dimensional kernel matrix can be obtained from Theorem 2, the partial distance lower bound sequence is

{ D i } i = 1 , 2 , , 2 n = { 1 , 2 , , 2 n , 3 , , 3 l ( 3 ) , 5 , , 5 l ( 4 ) , , 2 c 3 , , 2 c 3 l ( c ) , 2 n } . (17)

Therefore, E ( G ) = 1 2 n i = 1 2 n log 2 n ( D i ) 1 2 n [ 0 + 1 + l ( 3 ) log 2 n 3 + l ( 4 ) log 2 n 5 + + l ( c ) log 2 n 2 c 3 + 1 ] = 1 2 n [ 2 + 1 n i = 3 c l ( i ) log 2 ( 2 i 3 ) ] .

Namely

E ( G ) 1 2 n [ 2 + 1 n i = 3 c l ( i ) log 2 ( 2 i 3 ) ] . (18)

5. Comparative Analysis

Table 1 compares the scaling exponent of we proposed and [6], and the scaling index can be obtained according to Equation (9). Table 1 shows that the scaling exponent of the 32-dimensional kernel matrix is higher than that of [6], this is because we adjust these rows exceeding the upper bound of partial distance which has impact on the scaling exponent. The scaling exponent of 16-dimensional kernel matrix is slightly lower than [6]. Therefore, the 16-dimensional kernel matrix constructed by this paper is slightly better than [6] in terms of polarization speed.

Table 1. The sacling exponent of kernel matrix

Table 2. The lower bound of kernel matrix’s exponent.

Table 2 compares the lower bounds of the exponent of the proposed and G-V construction. G-V construction takes advantage of D ˜ i = max { D : j = 0 D 1 ( j l ) < 2 i } to obtain the lower bound sequence of the partial distance, and the lower bound of the exponent can be obtained by definition 4. Due to the difficulty of calculation, Table 2 only gives the lower bounds of the exponent of 8, 16, 32, and 64-dimensional kernel matrices. Table 2 shows that the lower bound of the exponent of the kernel matrix is higher than that of the G-V construction in these examples.

6. Conclusions

We use the primitive BCH codes to construct a 2 n -dimensional kernel matrix. Firstly, the generator matrix of primitive BCH codes with different error correction capabilities is intercepted to construct a ( 2 n 1 ) -order matrix, these sub-matrices are stacked from top to bottom, and then adding a column and row vector of all 1 s to form a 2 n -dimensional kernel matrix. Aiming at the problem of partial distance, a solution is proposed to solve the problem of partial distance exceeding the upper bound, through right cyclic shifting, it becomes the right cyclic shift vector of the previous row, which makes the sub-matrix still a lower triangular cyclic structure, ensuring the polarization of 2 n -dimensional kernel matrix, and the lower bound of 2 n -dimensional kernel matrix’s exponent is obtained.

The comparison result shows that the lower bound of 2 n -dimensional kernel matrix constructed in this paper is tighter than G-V construction, and the scaling exponent is better than [6] in 16-dimensional kernel matrix.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Arikan, E. (2009) Channel Polarization: A Method for Constructing Capacity-Achieving Codes for Symmetric Binary-Input Memoryless Channels. IEEE Transactions on Information Theory, 55, 3051-3073.
https://doi.org/10.1109/TIT.2009.2021379
[2] Şaşoğlu, E. (2011) Polar Coding Theorems for Discrete Systems. école Polytechnique Fédérale de Lausanne, Lausanne.
[3] Şaşoğlu, E., Telatar, E. and Arikan, E. (2009) Polarization for Arbitrary Discrete Memoryless Channels. Proceedings of 2009 IEEE Information Theory Workshop, Taormina, 11-16 October 2009, 144-148.
https://doi.org/10.1109/ITW.2009.5351487
[4] Mori, R. and Tanaka, T. (2010) Channel Polarization on q-ary Discrete Memoryless Channels by Arbitrary Kernels. Proceedings of 2010 IEEE International Symposium on Information Theory, Austin, 13-18 June 2010, 894-898.
https://doi.org/10.1109/ISIT.2010.5513568
[5] Korada, S.B., Şaşoğlu, E. and Urbanke, R. (2010) Polar Codes: Characterization of Exponent, Bounds, and Constructions. IEEE Transactions on Information Theory, 56, 6253-6264.
https://doi.org/10.1109/TIT.2010.2080990
[6] Moskovskaya, E. and Trifonov, P. (2020) Design of BCH Polarization Kernels with Reduced Processing Complexity. IEEE Communications Letters, 24, 1383-1386.
https://doi.org/10.1109/LCOMM.2020.2984382
[7] Miloslavskaya, V. and Trifonov, P. (2012) Design of Binary Polar Codes with Arbitrary Kernel. Proceedings of 2012 IEEE Information Theory Workshop, Lausanne, 3-7 September 2012, 119-123.
https://doi.org/10.1109/ITW.2012.6404639
[8] Shen, L.F. and Ye, Z.H. (2004) Information Theory and Coding. Science Press, Beijing.
[9] Fazeli, A., Hassani, H., Mondelli, M. and Vardy, A. (2020) Binary Linear Codes with Optimal Scaling: Polar Codes with Large Kernels. IEEE Transactions on Information Theory, 67, 5693-5710.
https://doi.org/10.1109/TIT.2020.3038806
[10] Hassani, S.H., Alishahi, K. and Urbanke, R.L. (2014) Finite-Length Scaling for Polar Codes. IEEE Transactions on Information Theory, 60, 5875-5898.
https://doi.org/10.1109/TIT.2014.2341919
[11] Fazeli, A. and Vardy, A. (2014) On the Scaling Exponent of Binary Polarization Kernels. Proceedings of 2014 52nd Annual Allerton Conference on Communication, Control, and Computing, Monticello, 30 September-3 October 2014, 797-804.
https://doi.org/10.1109/ALLERTON.2014.7028536
[12] Yao, H., Fazeli, A. and Vardy, A. (2019) Explicit Polar Codes with Small Scaling Exponent. Proceedings of 2019 IEEE International Symposium on Information Theory (ISIT), Paris, 7-12 July 2019, 1757-1761.
https://doi.org/10.1109/ISIT.2019.8849741
[13] Huffman, W.C. (2003) Fundamentals of Error correcting Codes. Cambridge University Press, New York.
https://doi.org/10.1017/CBO9780511807077
[14] Feng, K.Q. (2005) The Algebraic Theory of Error Correcting Codes. Tsinghua University Press, Beijing.
[15] Lin, H.P., Lin, S. and Abdel-Ghaffar, K.A.S. (2015) Linear and Nonlinear Binary Kernels of Polar Codes of Small Dimensions with Maximum Exponents. IEEE Transactions on Information Theory, 61, 5253-5270.
https://doi.org/10.1109/TIT.2015.2469298
[16] Xu, Y.C. and Ma, S.Y. (2013) Algebraic Coding and Cryptography. Higher Education Press, Beijing.

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.