Convolutional codes with a maximum distance profile based on skew polynomials
Abstract
We construct a family of convolutional codes with degree that have a maximum distance profile. The field size required for our construction is , which improves upon the known constructions of convolutional codes with a maximum distance profile. Our construction is based on the theory of skew polynomials.
I Introduction
Construction of codes with good distance has been central to the study of error-correcting codes. While there are quite a number of algebraic constructions of block codes with good distance properties, very few algebraic constructions are known for convolutional codes with good distance properties. In fact, there are several important distance measures for convolutional codes, suitable for assessing their error correction capability in various scenarios. In particular, free distance and column distance [1] are two of the most notable distance measures for convolutional codes.
Free distance is arguably the most well-known notion of distance measure for convolutional codes. An upper bound for the free distance of convolutional codes was presented in [2], which generalizes the Singleton bound for block codes, and thus convolutional codes with free distance attaining this upper bound are called maximum distance separable (MDS) convolutional codes. The existence of MDS convolutional codes is shown in [2] using methods from algebraic geometry. Relying on the intrinsic connection between quasi-cyclic block codes and convolutional codes [3, 4], the authors of [5] presented the first concrete construction of MDS convolutional code under certain restriction of the code parameters. A few other constructions of MDS convolutional codes for different regimes of the code parameters appeared later in [6], [7], [8].
Although both notions of free distance and column distance were born at the same time [1], the notion of column distance appears to be less known. Informally speaking, column distance measures the Hamming distance of a convolutional code truncated at certain time instant. Thus, a convolutional code is associated with a sequence of column distances at different time instants, and such a sequence is termed the distance profile of the convolutional code. In view of this, the distance profile of a convolutional code is particularly helpful for evaluating the error correction capability of the code when it is used for streaming over erasure channels [9], [10]. A comprehensive study of convolutional codes with the largest possible column distances was presented in [11]. Since a truncated convolutional code is a block code, the column distances of the code satisfy the corresponding Singleton bound for block codes. Moreover, the column distances can not exceed the free distance of the code since the latter is derived with no restriction on the time instant. Clearly, not every truncated code of a convolutional code can be an MDS block code. In other words, not every column distance of a convolutional code can attain the corresponding block code Singleton bound. Convolutional codes whose column distances attain the block code Singleton bound for the maximum number of times are said to have a maximum distance profile (MDP), and such codes are often referred to as MDP convolutional codes.
Algebraic constructions of MDP convolutional codes are scarce in the literature. In [11], the authors presented the first explicit construction of MDP convolutional codes over a large finite field of large characteristic. The idea that is instrumental in this construction is to first construct lower triangular Toeplitz supperregular matrices111Roughly speaking, a lower triangular Toeplitz matrix is called supperregular if its minors that are not trivially zero are nonzero. We refer the readers to [11] for a detailed discussion. and then utilize them as the building blocks for the construction of MDP convolutional codes. The same idea was also explored in [12] to construct another class of MDP convolutional codes over a large finite field of arbitrary characteristic. A third general class of explicit MDP convolutional codes was recently presented in [13], which uses generalized Reed-Solomon (RS) codes as the building blocks. These three families of codes are the only known explicit algebraic constructions of MDP convolutional codes hitherto but all of them require a finite field of gigantic size for the codes to be constructed. Specifically, to construct an MDP convolutional code with rate and degree222The degree of a convolutional code determines the implementation complexity of the code and is arguably one of the most fundamental parameters of the code. See Section II-A for a formal definition. , a finite field of size suffices for the construction in [11]. As for the construction in [12], the needed size of the finite field is . For the most recent construction in [13], it requires , which is comparable to that of [11]. Constructing explicit MDP convolutional codes over small finite fields has been challenging and it was conjectured in [14] that a finite field of size should suffice for constructing of MDP convolutional codes with rate and degree .
In this paper we present an explicit construction of MDP convolutional codes over a finite field of size , which improves over all the known explicit algebraic constructions. More precisely, we show that for any rate such that , there exists an explicit MDP convolutional code with rate and degree over a finite field of size . The idea behind our construction differs from the existing constructions and is based on the theory of skew polynomials [15], [16], which has seen important applications in other areas of coding theory such as distributed storage [17], [18].
II Preliminaries
II-A Convolutional codes
Let us first introduce basic concepts and requisite terminology of convolutional codes. We shall treat convolutional codes as submodules instead of subspaces since there is only marginal difference between this treatment and the classical approach [19] of taking convolutional codes as vector subspaces. We refer the readers to [16] for a recent survey of convolutional codes based on the module theoretic approach.
Let be a finite field and be the ring of polynomials with indeterminate and coefficients in .
Definition 1.
An convolutional code over is an -submodule of rank . A polynomial matrix over is called a generator matrix of the code if
An polynomial matrix over is called a parity check matrix of the code if
A polynomial matrix is called basic if it has a polynomial right inverse. The constraint length for the th input or the th row degree of the matrix is defined to be
Furthermore, the memory of the matrix is defined as
and the overall constraint length of the matrix is defined to be
The matrix is said to have generic row degrees if for .
A generator matrix of an convolutional code is called minimal or reduced if its overall constraint length is minimal over all generator matrices of the code . This minimum overall constraint length is called the degree of the code .333One can equivalently define the degree of as the highest degree of the minors of any generator matrix of . This is because the overall constraint length of a generator matrix is always at least the highest degree the of the minors of the generator matrix [20], and the sets of degrees of minors of any two generator matrices are the same since any two generator matrices of differs by left multiplication of a unimodular polynomial matrix. We shall call an convolutional code with degree an convolutional code.
The degree of convolutional codes stipulates the smallest number of memory elements needed to realize an convolutional code and is closely related to the decoding complexity of the code [19]. In this regard, given code parameters and , it is preferable to construct codes with a small degree in practice. However, it is in general not obvious to determine the degree of an convolutional code. The following result from [20] provides a criterion for a generator matrix to be minimal, which is helpful in determining the degree of the convolutional code generated by the matrix.
Lemma 1.
Let be a matrix over and define the matrix of the highest order coefficients for , denoted by , by
where denotes the coefficient of in the polynomial . Then is minimal if and only if has rank .
Next, let us briefly discuss distance properties of convolutional codes. To begin with, note that for , we may write . The Hamming weight of is then defined as
where denotes the Hamming weight of the length- vector over . Similarly to the distance of a linear block code, the free distance of a convolutional code is defined to be
The free distance was shown in [2] to satisfy the following upper bound that generalizes the Singleton bound for block codes.
Theorem 2.
Let be an convolutional code. Then
(1) |
Convolutional codes whose distance attains (1) with equality are called MDS convolutional codes.
In addition to the free distance, the column distance is another fundamental distance measure for convolutional codes. To define the column distance, let be a generator matrix of memory for an convolutional code . Further, let the corresponding semi-infinite matrix be
Let . Denote the th truncated generator matrix by
where for . Similarly, let be a parity check matrix of memory for the code . The th truncated parity check matrix is denoted by
where for .
With the th truncated matrix, the th column distance of a convolutional code can be defined as follows.
Definition 2.
Let be a generator matrix of memory for an convolutional code such that has full rank. For the th column distance of the code is given by
The Singleton bound for block codes implies that for all the column distance satisfies
(2) |
and equality for a given implies that all the other distances also attain their versions of the bound (2) with equality. The following result, given in [11], characterizes optimal column distances by the determinants of full-size square submatrices of .
Theorem 3.
Let be a basic and minimal generator matrix of an convolutional code and let be a basic parity check matrix of the code. Then the following are equivalent:
-
1.
;
-
2.
every full-size minor of formed by the columns with indices where for is nonzero;
-
3.
every full-size minor of formed by the columns with indices where for is nonzero.
For brevity, we shall call Item 2 of Theorem 3 the MDP property of , and Item 3 of Theorem 3 the MDP property of .
While for a matrix to generate an MDS block code, every submatrix should have a nonzero determinant, the MDP property of implies that for an convolutional code to have optimal th column distance, the full-size square submatrices that should have nonzero determinants are only those submatrices formed by at most columns of the first columns of for . In fact, the determinant of any other full-size submatrix of is always zero.
As there are various distance measures for convolutional codes, their error correction capability can be optimal in a number of different ways. We mention here several families of distance-optimal convolutional codes. Convolutional codes with a generator matrix of memory and distance attaining the equality in (2) for are called -MDS convolutional codes. Relating the bounds (1) and (2), [21] and [11] introduced two other types of distance-optimal convolutional codes, described below.
Definition 3.
Let be an convolutional code. Let and .
-
1.
The code is said to be strongly-MDS if
-
2.
The code is said to be MDP if
Clearly, strongly-MDS convolutional codes are also MDS by definition. Moreover, the column distance of strongly-MDS convolutional codes attains the Singleton bound (1) at the earliest possible time instant . But the column distance at the time instant prior to may not attain (2) . On the contrary, the column distances of MDP convolutional codes attain (2) for the maximum number of times although the free distance of the codes may not attain (1). We refer the readers to [11] for a detailed discussion of the difference between these two families of codes.
Theorem 3 suggests that one way of constructing MDP convolutional codes is to design a minimal and basic generator matrix such that the corresponding truncated matrix has the MDP property. Note that Lemma 1 provides a characterization of minimal matrices. As for basic matrices, it is observed in [22] that the assumption of being basic in Theorem 3 can be lifted and replaced with a weaker condition that the matrix has generic row degrees. This observation is summarized as follows.
Lemma 4.
Let be a minimal generator matrix with row degree and let . If has the MDP property then is basic and it generates an MDP convolutional code.
Finally, we note a duality result of MDP convolutional codes from [11].
Theorem 5.
An convolutional code over with generator matrix is MDP if and only if its dual code with parity check matrix is an MDP convolutional code over .
II-B Skew polynomials
Skew polynomials satisfy most basic properties of conventional polynomials but the product of skew polynomials is not commutative. We refer the readers to [23] for basic properties of skew polynomials over finite fields and [24, 15] for the theory of evaluation and interpolation on skew polynomials. The recent papers [25, 26] also have an accessible exposition of the theory of skew polynomials. In the following we shall only briefly introduce basic results of skew polynomials that suit the need of this paper.
Let be a finite field of size where is a prime power. Let be the Frobenius automorphism. Namely, for any .
Definition 4.
Let be the ring of skew polynomials with indeterminate in and coefficients in , where addition in is coefficient wise and multiplication in is distributive and satisfies that for any
As the product of skew polynomials is not commutative, evaluating skew polynomials over elements in a finite field is different from evaluating conventional polynomials. The following result, given in [24, 15], provides a simple approach to evaluate skew polynomials.
Definition 5.
For any , let
-
1.
;
-
2.
for .
Lemma 6.
Let . Then the evaluation of at any is given by
Different from conventional polynomials, a skew polynomial may have more than roots in . Moreover, the roots may belong to distinct conjugacy classes in induced by the field automorphism . The notion of conjugacy classes, introduced in [24, 15], is given below.
Definition 6.
Let . Then is a conjugate of with respect to the field automorphism if there exists such that
We also call the -conjugate of with respect to and write for brevity.
The notion of conjugacy above defines an equivalence relation in , and thus we can partition into distinct conjugacy classes. For any , let us denote by the conjugacy class with representative .
Proposition 7.
Let be a primitive element of . Then is a partition of . Moreover, and for .
The set of roots in of a skew polynomial has interesting structures closely related to the conjugacy classes of . The result below can be found in [26], [24, 15].
Theorem 8.
Let be a nonzero skew polynomial and be the set of roots of in . Further, let be the partition of into conjugacy classes and let where is some fixed representative in . Then is a vector space over where if and otherwise . Moreover,
Definition 7.
Let be a positive integer and let . The skew Vandermonde matrix with respect to , denoted by , is defined to be
The following corollary is a consequence of Theorem 8.
Corollary 9.
Let be an -subset of and be the partition of into conjugacy classes. Let and fix . Suppose further that . Then has full rank if and only if for each the set is linearly independent over where if and otherwise .
Note that evaluation of skew polynomials does not preserve linearity. Nevertheless, the following result provides a way to linearize the evaluation of skew polynomials on any conjugacy classes, which was first shown in [24].
Lemma 10.
Let , and . Then is an -linear map. In other words, for any and , we have .
Corollary 11.
Let be a nonzero skew polynomial and let . Further, let . If be a primitive element of then
III Construction
In this section we will construct MDP convolutional codes with degree . In the following, we will first present a construction of unit memory convolutional codes whose generator matrix is constructed based on skew Vandermonde matrices.
Construction III.1.
Let and let be a prime power. Let be a finite field and be a primitive element of .444By the assumption that , it follows from Proposition 7 that there exists at least three conjugacy classes in . Further, let be distinct elements and define using an arbitrary basis for over as
(3) | ||||
(4) |
Let be an convolutional code over with generator matrix where and are given by
(5) | |||
(6) |
Before proving the code is a unit memory MDP convolutional code, let us first show that its generator matrix has certain desirable properties.
Proposition 12.
Any submatrix of or defined in Construction III.1 has full rank.
Proof.
The construction (3) of ensures that any elements of are linearly independent over . By Corollary 9, any submatrix of the skew Vandermonde matrix has full rank. Observe that can be obtained by multiplying the th column of the nonzero element for all . It follows that any submatrix of has full rank. Similarly, one can show that, by the construction (4) of , any submatrix of has full rank. ∎
Proposition 13.
The generator matrix defined in Construction III.1 is minimal.
Proof.
Now we are ready to prove that is an MDP convolutional code whenever the rate .
Theorem 14.
Let . Then the code defined in Construction III.1 is an unit memory MDP convolutional code over .
Proof.
It is clear that the memory of the generator matrix is and we have
By Proposition 13, is minimal, and thus the degree of is . Since , we have . Note that has generic row degrees. Therefore, by Lemma 4, it suffices to show that has the MDP property. Namely, if satisfies Item 2 of Theorem 3, then is an MDP convolutional code with degree .
Let be such that and . Let be the matrix formed by columns of with indices in the set where means adding every element of by the integer . Further, let and . We would like to show that if and only if . Let be skew polynomials. Then, by Lemma 6 and Lemma 10, it is equivalent to showing that
(7) | |||
(8) |
if and only if . For clarity, let us write the matrix more explicitly as
where is the submatrix of with column indices in , and and are defined similarly.
Consider the case . In this case, we have . By Proposition 12, are matrices of full rank. It follows that if and only if .
Consider the case . In this case, we have . Suppose there exists such that . Suppose further that but . By Proposition 12, the matrix has full rank. Therefore, implies , which contradicts the assumption that . Now suppose that but . Similarly, by Proposition 12, the matrix has full rank. Therefore, implies , which contradicts the assumption that , and we are left with the only possibility that for to be nonzero.
The conclusion of this theorem for the case will follow from the following claim.
Claim 15.
If , , and then
(9) | ||||
(10) |
Before proving the claim, let us show that this claim indeed implies the theorem. Suppose there exists such that . Then Claim 15 holds. By (10) there exists with such that is linearly independent over . Moreover, since then by (8) we have for all . It follows that is linearly independent over , which contradicts (9) since . Therefore, if then , and the theorem follows. It remains to establish Claim 15.
Proof of Claim 15: Let us first establish (9). Note that we have . By construction (3) of , any elements of form a maximally linearly independent subset of . In other words, there exists with such that for any the element can be written as an -linear combination of . By Lemma 10, is an -linear map. Therefore, for any the element can be written as an linear combination of . Thus, .
Next, let us show (10). Note that is a nonzero skew polynomial with . Therefore, by Corollary 11, . At the same time, by (7) we have for all where . In addition, by construction (3) of , the elements in are linearly independent over . Therefore, we have . It follows that
(11) |
Suppose . Then there exists with such that for all
where . By Lemma 10, is an -linear map. Thus, for all we have
Let . Then . We claim that . Indeed, if , then can be written as a linear combination of over where . However, by construction (4) of , any elements of are linearly independent over , which is a contradiction. Furthermore, by construction (4) of , the set is linearly independent over . Since for all , we have . Therefore, from (11) we obtain , which implies and leads to a contradiction. This completes the proof of Claim 15 as well as the proof for the theorem. ∎
The above result gives an explicit construction of MDP convolutional code with rate and degree . It can be easily extended to the case of codes with higher rate and the same degree by applying Theorem 5.
Corollary 16.
Let and be the dual code of the code defined in Construction III.1. Then is an MDP convolutional code over .
Corollary 17.
Let . There exists a family of MDP convolutional codes that can be constructed explicitly over a finite field of size .
Furthermore, the duality of MDP convolutional codes implies that Construction III.1 also gives rise to a family of MDP convolutional codes with degree .
Corollary 18.
Let . There exists a family of MDP convolutional codes that can be constructed explicitly over a finite field of size .
IV Concluding remarks
A few observations can be made from Corollary 17 and 18. As mentioned before, it is desirable for convolutional codes to have small degree. Corollary 17 and 18 imply that, given any parameter and such that , there exists a family of MDP convolutional codes with degree that can be constructed explicitly over a finite field of size . In particular, if is a constant such that , then the codes can be constructed over a finite field of size . A comparison of the field size requirement for constructions of MDP convolution codes with constant rate and memory is given in Table I.
Field size | Reference |
---|---|
[11] | |
[12] | |
[13] | |
This paper | |
Conjecture in [14] |
It is worth noting that by Definition 3, when we have , and thus in that case the constructions given by Corollary 17 and 18 are also strongly-MDS convolutional codes.
Finally, it is interesting to generalize the approach in this paper to construct MDP convolutional codes for any and extend the result to any value of .
Acknowledgements
The author is grateful to Alexander Barg for discussion and comments on the first version of this paper.
References
- [1] D. J. Costello, “A construction technique for random-error-correcting convolutional codes,” IEEE Transactions on Information Theory, vol. 15, no. 5, pp. 631–636, 1969.
- [2] J. Rosenthal and R. Smarandache, “Maximum distance separable convolutional codes,” Applicable Algebra in Engineering, Communication and Computing, vol. 10, no. 1, pp. 15–32, 1999.
- [3] J. Justesen, “New convolutional code constructions and a class of asymptotically good time-varying codes,” IEEE Transactions on Information Theory, vol. 19, no. 2, pp. 220–225, 1973.
- [4] J. L. Massey, D. J. Costello, and J. Justesen, “Polynomial weights and code constructions,” IEEE Transactions on Information Theory, vol. 19, no. 1, pp. 101–110, 1973.
- [5] R. Smarandache, H. Gluesing-Luerssen, and J. Rosenthal, “Constructions of MDS-convolutional codes,” IEEE Transactions on Information Theory, vol. 47, no. 5, pp. 2045–2049, 2001.
- [6] H. Gluesing-Luerssen and B. Langfeld, “A class of one-dimensional MDS convolutional codes,” Journal of Algebra and its Applications, vol. 5, no. 04, pp. 505–520, 2006.
- [7] F. J. Plaza-Martín, J. I. Iglesias-Curto, and G. Serrano-Sotelo, “On the construction of 1-D MDS convolutional Goppa codes,” IEEE transactions on information theory, vol. 59, no. 7, pp. 4615–4625, 2013.
- [8] L. Julia and R. Pinto, “Constructions of MDS convolutional codes using superregular matrices,” Journal of Algebra Combinatorics Discrete Structures and Applications, vol. 7, no. 1, pp. 73–84, 2020.
- [9] V. Tomás, J. Rosenthal, and R. Smarandache, “Decoding of convolutional codes over the erasure channel,” IEEE Transactions on Information Theory, vol. 58, no. 1, pp. 90–108, 2012.
- [10] R. Mahmood, A. Badr, and A. Khisti, “Convolutional codes with maximum column sum rank for network streaming,” IEEE Transactions on Information Theory, vol. 62, no. 6, pp. 3039–3052, 2016.
- [11] H. Gluesing-Luerssen, J. Rosenthal, and R. Smarandache, “Strongly-MDS convolutional codes,” IEEE Transactions on Information Theory, vol. 52, no. 2, pp. 584–598, 2006.
- [12] P. Almeida, D. Napp, and R. Pinto, “A new class of superregular matrices and MDP convolutional codes,” Linear Algebra and its Applications, vol. 439, no. 7, pp. 2145–2157, 2013.
- [13] G. N. Alfarano, D. Napp, A. Neri, and V. Requena, “Weighted Reed-Solomon convolutional codes,” arXiv preprint arXiv:2012.11417, 2020.
- [14] R. Hutchinson, R. Smarandache, and J. Trumpf, “On superregular matrices and MDP convolutional codes,” Linear Algebra and its Applications, vol. 428, no. 11-12, pp. 2585–2596, 2008.
- [15] T.-Y. Lam and A. Leroy, “Vandermonde and Wronskian matrices over division rings,” Journal of Algebra, vol. 119, no. 2, pp. 308–336, 1988.
- [16] J. Lieb, R. Pinto, and J. Rosenthal, “Convolutional codes,” arXiv preprint arXiv:2001.08281, 2020.
- [17] U. Martínez-Peñas and F. R. Kschischang, “Universal and dynamic locally repairable codes with maximal recoverability via sum-rank codes,” IEEE Transactions on Information Theory, vol. 65, no. 12, pp. 7790–7805, 2019.
- [18] H. Cai, Y. Miao, M. Schwartz, and X. Tang, “A construction of maximally recoverable codes with order-optimal field size,” IEEE Transactions on Information Theory, 2021.
- [19] R. Johannesson and K. S. Zigangirov, Fundamentals of convolutional coding. John Wiley & Sons, 2015.
- [20] R. J. McEliece and R. P. Stanley, “The general theory of convolutional codes,” The Telecommunications and Data Acquisition Report, 1993.
- [21] R. Hutchinson, J. Rosenthal, and R. Smarandache, “Convolutional codes with maximum distance profile,” Systems & Control Letters, vol. 54, no. 1, pp. 53–63, 2005.
- [22] G. N. Alfarano and J. Lieb, “On the left primeness of some polynomial matrices with applications to convolutional codes,” Journal of Algebra and Its Applications, p. 2150207, 2020.
- [23] O. Ore, “Theory of non-commutative polynomials,” Annals of mathematics, pp. 480–508, 1933.
- [24] T.-Y. Lam, A general theory of Vandermonde matrices. Center for Pure and Applied Mathematics, University of California, Berkeley, 1985.
- [25] U. Martínez-Peñas, “Skew and linearized Reed–Solomon codes and maximum sum rank distance codes over any division ring,” Journal of Algebra, vol. 504, pp. 587–612, 2018.
- [26] S. Gopi and V. Guruswami, “Improved maximally recoverable LRCs using skew polynomials,” arXiv preprint arXiv:2012.07804, 2020.