 Research
 Open Access
 Published:
Research on correction algorithm of propagation error in wireless sensor network coding
EURASIP Journal on Wireless Communications and Networking volume 2020, Article number: 116 (2020)
Abstract
It is very difficult to deal with the problem of error correction in random network coding, especially when the number of errors is more than the mincut of the network. We combine a small field with rankmetric codes to solve this problem in this paper. With a small finite field, original errors are compressed to propagated errors, and their number is smaller than the mincut. Rankmetric codes are introduced to correct the propagated errors, while the minimum rank distance of the rankmetric code is hardly influenced by the small field. It is the first time to correct errors more than the mincut in network coding with our method using a small field. This new errorcorrecting algorithm is very useful for the environment such as a wireless sensor network where network coding can be applied.
Introduction
The nature of combining information in intermediate nodes makes network coding very susceptible to transmission errors. How to control and correct the errors in random network coding is therefore of great interest to researchers. This problem naturally motivates the topic of network error correction (NEC), pioneered by Yeung and Cai [1]. Since then, a variety of models were presented to combat the problem. For a determined network, its topology is stable over time, Guang et al. [2,3,4] proposed several NEC construction algorithms for correcting propagated errors in it based on the Hamming metric. For random networks, changing their topologies over time, original errors can be spread to the downstream nodes in the networks and cause the decoding failure when the Hamming metric is used. To solve the issue, Kotter [5] introduced a subspace/rankmetricbased method and developed an elegant approach. However, the capacity has reached the theoretical upperbound of NEC (C − 2t), where C and t are the mincut of the network and the number of corrupted links in the network, respectively. That is to say, t , the number of errors to be corrected is no more than C/2, here t = (C ‐ R)/2 and R is the transmission rate of the code. Guruswami et al. [6] generalized the rankmetric code to its listdecoding version. The method can correct nearly C errors when the transmission rate R is close to zero. In order to correct more errors, Guo et al. [7] introduced a nonlinear operation to network coding and proved that the transmission rate can be bigger than (C − t) in her method. But she did not show her concrete construction and corresponding decoding algorithm about the method in the paper we also noticed. The operations in the method have exponential complexity. So far, there are few practical methods that can correct more than C errors in the related literature. However, in the real communication environments, the number of original errors is usually larger than mincut C, since the number is directly proportional to total links in the networks.
In random network coding, the transport process can be depicted by Y = T ⋅ G ⋅ u + T_{Z → Y} ⋅ Z, where u is the original message intended to be transmitted. In a source node of a multicast network, u is encoded to G ⋅ u by a code Ω with its generate matrix G. Then, G ⋅ u is sent to the network and encounters a transfer matrix T, which represents the effect of network coding upon G ⋅ u. Simultaneously, the error Z occurred on the links also encounter their own transfer matrix T_{Z → Y}, and then, T_{Z → Y} ⋅ Z are injected into the messages Y received in the sink node. In the existing NEC models, packets must be collected enough to guarantee that T is a fullrank matrix. By multiplying T^{‐1} to both sides of the equation Y = T ⋅ G ⋅ u + T_{Z → Y} ⋅ Z, we can obtain T^{‐1} ⋅ Y = G ⋅ u + T^{‐1} ⋅ T_{Z → Y} ⋅ Z. Based on the code Ω with its generated matrix G, we can further get the original message u. Denoted by Z, the original error, as the effect of network coding, there are several variations of Z such as T^{‐1} ⋅ T_{Z → Y} ⋅ Z, T_{Z → Y} ⋅ Z, and they are called “propagated errors.” In essence, the existing NEC methods focused their aims on compressing the number of “propagated errors.” With respect to the determined network [2,3,4], the Hamming weight of the propagated error T^{‐1} ⋅ T_{Z → Y} ⋅ Z is compressed to the Hamming weight of the original error Z. The spread effect of the original error Z is compressed by the NEC method. But for a random network, it is difficult to compress the spread of original error from the perspective of the Hamming metric. However, from the view of the rank metric, the rank of propagated error T^{‐1} ⋅ T_{Z → Y} ⋅ Z is no more than the rank of the original error Z, and the spread of error in network coding is compressed naturally [5]. Once the listdecoding method is introduced, the range of the number of original errors that can be corrected increases from the interval (0, C/2) to (C/2, C), see [5] or [6] for a reference. In order to guarantee T is a fullrank matrix, the size of the finite field should be big enough [6]. The rank of propagated error T^{‐1} ⋅ T_{Z → Y} ⋅ Z could not be compressed to a smaller one than C when the rank of the original error Z is bigger than C. Even if the list decoding of rankmetric code [6] has a strong decoding ability, it still cannot correct the propagated errors based on the rationale of coding theory when the rank of propagated error T^{‐1} ⋅ T_{Z → Y} ⋅ Z is bigger than C.
In this paper, we propose a new method, which can correct more than C errors. In our experiments, we test and verify that if the size of a finite field for network coding is smaller than a threshold number, the rank of T_{Z → Y} ⋅ Z can be smaller than C. But T unfortunately is usually not full rank, so we cannot use the decoding algorithm of code Ω with the generated matrix G to decode u, where the corresponding decoding equation is T^{‐1} ⋅ Y = G ⋅ u + T^{‐1} ⋅ T_{Z → Y} ⋅ Z. For T is not full rank in a small field, we decode u based on the equationY = T ⋅ G ⋅ u + T_{Z → Y} ⋅ Z. We will use the new code Ω’ with it generates matrix, T ⋅ G, instead of G to decode u. Here, the rank of T_{Z → Y} ⋅ Z is smaller than C. Now, the only question remaining is how much the minimum rank distance of code Ω’ declines, compared to the minimum rank distance of code Ω. If the minimum distance of code Ω’ declines too much, we still cannot correct the propagated error T_{Z → Y} ⋅ Z with the listdecoding algorithm of rankmetric code, just like that in [7]. Fortunately, based on our experiments to be shown in Section 5, we found that the minimum distance of T ⋅ G is very close to the minimum distance of G even if T is not full rank when a small field is used. In the traditional research paradigm of network coding, we use a big field to make sure T is full rank. In our work, we deal with the problem by using a small field way to compress the spread of original error based on the rank metric. Even the number of nonzero components in original error Z is far larger than C, the rank of propagated error T_{Z → Y} ⋅ Z also can be compressed to a smaller number than C as long as the size of the field is small enough. Our method not only can correct more than C errors in random network coding, but also can correct lots of errors valuable to practical applications of network coding.
The rest of this paper is organized as follows. We first introduce the related works including rankmetric codes and list decoding briefly in Section 2. Then, we present our method in Section 3. In Section 4, we give the experiment results and make a detailed theoretical analysis based on a combinatorial theory of probabilities and simulation. Finally, we draw a conclusion about our work in Section 5.
Related works
In this section, as the background of our work, we first introduce the rankmetric codes and the listdecoding technique. Then, we explain a model about NEC, in which list decoding of rankmetric codes is involved.
Rankmetric codes
Denote the set of all n × m matrices over F_{q} by \( {F}_q^{n\times m} \) and suppose a rankmetric code, Ω, is a subset of \( {F}_q^{n\times m} \). Define the distance between \( X\in {F}_q^{n\times m} \) and \( Y\in {F}_q^{n\times m} \) as rank(X − Y). Obviously, we can also take the matrix \( X\in {F}_q^{n\times m} \) as a vector \( x\in {\left({F}_{q^m}\right)}^n \) in the field \( {F}_{q^m} \), and this means there is a bijection between the vector set of \( {F}_{q^m} \) and the matrix set of \( {F}_q^{n\times m} \). We further define rank(x) = rank(X).
List decoding of rankmetric codes
In the traditional decoding method of linear block codes, the solution is unique, and the number of errors that can be corrected is less than the halfsize of the codeword. With the method of list decoding, we can correct errors more than half of a codeword size, but the solutions of listdecoding method are not unique. Based on the rank codes, Guruswami in [6] proposed the list decoding of rank codes and approximate the number of the corrected errors to the word size when the transmission rate approximates zero. We can utilize the method with its strong errorcorrecting ability to correct the propagated error in network coding, where the codeword size is exactly the mincut C.
Graph model about network coding
Consider an acyclic directed graph \( \mathcal{G}=\left\{\mathcal{V},\mathrm{\mathcal{E}}\right\} \), where \( \mathcal{V} \) is a node set and, ℰ is an edge set whose elements represent network channels. A channel e = (i, j) is a directed edge starting from node i and ending at node j, i.e., tail(e) = i and head(e) = j. For a nodei, the collection of incoming channels is In(i) {e : e ∈ ℰ, head(e) = i} and the collection of outgoing channels is out(i) = {e : e ∈ ℰ, tail(e) = i}. Each channel has a unit capacity in the network. NEC is specified by the related encoder at the source, encoders at intermediate nodes, and decoders at sink nodes. The coding and decoding operations for messages are performed over the field \( {F}_{q^m} \), and the network coding is F_{q}.
Proposed methods
Some claims based on numerical experiments
In this section, we give a set of numerical experiment results to be used in our method. The results are summed up in our three claims. The experiment details will be given in the next section. The three claims establish the base of our method in Section 3. Our claims are based on plenty of experiments rather than rigorous mathematical deducing. However, we also give out auxiliary mathematical analysis about the claims in Section 4.
Claim 1:
Suppose ∣F_{q} ∣ ∈ {2, 3, 4, 5}, there is rank(T_{Z → Y} ⋅ Z') < C with a high probability, where Z ' ∈ (F_{q})^{t × m}, T_{Z → Y} ∈ (F_{q})^{C × t}, and t > 0.
Claim 2:
For rankmetric code Ω with the generated matrix G and minimum rank distance d_{min}, if rank(T) = C ‐ k, then d_{min} ' = d_{min} − k is the minimum rank distance d_{min}' of code Ω’ with the generated matrix T ⋅ G, where T ∈ (F_{q})^{C × C}, 0 < k < C.
Claim 3:
If ∣F_{q} ∣ ∈ {2, 3, 4, 5}, we have rank(T) = C − 1, where T ∈ (F_{q})^{C × C} with high probability.
Data collection strategy with delay constraint
In this section, we formally propose our method. The method is specified by its encoder at the source, encoders at intermediate nodes, and decoders at sink nodes. The coding operation is over the field F_{q}.
Source
A source message \( u\in {\left({F}_{q^m}\right)}^R \) is encoded with a rankmetric code Ω equipped with the generated matrix \( G\in {\left({F}_{q^m}\right)}^{C\times R} \), where m ≥ C and 0 < R < C. The coded message is a vector \( x\in {\left({F}_{q^m}\right)}^C=G\cdot u \). Then, x is then sent to the network from the source. The minimum rank distance of Ω is d_{min}.
Coding at intermediate nodes
Every intermediate node combines its received packets from incoming edges with its own local coding kernel in field F_{q}, creates a new packet, and sends the packet to its successors via outgoing edges. At a more macroscopic level, the network coding at that time leads that messages encounter a socalled transfer matrix [8]. In our method, the messages \( x\in {\left({F}_{q^m}\right)}^C=G\cdot u \) encounter a transfer matrix T ∈ (F_{q})^{C × C} and the error \( Z\in {\left({F}_{q^m}\right)}^t \) encounters a corresponding transfer matrix T_{Z → Y} ∈ (F_{q})^{C × t}, where t is the number of edges, on which errors occur. Roughly speaking, the received messages Y in the sink can be expressed as Y = T ⋅ G ⋅ u + T_{Z → Y} ⋅ Z, if T is not polluted by errors. T is gotten from the global coding kernels in the packets. However, if errors occur, the global coding kernels may also be polluted. At this case, T is then naturally polluted and it is unknown. Though T is unknown, the polluted version of it, denoted by \( \hat{T} \), is known. The transmission procedure can be expressed by a more delicate equation \( Y=\hat{T\cdot }G\cdot u+{T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right) \). Here, L ∈ (F_{q})^{t × C} is a matrix, formed by grouping the global coding kernel vectors together.
Decoding in the sink
Let \( \hat{T}\cdot G \) as the generated matrix of the newly formed rankmetric code Ω' with a minimum rank distance of d_{min}' and Y as the received messages, and the errors can be evaluated as T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u). We utilize the listdecoding method of rankmetric codes [6] to perform decoding and get u, as long as rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) < C and d_{min}' is not smaller than d_{min} too much. This means u^{d}, corresponding to \( {Y}^d=\underset{Y^d\in \Omega \hbox{'}}{\arg}\min \operatorname {rank}\left({Y}^dY\right) \), is the estimate of u. So, u^{d} is the solution of the decoding algorithm for the original message. In the list decoding, u^{d} is not unique.
The feasibility of the decoding
We discuss the feasibility of our method here. The feasibility depends on claims 1 to 3. As mentioned in Section 2, the model in [6] to correct the errors is based on the equation \( {\hat{T}}^{1}\cdot Y=G\cdot u+{\hat{T}}^{1}\cdot {T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right) \) in the random network coding. \( {\hat{T}}^{1}\cdot Y=G\cdot u+{\hat{T}}^{1}\cdot {T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right) \) is induced by multiplying \( {\hat{T}}^{1} \) on both sides of the equation \( Y=\hat{T\cdot }G\cdot u+{T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right) \), and \( {\hat{T}}^{1} \) is invertible in the field F_{q}. We use the decoding algorithm of code Ω with its generated matrix G to perform list decoding. The listdecoding produce can correctly work as long as \( \operatorname{rank}\left({\hat{T}}^{1}\cdot {T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right)\right)<C \). It is a dilemma to make a choice, using a bigger field size or using a smaller field size in the context of random network coding. If \( {\hat{T}}^{1} \) is invertible, the size of field F_{q} would be big enough, and it is usually ∣F_{q} ∣ ≥ 256 [9]. On the other hand, claim 1 shows \( \operatorname{rank}\left({\hat{T}}^{1}\cdot {T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right)\right)>C \) if ∣F_{q}∣ is bigger than 5. In this case, the decoding certainly fails if the listdecoding method [6] is used.
As discussed above, we use code Ω' with the generated matrix \( \hat{T\cdot }G \) to perform list decoding, where the corresponding error is T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) rather than \( {\hat{T}}^{\hbox{} 1}\cdot {T}_{Z\to Y}\cdot \left(ZL\cdot G\cdot u\right) \). The two preconditions for successfully decoding are that rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) < C and d_{min}' is not smaller than d_{min} too much, respectively. The first condition, rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) < C, can be naturally satisfied, and it is necessary according to the inherent nature of the linear block code. For the second condition, if d_{min}' is not smaller than d_{min} too much, we can consider that the code Ω' has a nearly identical errorcorrecting ability with the code Ω. When rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) < C, we also perform list decoding of rankmetric just like done in [6]. If d_{min}' is smaller than d_{min} too much, the errorcorrecting ability of code Ω' declines sharply, compared with code Ω. In this case, it becomes unmeaningful in practice, even if code Ω' can correct nearly C errors in theory. After the theoretical analysis, we explain that the two preconditions can be met in the following case if ∣F_{q} ∣ ∈ {2, 3, 4, 5}. In order to obtain better results, we set ∣F_{q} ∣ = 2 in this paper.
Based on claim 1, rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) is smaller than C if ∣F_{q} ∣ = 2. In essence, rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) ≤ rank(T_{Z → Y} ⋅ Z) because rank(Z − L ⋅ G ⋅ u) ≤ rank (Z).
Based on claims 2 and 3, it can be guaranteed that d_{min}' is not smaller than d_{min} too much. In most cases, d_{min} ' = d_{min} or d_{min} ' = d_{min} − 1, and in a few cases, d_{min} ' = d_{min} − 2. The specific situation depends on the sizes of C and m. The details of it will be introduced in the experiment section.
Advantages and disadvantages
The advantages of our method are as follows: (1) More than C errors can be corrected with list decoding of rankmetric codes in a random network coding, when we adopt a small network coding field, for example, ∣F_{q} ∣ = 2. (2) Our method has the ability of correcting the original more than mincut C errors, and it is very important in the real applications of network coding. (3) The small field can avoid a low computational burden. In the rankmetric code for network coding, we can make the extension field \( {F}_{q^m} \) big enough to get a bigger d_{min}. In this case, m is usually bigger than mincut C [10]. In [9], ∣F_{q}∣ is set more than 256 to guarantee \( \hat{T} \) is invertible. Naturally, a bigger \( {F}_{q^m} \) obviously lead to a heavy computation. In our approach with a small field, \( {F}_{q^m} \) should be smaller because ∣F_{q}∣ is very small, and a small \( {F}_{q^m} \) can lead to a serious computational burden. (4) The disadvantage of our approach is that the transmission rate is a little smaller than the rate in [6] because d_{min}' is 1 or 2 and it is smaller than d_{min}. But the problem can be alleviated as C becomes bigger because 1 or 2 is a few ratios of C. In this case, d_{min}' is also sufficient to successfully decode. On the other hand, the more components about the original error Z, i.e., t, the bigger the rank of the propagated errors T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) it takes. In this case, the transmission rate is usually low.
Experimental results and discussion
In this section, we give a set of experiments to support claims 1 to 3 in Section 3. We also give corresponding theory analysis about experiments. Because many mathematical operations are performed in the finite field, we design our program about the finite field mathematical operation based on MATLAB, such as inversing a matrix, computing the inverse of a number in the finite field, and computing rank of a matrix in the finite field. In the experiments, we use newly designed finite operation methods to verify claims 1 to 3.
Experimental results
Claim 1
Assume the mincut C = 5 ∗ ∣ F_{q}∣.
In Fig. 1, the different numbers of original errors are illustrated by different curves. We can find that, if ∣F_{q} ∣ = 2, the rank of the propagated errors T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) can be compressed to 0.7 ∗ C, even if t = 20 ∗ C, where t is the number of original errors Z. In this case, the list decoding of rankmetric codes can correct the propagated errors T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) easily. This character is good for correcting the dense errors in the random network coding.
Because it is difficult to do finite field programming with MATLAB when the size of the finite field is even, we use C programming language to do finite field programming. Like Fig. 1, where the size of the finite field is odd, Fig. 2 shows how the errors are propagated in the finite field when the size of the finite field is even.
We now analyze the experimental results theoretically and find that there is a satisfactory coincidence with the theory and the experiment results. The transfer matrix T_{Z → Y} ∈ (F_{q})^{C × t} is known in advance, where t is the number of edges on which errors occur, and (Z − L ⋅ G ⋅ u) takes its value in \( {\left({F}_{q^m}\right)}^t \). Based on the theory of the extension field and the base field, we can take the vector (Z − L ⋅ G ⋅ u) in field \( {F}_{q^m} \) as a matrix (F_{q})^{t × m} in field F_{q} [10]. T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) means that when we multiply two matrices together in the field F_{q}, they should satisfy (C ⋅ t) × (t ⋅ m) = (C × m). Finally, we can assert T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) ∈ (F_{q})^{C × m}. Obviously, no matter how big the value t is, the rank of T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) is no more than C. The reason is that a small field F_{q} usually makes rank(T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u)) ≤ C, no matter how big the value of t is. If the Hamming metric is adopted, no matter how small the size of F_{q} is, we cannot compress Hamming weight of T_{Z → Y} ⋅ (Z − L ⋅ G ⋅ u) to the number smaller than C. Naturally, we cannot correct the propagated errors even though the list decoding is used based on the Hamming metric.
Claim 2
According to the experiments we have done, we found that claim 2 always holds on all the parameter combinations. So far, no counterexample has been found to claim 2 in our work.
Claim 3
In Table 1, we show several examples about the decline of the rank of T. We can see the declined amount is 0, 1, or 2.
In Fig. 3, we illustrated the ratio of the declined amount of the rank of T in a small field to the full rank. We can find the ratio very low, which means d_{min}' is very close to d_{min} when the size of the finite field is odd.
Because it is difficult to do finite field programming with MATLAB when the size of the finite field is even, we use C programming language to do finite field programming. Like Fig. 3 where the size of the finite field is odd, Fig. 4 shows how the rank reduced in the finite field when the size of the finite field is even (Table 1).
Whether T is a full rank depends on the size of the finite field. Figure 5 shows the probability that T ∈ (F_{2})^{C × C} is a full rank in field F_{2}.
Discussion
Consider the rank of a square matrix T in the finite field F_{q}. The probability that T ∈ (F_{q})^{C × C} is a full rank in field F_{q} is (1 − F_{q}^{−C}) × (1 − F_{q}^{−(C − 1)}) × ⋯(1 ‐ F_{q}^{−2}) × (1 ‐ F_{q}^{−1}) [9]. For the first selected row of T, the probability that this row is a nonzero vector is 1 − F_{q}^{−C}. For the second selected row of T, the probability this row is linearly independent with the first selected row is 1 − F_{q}^{−(C − 1)}. The rest rows can also be considered in a similar way. Consider the case that ∣F_{q} ∣ = 2 and C approximate infinity, the probability that T is a full rank is about 0.289 based on [9]. The probability rank of T that declines no more than 2 is (1 − 2^{−C}) × (1 − 2^{−(C − 1)}) × ⋯(1 ‐ 2^{−3}), where (1 ‐ 2^{−2}) × (1 ‐ 2^{−1}) is not included. If C approximates infinity, (1 − 2^{−C}) × (1 − 2^{−(C − 1)}) × ⋯(1 ‐ 2^{−3}) is about 0.7707. In Fig. 2, when ∣F_{q} ∣ = 2 and C = 50, the normalized declined amount about rank is very low. So, the rank of T also did not decline too much when ∣F_{q} ∣ = 2, and then, it means d_{min}' is very close to d_{min}.
This work mainly depends on information methods, if we adopt the methods from the machine learning field [11,12,13,14], we may correct more errors. The methods in [15,16,17,18] are also worth learning for the network coding error correction.
Conclusions and future work
With a small field ∣F_{q} ∣ = 2, the number of original errors can be compressed to less than the mincut of the network when their number is far more than the mincut of the network. The minimum distance of the newly formed rankmetric code in the small field also did not decline sharply. So, the propagated errors can be corrected by the listdecoding method of the newly formed rankmetric code. Our scheme can correct more than the mincut errors in network coding. A small field is also useful to reduce the computational burden compared to the bigger fields.
In our future research, we will try to optimize the scheme in real scenarios. We also will combine deep learning methods with network coding to improve the effectiveness of this method.
Availability of data and materials
Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.
Abbreviations
 NEC:

Network error correction
References
R.W. Yeung, N. Cai, Network error correction, I: Basic concepts and upper bounds. Commun. Inf. Syst. 6(1), 19–31 (2006)
G. Xuan, F.W. Fu, Z. Zhang, in International Symposium on Networking Coding. Construction of network error correction codes in packet networks (2011), pp. 1–6
Z. Zhang, Linear network error correction codes in packet networks. IEEE Trans Inform Theory 54, 209–218 (2008)
S. Yang, R.W. Yeung, K.N. Chi, Refined coding bounds and code constructions for coherent network error correction. IEEE Trans Inform Theory 57, 1409–1424 (2010)
R. Koetter, F.R. Kschischang, Coding for errors and erasures in random network coding. IEEE Trans Inform Theory 54, 3579–3591 (2007)
V. Guruswami, C. Wang, C. Xing, Explicit listdecodable rankmetric and subspace codes via subspace designs. IEEE Trans Inform Theory 62, 2707–2718 (2016)
W. Guo, H. Dan, C. Ning, On capacity of network error correction coding with random errors. IEEE Commun. Lett. 22, 696–699 (2018)
R. Koetter, M. Medard, in Joint Conference of the IEEE Computer and Communications Societies. Beyond routing: an algebraic approach to network coding (Proceedings, IEEE, 2002), pp. 122–130
P.A. Chou, Practical network coding. Allerton Conference on Communication, Control, and Computing, Monticello (2003)
È. Gabidulin, Theory of codes with maximum rank distance. Probl. Inf. Transm. 21, 3–16 (1985)
W.T. Cheng, Y. Sun, G.F. Li, G.Z. Jiang, H.H. Liu, Jointly network: A network based on CNN and RBM for gesture recognition. Neural Comput. Applic. 31(Suppl 1), 309–323 (2019)
G.F. Li, D. Jiang, Y.L. Zhou, G.Z. Jiang, J.Y. Kong, G. Manogaran, Human lesion detection method based on image information and brain signal. IEEE Access 7, 11533–11542 (2019)
G.F. Li, L.L. Zhang, Y. Sun, J.Y. Kong, Towards the sEMG hand: Internet of things sensors and haptic feedback application. Multimed. Tools Appl. 78(21), 29765–29782 (2019)
J. X Qi, G.Z Jiang, G.F Li, Y. Sun, B. Tao, Intelligent humancomputer interaction based on surface EMG gesture recognition, IEEE Access, 2019, 7: 6137861387.
Z. Huang, X. Xu, J. Ni, H. Zhu, W. Cheng, Multimodal representation learning for recommendation in internet of things. IEEE Internet Things J. 6(6), 10675–10685 (2019)
T. Zhou, J. Zhang, Analysis of commercial truck drivers’ potentially dangerous driving behaviors based on 11month digital tachograph data and multilevel modeling approach. Accid. Anal. Prev. 132, 105256 (2019)
Z. Chen, Y. Zhang, C. Wu, B. Ran, Understanding individualization driving states via latent Dirichlet allocation model. IEEE Intell. Transp. Syst. Mag. 11(2), 41–53 (2019)
Z. Huang, J. Tang, G. Shan, J. Ni, Y. Chen, C. Wang, An efficient passengerhunting recommendation framework with multitask deep learning. IEEE Internet Things J. (2019). https://doi.org/10.1109/JIOT.2019.2901759
Acknowledgements
The authors acknowledged the anonymous reviewers and editors for their efforts in valuable comments and suggestions.
Funding
This work was supported in part by Natural Science Foundation of Heilongjiang Province under Grant LH2019F052, and in part by Key topics of the 13th five year plan for Education Science in Heilongjiang Province in 2019 under Grant GJB1319161, in part by Key topics of the Ministry of education in 2017 in the 13th five year plan of National Education Science under Grant DCA170302, in part by Key Project of Natural Science Foundation of China under Grant 41631175, Innovation and entrepreneurship project of college students in Heilongjiang Province under Grant 201810236051.
Author information
Authors and Affiliations
Contributions
Dongqiu Zhang and Guangzhi Zhang conceived the idea almost at the same time when they are discussing the problem. Guangzhi Zhang designed the algorithm of the experiment. Mingyong Pang organized the work and wrote the paper. Dandan Huang assisted in the laboratory work and analyzed the results. The authors read and approved the final manuscript.
Authors’ information
Dongqiu Zhang is pursuing a doctorate degree in the Education Science Department of Nanjing Normal University. Her primary research interests include network coding, deep learning, and education technology.
Mingyong Pang was born in 1968, Ph.D., a professor in the Computer Department, Nanjing Normal University. Post doctor in the Computer Department, Nanjing University. His primary research interests include digital set processing and geometrybased modeling and rendering.
Guangzhi Zhang was born in 1982, Ph.D. in Computer Science, and an associate professor of Suihua University, China. His primary research interests include ad hoc networks and deep learning.
Dandan Huang is pursuing a master’s degree in the Education Science Department of Nanjing Normal University. Her primary research interests include digital image processing.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zhang, D., Pang, M., Zhang, G. et al. Research on correction algorithm of propagation error in wireless sensor network coding. J Wireless Com Network 2020, 116 (2020). https://doi.org/10.1186/s13638020017331
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13638020017331
Keywords
 Network coding
 Error correction
 Wireless sensor network
 Small field