 Research
 Open access
 Published:
Soft information acceleration aided subspace suppression MIMO detection
EURASIP Journal on Wireless Communications and Networking volume 2024, Article number: 51 (2024)
Abstract
In this paper, a multipleinput multipleoutput detection structure called soft information acceleration (SIA) is proposed, which is suitable for simplifying the twostage subspace marginalization with interference suppression (SUMIS) into one stage. The proposed onestage method outperforms the conventional twostage SUMIS when the subspace size is large enough. The performance advantage of the proposed SUMISSIA mainly results from the average number of soft information updates being equal to the ’subspace size,’ instead of only once during the twostage SUMIS detection. Thus, the SUMISSIA achieves a better tradeoff between performance and complexity. To further reduce the complexity, a channelshortening method based on subspace suppression is proposed. Simulation results show that the proposed channelshortening onestage method also outperforms SUMIS, which benefits from SIA.
1 Introduction
The fifth generation of mobile communication scenarios demands higher data transmission rates. Multipleinput multipleoutput (MIMO) [1,2,3,4,5] technology is a powerful means to adapt to this trend. MIMO systems enhance spectral efficiency through spatial multiplexing and use MIMO detection techniques at the receiver to recover the superimposed symbols in the spatial domain. However, the scaling up of MIMO presents a challenge to the base station receivers in the uplink scenario. The linear detection algorithms [6]—including zeroforcing (ZF) and minimum mean square error (MMSE)—can better cope with complexity rather than performance. In contrast, nonlinear detection can achieve better performance, but the number of symbol combinations traversed by nonlinear detection increases exponentially. As a result, receivers tend to adopt nonlinear detection for better performance. To reduce the complexity, many nonlinear detection schemes approximating the maximum likelihood (ML) are proposed, such as Kbest decoding [7] and likelihood ascent search based on K symbols (KLAS) [8]. These are both successful detection algorithms that offer a good tradeoff between performance and complexity, coming close to the performance of ML [9] with adjustable complexity.
Subspace suppression is another technique in tradeoff schemes, such as the twostage detection algorithm ‘subspace marginalization for interference suppression’ (SUMIS) [10]. SUMIS constructs an \(n_s\)dimensional subspace for each symbol and each subspace outputs soft information for that symbol only. Subspace partitioning and the precision of soft cancellation determine the quality of each symbol. Furthermore, both stages of SUMIS are well suited for parallel processing [11]. However, considering the coupling between different symbols, the two stages of SUMIS can be compressed into one stage by serial processing of symbols, subspace sorting, and result updating. Another method to build a tradeoff is the Ungerboeck observation model (UOM) with finite memory length. The UOM used to be a treebased approach. Since Rusek proposed parameter optimization and channel shortening method, the UOM can shorten the memory length based on the information theory [12] and run on the trellis.
The proposed soft information acceleration (SIA) structure improves SUMIS by introducing serial processing. It directly obtains soft information for all \(n_s\) symbols within the subspace through classification operations. Therefore, on average each symbol will get \(n_s\) versions of soft information. SIA corrects the cumulative soft information of symbols by utilizing multiple soft information versions per symbol, enhancing the accuracy of soft cancellation. Under the SIA structure, SUMIS detection is compressed into one stage with serial detection. Before detecting the subspace, the covariance matrix of interference and noise is updated based on previous detection results, achieving accelerated convergence of detection performance. Although the algorithm has only one stage, SUMISSIA outperforms SUMIS as \(n_s\) increases. This improvement is mainly due to the average number of soft information updates being \(n_s\), instead of only once as in SUMIS. We then proposed a channelshortening method based on subspace detection. Using this method, we combined SUMISSIA and UOM to propose another algorithm, USUMISSIA, to further reduce complexity. The results show that USUMISSIA can also achieve a good tradeoff and adjust the complexity by both \(n_{s}\) and memory length v, which is more flexible.
The notations in this paper are described below. Lowercase bold letters represent vectors, and uppercase bold letters represent matrices. \(\{\cdot \}^{T}\) and \(\{\cdot \}^{1}\) stand for matrix transpose and inverse, respectively. \(M_{i,j}\) is the entry at the ith row and jth column of matrix \(\textbf{M}\). \({\mathbb {E}}\{\cdot \}\) represents mathematical expectation. \(\Vert \cdot \Vert\) is the norm calculation for vectors. \(\cdot \) is the modular computation of the set.
2 Preliminaries
For the consistency with SUMIS algorithm, the real signal model is used in this paper. Considering an uplink MIMO system with \(N_T /2\) transmitting antennas and \(N_R /2\) receiving antennas, the real reception model is
where \(\textbf{y}\in {{\mathbb {R}}^{{{N_{R}}}\times 1}}\), \(\textbf{H}\in {{\mathbb {R}}^{ {N_{R} }\times N_T }}\), \(\textbf{x}\in {{\mathbb {R}}^{{N_T}\times 1}}\), \(\textbf{n}\in {{\mathbb {R}}^{{{N_{R}}}\times 1}}\) are the real received vector, real channel matrix, transmitted real symbol vector, and real white Gaussian noise, respectively. \(\textbf{x}\), \(\textbf{y}\), \(\textbf{n}\), and \(\textbf{H}\) are calculated from the corresponding complex form matrices \(\textbf{x}^\mathbb {C}\), \(\textbf{y}^\mathbb {C}\), \(\textbf{n}^\mathbb {C}\), \(\textbf{H}^\mathbb {C}\) as follows
Each element of \(\textbf{n}\sim \mathcal {N}(0,\gamma )\), where \(\gamma =\frac{N_T{N}_{0}}{4}\) and \(N_0\) is the power of noise. The element \({x}_{s}\) of \(\textbf{x}\) belongs to the alphabet \(\chi\) (such as 2PAM, \({{x}_{s}}\in \chi = \{1,1\}\)), \({\mathbb {E}}\{ {x}_{s} \}=0\), variance of \({x}_{s}\) is 1, where \(s\in {\{1, \cdots , N_T \}}\).
The optimal method to detect \(\textbf{x}\) from (1) is ML [9] with exponential complexity. To avoid the huge complexity of ML, SUMIS [10] was proposed to reduce the complexity, which constructs a subspace of size \(n_s\) for the sth symbol and detects sth symbol within it. In twostage SUMIS, the sth subspace must contain the sth symbol. Thus, \(N_T\) subspaces of twostage SUMIS can also be indexed by s. The subspace division is based on
where \({\rho }_{s,j}\) is the inner product of the sth and jth real channel of \(\textbf{H}\), and the larger \({\rho }_{s,j}\) is, the stronger the correlation between sth channel and jth channel is. Subspace s contains sth symbol and other \(n_{s}\)1 symbols with the strongest correlation to sth symbol. Thus the reception model becomes
where \({\bar{\textbf{H}}}{\bar{\textbf{x}}}\) and \(\tilde{\textbf{H}}\tilde{\textbf{x}}\) are the detection components and interference components of \(\textbf{y}\), respectively. Obviously, the elements associated with sth symbol are contained in \(\bar{\textbf{x}}\) and \(\bar{\textbf{H}}\).
SUMIS contains two stages. In the first stage (S1), SUMIS detects \(N_T\) subspaces without (priori) soft information of interference subspace. The result of subspace s only contains loglikelihood ratios (LLRs) of sth symbol. The ith LLR of s calculated by maxlog is
where \({{b}_{s,i}}\) is the ith bit of sth symbol and \(\textbf{Q}\) is the covariance matrix of ‘\(\textbf{y} {{\bar{\mathbf{H}}}{\bar{x}}}\)’. \(\left\ \textbf{y} {{\bar{\mathbf{H}}}{\bar{x}}} \right\ _{\textbf{Q}}^{2}\) is the shorthand of the inner product \((\textbf{y} {{\bar{\mathbf{H}}}{\bar{x}}})^{T}{\textbf{Q}}^{1}(\textbf{y} {{\bar{\mathbf{H}}}{\bar{x}}})\), which can be simplified as
where \({{\bar{\textbf{H}}}^{r}}={{\left( {{\textbf{Q}}^{1}} \right) }}\bar{\textbf{H}}\), \({{\bar{\textbf{G}}}^{r}}={{\bar{\textbf{H}}}^{T}}{{\textbf{Q}}^{1}}\bar{\textbf{H}}\) and \(\textbf{Q}=\tilde{\textbf{H}}\tilde{\textbf{H}}^{T}+\gamma \textbf{I}\).
In the second stage (S2), SUMIS detects \(N_T\) subspaces after canceling soft information of the interference subspace. The posteriori expectation vector \({\mathbb {E}}\left\{ \tilde{\textbf{x}}\textbf{y} \right\}\) (calculated by S1’s result) of symbols in the interference subspace should be canceled from \(\textbf{y}\) to get \(\textbf{y}'\) by \(\mathbf {{y}'}\triangleq \textbf{y}\tilde{\textbf{H}}{\mathbb {E}}\left\{ \tilde{\textbf{x}} \right\}\) and then \(\textbf{Q}'=\tilde{\textbf{H}}\bar{\varvec{\Phi }}\tilde{\textbf{H}}^{T}+\gamma \textbf{I}\) instead of \(\textbf{Q}\). The operation \(\mathbf {{y}'}\triangleq \textbf{y}\tilde{\textbf{H}}{\mathbb {E}}\left\{ \tilde{\textbf{x}} \right\}\) means subspace suppression. \(\bar{\varvec{\Phi }}\) is a diagonal matrix composed of the posteriori variances of the interference symbols. The probability of s at constellation x is defined as \({p}_{s,x}\) which is the ‘soft information’ throughout the paper. In SUMIS, the \({p}_{s,x}\)s is calculated from the LLRs of S1. So the posteriori expectation and posteriori variance of sth symbol are \(\sum \limits _{x\in \chi }{x\cdot {{p}_{s,x}}}\) and \(\sum \limits _{x\in \chi }{x^{2}\cdot {{p}_{s,x}}}\{ \sum \limits _{x\in \chi }{x\cdot {{p}_{s,x}}} \} ^2\), respectively. It is worth noting that in S1, the soft information of the sth symbol is directly from the sth subspace. Thus, the soft information softly canceled by S2 is updated only once.
The LLRs in (4) is calculated only for the sth symbol while all cases of \(\bar{\textbf{x}}\) and \(\left\ \textbf{y} {{\bar{\mathbf{H}}}{\bar{x}}} \right\ _{\textbf{Q}}^{2}\) within such subspace are traversed. If all \(\bar{\textbf{x}}\) are properly classified, the LLRs of the other \(n_{s}\)1 symbols can also be obtained directly. In each stage of SUMIS, the detection of subspaces is carried out independently, so each of the two stages of SUMIS can be processed parallelly. The core idea of this paper is that by only one stage, detecting the \(N_T\) subspaces serially and cumulatively updating the soft information for better performance. Specifically, the sth subspace detection comprehensively utilizes the soft information from the previous \(s1\) subspaces and executes soft cancellation in S1, instead of S2.
3 Methods
First, a onestage SUMISSIA detection method through ‘serial detection’ and ‘utilizing accumulated soft information to perform interference cancellation’ is proposed. Then parameter optimization and channel shortening methods based on subspace suppression are derived. Based on these methods, USUMISSIA is proposed to further reduce the complexity.
Since SIA involves the adjustment of subspace order, define t as the index of sorted subspaces. The left side of Fig. 1 is an example of subspaces division while \(N_T\)=8 and \(n_s\)=3. In the subspace [1, 6, 4], SUMIS only calculates the LLRs of symbol 1 while SUMISSIA and USUMISSIA directly output the LLRs of symbols 1, 6, and 4 with almost no extra overhead. Define \(\varvec{\Sigma }^{t}_{a}\in {{\mathbb {R}}^{\log _{2}(\left \chi \right )\times {{N_T}}}}\) (initialized by \(\varvec{\Sigma }^{0}_{a}=\textbf{0}\)) for each symbol to store the accumulative LLRs throughout the detection process. Due to the serial detection architecture, the superscript t here in \(\varvec{\Sigma }^{t}_{a}\) represents the number of updates as the number of subspace detections increases. And \(\varvec{\Sigma }^{t}_{a}\)’s sth column \(\varvec{\sigma }^{t}_{s}\) is the accumulative LLRs of sth symbol. Define \(\varvec{\lambda }^{t}_{s}\) (\(s \in {1,\cdots ,n_s}\)) as the output LLRs from the tth subspace. If the tth subspace is detected, the \(\varvec{\lambda }^{t}_{s}\)s of \(n_s\) symbols will be merged into the corresponding \(\varvec{\sigma }^{t1}_{s}\)s to get \(\varvec{\sigma }^{t}_{s}\)s. The soft cancellation operation of (t+1)th subspace will be based on \(\varvec{\Sigma }^{t}_{a}\), which increases the performance of (t+1)th subspace.
3.1 SUMISSIA
The details of SUMISSIA are given in Algorithm 1. Line 4 indicates that each subspace is detected only once. SUMISSIA detects subspaces in sorted serial order. Before a subspace detection, it is necessary to cancel the cumulative soft information of the interfering symbol. After a subspace detection, the cumulative LLRs need to be merged with the LLRs detected by the current subspace. The rest of this subsection will introduce the changes of SUMISSIA from SUMIS in three aspects, which include the sorting of subspaces, the improved subspace detection, and the LLR merging.
3.1.1 Sorting of subspaces
The division of subspace also uses \(\textbf{H}^{T}\textbf{H}\) as SUMIS. In SUMIS, subspaces are detected in natural order (and parallel), and the symbol s must be included in sth subspace. SUMISSIA needs to adjust the above natural order first. By counting the total \(n_{s} \times N_T\) symbols of all detection subspaces, it can be found that the occurrence times of each symbol may be different. SUMISSIA sorts the subspaces according to the descending order of occurrence times as Fig. 1 shows. In Fig. 1, Stream 4 (and 8) has a total of 5 occurrences in all subspaces, so detect the Subspace 4 (or 8) first. The occurrence times of the 8 real symbols are “1, 4, 2, 5, 1, 4, 2, 5” respectively. Arranging the above occurrence times in descending order, SUMISSIA obtains the detection order of the serial subspace as: “4, 8, 2, 6, 3, 7, 1, 5”. The LLRs in \(\varvec{\Sigma }^{t}_{a}\) will gradually become more accurate; serially detecting the subspaces with ‘descending order’ will get more accurate LLRs in \(\varvec{\Sigma }^{t}_{a}\) of the symbols with fewer occurrence time, so that increasing the accuracy of accumulative LLRs comprehensively. Figure 4 in Section Results and Discussion proves the above statement.
3.1.2 Subspace detection
Besides \(\varvec{\Sigma }^{t}_{a}\), define \(\varvec{\Sigma }^{t}_{c}\in {{\mathbb {R}}^{ \log _{2}(\left \chi \right ) \times N_T}}\) (initialized by \(\textbf{0}\)) to store the last LLRs detected from a subspace. The reason for \(\varvec{\Sigma }^{t}_c\) is needed is that SUMISSIA outputs ‘\(\varvec{\lambda }\)’ consistent with SUMIS rather than ‘\(\varvec{\sigma }\)’, which may not belong to the \(N_T\)th subspace (such as the right side of Fig. 1, the 1st symbol doesn’t belong to the 8th subspace).
Before tth subspace detection, SUMISSIA selects \(\varvec{\sigma }^{t1}_{m}\)s from \(\varvec{\Sigma }^{t1}_{a}\) (index m is corresponding to the interference subspace) and calculates \(\bar{\varvec{\Phi }}\) to give the most accurate soft information so far. The \(\bar{\textbf{x}}\) traversal of tth subspace detection is consistent with twostage SUMIS. Then the LLR vectors of the tth detection \(\varvec{\lambda }^{t}_{s}\)s shall replace the corresponding columns of \(\varvec{\Sigma }^{t1}_{c}\) to get \(\varvec{\Sigma }^{t}_{c}\) and merged with the accumulative LLR \(\varvec{\sigma }^{t1}_{s}\)s through the method of 3.1.3 to get \(\varvec{\sigma }^{t}_{s}\)s and \(\varvec{\Sigma }^{t}_{a}\). \(\varvec{\Sigma }^{N_T}_{c}\) is the final output. The rationality here also lies in the growing accuracy of soft information.
3.1.3 The merger of LLRs
SUMISSIA uses a damped merge approach as [13]. When subspace detection obtains a series of new LLRs, they are merged with the previous accumulative LLRs as
Through a large number of tests, \(\zeta = 0.6\) gives good performance in various simulations. Thus the LLRs of a symbol from all detection subspaces are merged together smoothly. This means that SIA needs to gradually correct LLR from the a priori equivalency of two cases of each bit.
3.2 USUMISSIA with trellis structure
Taking advantage of UOM’s reduction complexity method by shortening memory length, we then propose USUMISSIA, which also takes one stage as SUMISSIA while the trellis of UOM is introduced to the subspace detection at line 5 to 7 in Algorithm 1. The channel input–output relationship based on \(\mathbf {{y}}\) and \(\textbf{Q}\) (or \(\mathbf {{y}'}\) and \(\textbf{Q}'\)) of (3) is
where \({{\bar{\textbf{H}}}^{r}}\) and \({{\bar{\textbf{G}}}^{r}}\) are same as them in (5). And the recursive factorization of (7) is
Combining (7) with channel shortening can further reduce the complexity of subspace detection under the same \(n_s\). For a limited memory v [14], the number of states of trellis is equal to \({{\left( \sqrt{\left \chi \right } \right) }^{v}}\), and there are \({{\left( \sqrt{\left \chi \right } \right) }^{v+1}}\) branches within a trellis unit. if \(v<n_{s}1\), \(\bar{\textbf{G}}^{r}\) is a symmetric band matrix, i.e., \({{{\bar{G}}}^{r}_{m,n}}=0\), if \(mn>v\). Recursive factorization [15] of (7) with v is
where \(\textbf{r}={( {{{\bar{\textbf{H}}}}^{r}} )^{T}}\textbf{y}\). By comparing the changes from (8) to (9), channel shortening is manifested in the fact that the kth symbol is only related to the forwardneighboring v1 symbols. Based on (9), BCJR in [16] is adopted to complete the detection. The branch metric [17] (gamma) is
Combining with (10), other operations (recursive calculation of \(\alpha\)s, \(\beta\)s and to calculate the symbol’s marginal probability (decision) of the trellis) of BCJR can be run on the trellis.
The following of this subsection will introduce how to shorten the channel under subspace suppression. Unlike the description in [14], on the premise that the channel is shortened, \({{\bar{\textbf{H}}}^{r}}\) and \({{\bar{\textbf{G}}}^{r}}\) are given as
Upper triangular matrix \(\textbf{V}\) satisfies \({V_{m,n}}=0\), if \(nm>v\), and its calculation is consistent with that in [14].
Proposition 1
Define
Define \(\textbf{B}_{k}^{v}\) as the submatrix of \(\textbf{B}\),
Define \(\textbf{b}_{k}^{v}=\left[ B_{k,k+1}^{},\cdots ,B_{k,\min (n_{s},k+v)} \right]\) as the row submatrix of \(\textbf{B}\), and \(\textbf{v}_{k}^{v}=\left[ \textbf{V}_{k,k+1}^{{}},\cdots ,\textbf{V}_{k,\min (n_{s},k+v)} \right]\) as the row submatrix of \(\textbf{V}\). Define
Finally, define \({{v}_{k,k}}={{\left( {{c}_{k}} \right) }^{\frac{1}{2}}}\) as the diagonal element of the matrix \(\textbf{V}\), and
together with \({v}_{k,k}\) form the row submatrix of \(\textbf{V}\). \(\square\)
The proof of (11) is given in Appendix A. Due to (3), derivation process only analyze \(\bar{\textbf{H}\bar{x}}\), which leads to \(\gamma \textbf{I}\) changing to \(\textbf{Q}\). It should be noted that although UOM directly obtains symbollevel probabilities, USUMISSIA still follows the LLR merger method described in 3.1.3. This requires converting the symbollevel probabilities into bitlevel LLRs
where \(p_{s,x}\) is the symbollevel probability from the BCJR decision. The other operations are no different from Algorithm 1.
3.3 Complexity analysis
This subsection discusses the complexity of the SUMIS, SUMISSIA, and USUMISSIA based on real addition and real multiplication. Firstly, partial ML in a subspace requires preprocessing of \({{\bar{\textbf{H}}}^{T}}{{\textbf{Q}}^{1}}\bar{\textbf{H}}\). Based on (5), traversal (\(\bar{\textbf{x}}\)) complexity of a partial ML subspace detection is
which is the complexity of detecting a subspace once. For each of the \(N_T\) subspaces, SUMIS requires two times the above preprocessing and (18) while SUMISSIA requires only once.
Secondly, USUMISSIA only needs to preprocess and traverse each \(N_T\) subspace once. BCJR detection in subspace requires preprocessing of \({{\textbf{Q}}}\), (11) and (12). Due to the existence of finite v, the BCJR detection structure is a combination of a tree and trellis. The BCJR algorithm adopted by a USUMISSIA’s subspace needs to iterate over a trellis with branches of
within a subspace detection. The BCJR algorithm can use maxlog algorithm [18] to merge \(\alpha\)s, \(\beta\)s and to calculate the symbol’s marginal probability (decision) of each trellis unit, these above operations are directly related to the number of branches in (19). The gamma computation of a subspace is
At the same time, total \(\alpha\) and \(\beta\) computation need multiplication of two times (19) respectively, and total decision computation needs the addition of one time (19).
4 Results and discussion
This section introduces the experimental results of SUMISSIA and USUMISSIA in fastfading MIMO channels. In the MIMO scenario, \(N_T/2\) codewords are transmitted in parallel over complex MIMO channels. Under fastfading Rayleigh channels, modulation symbols mapped from the same codeword will experience different channel fading coefficients. The elements in matrix \(\textbf{H}\) are independently and identically distributed, following a real Gaussian distribution with \(\mathcal {N}(0,\frac{1}{2})\). For ease of expression, this section describes the configuration in complex MIMO. For example, the realform “32 (\(N_T\)) \(\times\) 32 (\(N_R\)) MIMO 4PAM” will be described as the complexform “\(16\times 16\) MIMO 16QAM”. The discussion in this section focuses on \(E_b/N_0\), where the bit error rate (BER) being discussed of the performance curve is \(10e^{4}\), and each point is based on 300 codeword errors. The codewords are encoded using LDPC codes with a length of 576 from 802.16e [20], and the output of the detector is decoded using a layered decoder with up to 25 iterations. The code rate is 1/2 (for Fig. 2) or 3/4 (for Fig. 3, 5, 6, 7, 8, 9, 10, 11 and 12).
Fig. 2 presents the simulation results for complex 6x6 MIMO 4QAM with 1/2 LDPC, including SUMIS, SUMISSIA, ML, MMSE, and Achievable Information Rate based Partial Marginalization (AIRPM) [21]. The size of r determines the complexity of maximum likelihood search in AIRPM, and r satisfies \(n_s\)=r+1. In the experimental results, SUMIS has matched the detection performance of ML at \(n_s\)=3 (with a gap of less than 0.1 dB). SUMISSIA needs to achieve performance close to ‘\(n_s\)=3 SUMIS’ with \(n_s\)=4 (with a gap of 0.05 dB), and the search complexity of ‘\(n_s\)=3 SUMIS’ and ‘\(n_s\)=4 SUMISSIA’ is similar. This indicates that under smallscale MIMO and loworder modulation, SUMISSIA has fewer soft information updates, which cannot reflect its advantages. Furthermore, SUMISSIA has at least a 1 dB advantage over linear detection. AIRPM requires several times the preprocessing overhead of SUMISSIA to achieve the performance of ‘\(n_s\)=4 SUMISSIA’ at r=5. To demonstrate the advantages of SUMISSIA, the next experimental configuration will increase the MIMO scale and modulation order.
Figure 3 focuses on the impact of \(n_{s}\) on changes in the performance gap between USUMISSIA and SUMIS under \(16 \times 16\) MIMO 16QAM. It can be found that as \(n_{s}\) increases from 2 to 4, only when \(n_s = 2\), SUMIS has smaller BER, and when \(n_{s}\) becomes larger, SUMISSIA outperforms SUMIS. As \(n_s\) increases, SUMISSIA has faster performance convergence. This is because the larger \(n_s\) is, the more the average number of updates of each symbol’s soft information is, and the more accurate the obtained symbol soft information is.
In order to prove the correctness of the above statement, this section gives a performance comparison of the subspaces. According to the description of SUMIS, S2 performs a soft cancellation operation of interference subspace symbols from the received signal. It is precisely because the symbol soft information of the interference subspace is close to the correct transmitted symbol that S2 of SUMIS can obtain better results than S1. If the accuracy of the symbol is higher, the symbollevel mathematical expectation is closer to the correct symbol. Based on this, if the soft information of SUMISSIA gradually becomes more accurate, the accuracy of subspace detection should also gradually increase. Figure 4 shows the SER simulation results under \(4 \times 4\) MIMO 16QAM. The corresponding subspace index of the performance curve in Fig. 4 is the reordered subspace index as 3.1.1. From this result, it can be seen that in this group of 8 subspace detections, the accuracy of the hard decision of the symbol gradually increased, which confirmed that the soft information of the symbol gradually became more accurate.
Figures 3, 5, 6, 7 and 8 show the performance gains of SUMISSIA under different simulation configurations. As the modulation orders or MIMO scales increase, the advantages of SUMISSIA over SUMIS become increasingly significant. This means that with only a small \(n_s\) required, the BER performance of SUMISSIA can be superior to SUMIS. For example, in the \(16 \times 16\) MIMO 4QAM scenario, when \(n_s = 5\), the BER performance of SUMISSIA is better than that of SUMIS, while in the \(16 \times 16\) MIMO 16QAM scenario, only \(n_s = 3\) is needed. In the \(8 \times 8\) MIMO configuration, the BER performance of SUMISSIA is better than that of SUMIS only when the modulation order reaches 64. It is worth noting that Figs. 3, 5, 6, 7 and 8 are only for displaying the performance changes of SUMISSIA. The BER performance and computational complexity of SUMISSIA are better than SUMIS in most scenarios. The reason SUMISSIA can outperform SUMIS is entirely due to the sufficient number of symbol updates provided by \(n_s\). Under the SIA structure, the average number of updates for the soft information of symbols is \(n_s\), rather than just once in SUMIS.
The complexity of SUMISSIA is half that of SUMIS. When \(n_s\) increases enough to enable SUMISSIA to outperform SUMIS, the performancecomplexity tradeoff of SUMISSIA can outperform that of SUMIS: achieving better performance than SUMIS with lower complexity than SUMIS.
USUMISSIA changes the subspace detection method of SUMISSIA to a tree or trellis structure based on UOM. In the SIA architecture, the average number of symbol soft information updates is still \(n_s\). In the previous analysis, SUMISSIA has already outperformed SUMIS by utilizing a sufficient \(n_s\). The following experiments demonstrate that as long as \(n_s\) is large enough, USUMISSIA can still outperform SUMIS by moderately reducing the memory length v. Figures 9, 10, and 11 show the performance of USUMISSIA with subspace detection using channel shortening. Firstly, when \(n_s\)=3 and v=1, USUMISSIA can achieve performance comparable to ‘\(n_s\)=3 SUMIS’. When \(n_s\)=4 and \(n_s\)=5, even if the memory length of subspace detection is shortened to v=2, USUMISSIA can still outperform SUMIS. These experimental results fully demonstrate the effect of \(n_s\) on performance improvement in the SIA: a sufficiently large \(n_s\) provides sufficiently accurate soft information, allowing subspace detection to tolerate moderate channel shortening.
In order to more intuitively understand the complexity of the algorithms in this figure, the subspace detection complexity corresponding to all involved configurations is listed in Table 1. Each data in the table represents the sum of real multiplication and real addition required for subspace detection under such algorithm and configuration. As shown in Table 1, when measuring complexity in terms of real addition and multiplication, the complexity relationship under the same configuration is “SUMIS > SUMISSIA > USUMISSIA”. Firstly, the preprocessing and traversal complexities of SUMISSIA are both half of those of SUMIS. The computational complexity of USUMISSIA is further reduced. This is mainly because USUMISSIA employs the maxlog BCJR algorithm, where some multiplications and additions are simplified to additions and comparisons.
If \(n_s\) is sufficiently large, even with a moderate reduction in memory length v, USUMISSIA still outperforms SUMIS. Combining the complexity statistics in Table 1, the performancecomplexity tradeoff of USUMISSIA can also outperform that of SUMIS: achieving better performance than SUMIS with lower complexity than SUMIS.
The results show that when \(n_s\) is greater than a certain threshold, SUMISSIA based on SIA’s reasonable utilization and combination of soft information can outperform SUMIS. As previously mentioned, under SIA architecture, the symbol’s soft information needs to be updated many times (instead of SUMIS’s only one time in its S1). The above situation allows the detector to achieve better performance than SUMIS at a computational complexity lower than SUMIS. Situation results show that both SUMISSIA and USUMISSIA are very good tradeoff cases between complexity and performance.
The following experiments will evaluate the detection performance of moderately parallelized SIA. According to the Algorithm 1, SIA is a fully serialized detection algorithm. In this experiment, the parallelism of SIA is increased to 2. In the improved SIA, subspaces \(t_o\) and \(t_o\)+1 (where \(t_o\) is an odd number) share the same set of soft information for soft cancellation. After the two parallel subspace detections are completed, the subspace output results are merged into the accumulated soft information. Figure 12 compares the performance of SUMISSIA and SUMISSIA with a parallelism of 2 under \(8\times 8\) MIMO 16QAM. When \(n_s\)=2, SUMISSIA shows a performance advantage of approximately 0.12dB. When \(n_s\) increases to 3 and 4, the performance gap between SUMISSIA and ‘SUMISSIA with parallelism 2’ is only 0.07dB. The experimental results show that as the average number of symbol updates increases, the performance loss caused by parallelism can be partially compensated. This is also attributed to the multiple updates of symbol soft information in SIA.
5 Conclusion
Two detection algorithms, SUMISSIA and USUMISSIA, are proposed by combining SUMIS architecture and soft information acceleration as the main contribution of this paper. For USUMISSIA, channel parameter optimization and channel shortening methods suitable for subspace suppression are also proposed so that the proposed algorithm can be flexibly adjusted in two aspects of subspace size \(n_{s}\) and memory length v. Compared with the SUMIS under the same \(n_{s}\), the complexities of these two algorithms are lower. Under scalingup MIMO and high order modulation, with the increasing of \(n_s\), the performance convergence of SUMISSIA is faster than that of SUMIS; under sufficiently large \(n_s\), the performance of USUMISSIA after moderate channel shortening is also better than that of SUMIS. To sum up, the two algorithms proposed in this paper can rely on SIA to achieve a better tradeoff between performance and complexity.
Availability of data and materials
The datasets used during the current study are available from the corresponding author on reasonable request.
Materials availability
Not applicable.
Code availability
Not applicable.
Abbreviations
 MIMO:

Multiple input multiple output
 SIA:

Soft information acceleration
 SUMIS:

Subspace marginalization with interference suppression
 ZF:

Zero forcing
 MMSE:

Minimum mean square error
 ML:

Maximum likelihood
 KLAS:

Likelihood ascent search based on K symbols
 UOM:

Ungerboeck observation model
 LLR:

Loglikelihood ratios
 AIRPM:

Achievable information rate based partial marginalization
 BER:

Bit error rate
References
E.G. Larsson, O. Edfors, F. Tufvesson, T.L. Marzetta, Massive MIMO for next generation wireless systems. IEEE Commun. Mag. 52(2), 186–195 (2014). https://doi.org/10.1109/MCOM.2014.6736761
L. You, K.X. Li, J. Wang, X. Gao, X.G. Xia, B. Ottersten, Massive MIMO transmission for Leo satellite communications. IEEE J. Sel. Areas Commun. 38(8), 1851–1865 (2020). https://doi.org/10.1109/JSAC.2020.3000803
Z. Lin, M. Lin, T. Cola, J.B. Wang, W.P. Zhu, J. Cheng, Supporting IoT with ratesplitting multiple access in satellite and aerialintegrated networks. IEEE Internet Things J. 8(14), 11123–11134 (2021). https://doi.org/10.1109/JIOT.2021.3051603
Z. Lin, H. Niu, K. An, Y. Wang, G. Zheng, S. Chatzinotas, Y. Hu, Refracting risaided hybrid satelliteterrestrial relay networks: Joint beamforming design and optimization. IEEE Trans. Aerosp. Electron. Syst. 58(4), 3717–3724 (2022). https://doi.org/10.1109/TAES.2022.3155711
Z. Lin, M. Lin, B. Champagne, W.P. Zhu, N. AlDhahir, Secrecyenergy efficient hybrid beamforming for satelliteterrestrial integrated networks. IEEE Trans. Commun. 69(9), 6345–6360 (2021). https://doi.org/10.1109/TCOMM.2021.3088898
K.K. Gill, K. Kaur, Comparative analysis of ZF and MMSE detections for nakagamim faded MIMO channels, in Paper presented at 2015 annual IEEE India conference (INDICON), (2015), pp. 1–5. https://doi.org/10.1109/INDICON.2015.7443281
Z. Guo, P. Nilsson, Algorithm and implementation of the kbest sphere decoding for mimo detection. IEEE J. Sel. Areas Commun. 24(3), 491–503 (2006). https://doi.org/10.1109/JSAC.2005.862402
Z. Qin, J. Xu, X. Tao, Z. X., Improved DepthFirstSearch Sphere Decoding Based on LAS for MIMOOFDM Systems, in Paper presented at 2015 IEEE 82nd vehicular technology conference (VTC2015Fall), (2015), pp. 1–5. https://doi.org/10.1109/VTCFall.2015.7390976
M.O. Damen, H.E. Gamal, G. Caire, On maximumlikelihood detection and the search for the closest lattice point. IEEE Trans. Inf. Theory 49(10), 2389–2402 (2003). https://doi.org/10.1109/TIT.2003.817444
M. Čirkić, E.G. Larsson, Sumis: nearoptimal softin softout mimo detection with low and fixed complexity. IEEE Trans. Signal Process. 62(12), 3084–3097 (2014). https://doi.org/10.1109/TSP.2014.2303945
W. Haselmayr, G. Möstl, S. Seeber, A. Springer, Hardware implementation of the SUMIS detector using highlevel synthesis, in Paper presented at 2015 IEEE international symposium on circuits and systems (ISCAS), (2015), pp. 29722975. https://doi.org/10.1109/ISCAS.2015.7169311 (2015)
F. Rusek, O. Edfors, An information theoretic charachterization of channel shortening receivers, In Paper presented at 2013 asilomar conference on signals, systems and computers, (2013), pp. 2108–2112. https://doi.org/10.1109/ACSSC.2013.6810679
J. Gao, D. Zhang, J. Dai, K. Niu, C. Dong, Resnetlike beliefpropagation decoding for polar codes. IEEE Wirel. Commun. Lett. 10(5), 934–937 (2021). https://doi.org/10.1109/LWC.2021.3050819
F. Rusek, A. Prlja, Optimal channel shortening for mimo and ISI channels. IEEE Trans. Wirel. Commun. 11(2), 810–818 (2012). https://doi.org/10.1109/TWC.2011.121911.110809
F. Rusek, G. Colavolpe, C.E.W. Sundberg, 40 years with the ungerboeck model: a look at its potentialities [lecture notes]. IEEE Signal Process. Mag. 32(3), 156–161 (2015). https://doi.org/10.1109/MSP.2014.2374221
L. Bahl, J. Cocke, F. Jelinek, J. Raviv, Optimal decoding of linear codes for minimizing symbol error rate (corresp.). IEEE Trans. Inform. Theory 20(2), 284–287 (1974). https://doi.org/10.1109/TIT.1974.1055186
F. Rusek, M. Loncar, A. Prlja, A comparison of ungerboeck and forney models for reducedcomplexity ISI equalization, in Paper presented at IEEE GLOBECOM 2007—IEEE global telecommunications conference, (2007), 14311436. https://doi.org/10.1109/GLOCOM.2007.275
D. Zhang, J. Liu, X. Yang, H. Ji, MaxLogMAP and LogMAP Decoding Algorithms for UST Symbol Based Turbo Code, in Paper presented at 2008 4th international conference on wireless communications, networking and mobile computing, (2007), pp. 15 https://doi.org/10.1109/WiCom.2008.355
A. Krishnamoorthy, D. Menon, Matrix inversion using Cholesky decomposition, in Paper presented at 2013 signal processing: algorithms, architectures, arrangements, and applications (SPA), (2013), pp. 70–72
IEEE p802.16e: Part 16: air interference for fixed and mobile broadband wireless access systems. IEEE 802.16 document (2005)
S. Hu, F. Rusek, A softoutput mimo detector with achievable information rate based partial marginalization. IEEE Trans. Signal Process. 65(6), 1622–1637 (2017). https://doi.org/10.1109/TSP.2016.2641393
Acknowledgements
Not applicable.
Funding
This research is funded by the Key Program of the National Natural Science Foundation of China of funder grant number 92067202.
Author information
Authors and Affiliations
Contributions
An algorithm jointly proposed by four authors. Xiaoxiong Xiong completed the simulation work and wrote the paper.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
USUMISSIA with trellis structure
USUMISSIA with trellis structure
In this paper, the lower bound to the information rate \({{I}_{\text {LB}}}\triangleq {{\mathbb {E}}_{\textbf{Y}}}[{{\log }_{2}}(\tilde{p}(\textbf{y}))]+{{\mathbb {E}}_{\textbf{Y},\bar{\textbf{X}}}}[{{\log }_{2}}(\tilde{p}(\textbf{y}\bar{\textbf{x}}))]\) [14] is used for derivation. Integrating (7) gives the expectation of \(\tilde{p}(\textbf{y})\). Define \({{\bar{\textbf{G}}}^{\text {r}}}=\mathbf {U\Delta }{{\textbf{U}}^{T}}\) (\(\textbf{U}^{T}\textbf{U}=\textbf{I}\)) as the eigenvalue decomposition expression of \(\bar{\textbf{G}}^{\text {r}}\) and \(\textbf{z}={{\textbf{U}}^{T}}\bar{\textbf{x}}\).
Firstly, the follwing is the derivation of the part \({{\mathbb {E}}_{\textbf{Y}}}[{{\log }_{2}}(\tilde{p}(\textbf{y}))]\). Given no priori probability about \(\bar{\textbf{x}}\), \(\mathbb {E}\{\bar{\textbf{x}}\}=\textbf{0}\) and the variance of the elements of \(\bar{\textbf{x}}\) are 1. Given \(\int {e^{x^{2}}}dx = \sqrt{\pi }\), substitute the above disassembly into (7) and integrate it to obtain the following expression,
where \({d}_{k}\) is from vector \(\textbf{d} \triangleq \textbf{U}^{T} \left( \bar{\textbf{H}}^{r} \right) ^{T} \mathbf {{y}}\) (or \(\textbf{d} \triangleq \textbf{U}^{T} \left( \bar{\textbf{H}}^{r} \right) ^{T} \mathbf {{y}'}\) for S2). According to (A1), the expectation \({{\mathbb {E}}_{\textbf{Y}}}[{{\log }_{2}}(\tilde{p}(\textbf{y}))]\) is
Define the covariance matrix \(\textbf{R}\) of \(\textbf{d}\) as
And \({{{\mathbb {E}}_{\textbf{Y}}}[{{d}_{k}}{{}^{2}}]}\) is a diagonal element of the \(\textbf{R}\) matrix.
Secondly, the follwing is the derivation of the part \({{\mathbb {E}}_{\textbf{Y},\bar{\textbf{X}}}}[{{\log }_{2}}(\tilde{p}(\textbf{y}\bar{\textbf{x}}))]\). The expectation in trace form is
and \(\textbf{y}'\) is used when S2.
Notice that the terms related to \(\bar{\textbf{H}}^{r}\) are \(\sum \limits _{k=1}^{{{n}_{s}}}\frac{{{R}_{kk}}}{2(\delta _{k}+1)}\) and \(\text {Tr}({{({{\bar{\textbf{H}}}^{\text {r}}})}^{T}}\bar{\textbf{H}})\). In S1, for example, rewrite \(\sum \limits _{k=1}^{{{n}_{s}}}\frac{{{R}_{kk}}}{2(\delta _{k}+1)}\) in matrix form as
Define objective function as
Given the gradient calculation methods \(\frac{\partial \text {Tr}\left( {{\textbf{Z}}^{T}}\textbf{A} \right) }{\partial \textbf{Z}}=\textbf{A}\) and \(\frac{\partial \text {Tr}\left( {{\textbf{Z}}^{T}}\textbf{BZA} \right) }{\partial \textbf{Z}}={{\textbf{B}}^{T}}\textbf{Z}{{\textbf{A}}^{T}}+\textbf{BZA}\), the gradient of (A7) is calculated as
At zero point of the first derivative, \(\bar{\textbf{H}}^{\text {r}}\) is as follow
which guarantees \(I_{LB}\) is optimal. What’s more, \(\textbf{Q}'\) is used in (A9) when S2. So far, (11) has been proved.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Xiong, X., Dong, C., Bian, Y. et al. Soft information acceleration aided subspace suppression MIMO detection. J Wireless Com Network 2024, 51 (2024). https://doi.org/10.1186/s13638024023815
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13638024023815