Skip to main content

Variable-step-size based sparse adaptive filtering algorithm for channel estimation in broadband wireless communication systems

Abstract

Sparse channels exist in many broadband wireless communication systems. To exploit the channel sparsity, invariable step-size zero-attracting normalized least mean square (ISS-ZA-NLMS) algorithm was applied in adaptive sparse channel estimation (ASCE). However, ISS-ZA-NLMS cannot achieve a good trade-off between the convergence rate, the computational cost, and the performance. In this paper, we propose a variable step-size ZA-NLMS (VSS-ZA-NLMS) algorithm to improve the ASCE. The performance of the proposed method is theoretically analyzed and verified by numerical simulations in terms of mean square deviation (MSD) and bit error rate (BER) metrics.

1 Introduction

Broadband transmission is one of the key techniques in wireless communication systems [13]. To realize reliable broadband communication, one challenge is accurate channel estimation in order to mitigate inter-symbol interference (ISI). Conventional normalized least mean square (ISS-NLMS) algorithm using invariable step size was considered as one of the effective methods for channel estimation due to its easy implementation [4]. However, ISS-NLMS does not take the channel characteristic into consideration and cannot take the advantage of the inherent channel prior information. During the last few years, more and more channel measurements have validated and indicated that broadband channels are most likely to have sparse or cluster-sparse structures [57], as shown in Figure 1 as an example. In particular, channel sparsity in different mobile communication systems are summarized in Table 1. Inspired by least absolute shrinkage and selection operator (LASSO) algorithm [8], an 1-norm sparse constraint function can be used to take the advantage of channel sparsity in adaptive sparse channel estimation (ASCE); zero-attracting ISS-NLMS (ZA-ISS-NLMS) has been proposed for ASCE [9, 10] to improve the estimation performance.

Figure 1
figure 1

Two kinds of channel structures: dense and sparse.

Table 1 Channel structures in different mobile communication systems

It is well known that step size is a critical parameter which determines the estimation performance, convergence rate, and computational cost. However, ISS-NLMS and ZA-ISS-NLMS adopt a fixed step size, and as a result, they are unable to achieve a good balance between steady-state estimation performance and convergence speed. Different from ISS-NLMS [4], variable step-size NLMS (VSS-NLMS) was first proposed to improve the estimation performance [11] without sacrificing the convergence speed. Variable step size is controlled by the instantaneous square error of each iteration, i.e., lower error will decrease the step size and vice versa. To the best of our knowledge, the application of sparse VSS-NLMS to simultaneously exploit the channel sparsity and control the step size has not been reported in the literature.

In this paper, we propose a zero-attracting VSS-NLMS (ZA-VSS-NLMS) algorithm for sparse channel estimation. The main contribution of this paper is to propose the ZA-VSS-NLMS using VSS rather than ISS for estimating spare channels. In addition, the step size of the proposed algorithm is updated in each iteration according to the error information. In the following, conventional ZA-ISS-NLMS is introduced and its drawback is analyzed at first. ZA-VSS-NLMS is then proposed using an adaptive step size to achieve a lower steady-state estimation error. To derive the adaptive step size, different from the traditional VSS-NLMS algorithm in [11], two practical problems are considered: sparse channel model and tractable independent assumptions [12]. At last, numerical simulations are carried out to evaluate the proposed algorithm in terms of two metrics: mean square deviation (MSD) and bit error rate (BER).

The remainder of this paper is organized as follows. A system model is described and ZA-ISS-NLMS algorithm is introduced in Section 2. In Section 3, ZA-VSS-NLMS algorithm is proposed. Numerical results are presented in Section 4 to evaluate the performance of the proposed ASCE method. Finally, we conclude the paper in Section 5.

2 ZA-ISS-NLMS algorithm

Consider a frequency-selective fading wireless communication system where FIR sparse channel vector h = [h0, h1, …, hN −1]T has length N and it is supported only by K nonzero channel taps. Assume that an input training signal x(t) is used to probe the unknown sparse channel. At receiver, equivalent-baseband observed signal y(t) at time t is given by:

y t = h T x t +z t ,
(1)

where x(t) = [x(t), x(t −1), …, x(t − N +1)]T denotes the vector of training signal x(t); z(t) is the additive white Gaussian noise (AWGN), which is assumed to be independent to x(t); ()T denotes the vector transpose operation. The objective of ASCE is to adaptively estimate the unknown sparse channel vector h using the training signal vector x(t) and the observed signal y(t). According to Equation 1, instantaneous error e(n) is defined as,

e n = y t y n = h h ˜ n T x t + z t = v T n x t + z t ,
(2)

where v n =h h ˜ n denotes the channel estimation error in the n-th iteration. In the sequel, one can apply ZA-ISS-LMS algorithm to exploit channel sparsity in time domain. First of all, cost function of ZA-ISS-LMS is given by:

G n = 1 2 e 2 n +λ h ˜ n 1 ,
(3)

where λ is the regularization parameter to balance the updating square error e2(n) and sparse penalty of the n-th updated channel estimator h ˜ n ; 1 denotes 1-norm operation, e.g., h 1 = l = 0 N 1 h l . The update equation of ZA-ISS-LMS at time t is:

h ˜ n + 1 = h ˜ n μ G n h ˜ n = h ˜ n + μ e n x t ρ sgn h ˜ n ,
(4)

where μ is the ISS which determines the convergence speed; ρ = μλ is a parameter which depends on the step-size μ and the regularization parameter λ; and sgn() is a component-wise function which is defined by:

sgn h = 1 , h > 0 0 , h = 0 1 , h < 0 .
(5)

Observing the update Equation 4, its second term attracts small-value channel coefficients to zero in high probability. In other words, most of the small-value channel coefficients can be replaced by zero. This will speed up the convergence and mitigate the noise on zero positions as well. However, the performance of the ZA-ISS-LMS is often degraded by random scaling of training signals. To avoid the randomness as well as to improve the estimation performance, we proposed an improved algorithm (i.e., ZA-ISS-NLMS) in our previous works in [9] and [10]. The update equation of ZA-ISS-NLMS [9] was proposed as follows

h ˜ n + 1 = h ˜ n +μ e n x t x T t x t ρsgn h ˜ n .
(6)

The ZA-ISS-NLMS algorithm in Equation 6 adopts one step size and its convergence speed is fixed as shown in Figure 2a. As a result, one drawback of ZA-ISS-NLMS is the lack of ability to trade off between the estimation performance and convergence speed.

Figure 2
figure 2

Illustrations of gradient descend and zero attracting of (a) ZA-ISS-NLMS and (b) ZA-VSS-NLMS.

3 Proposed algorithm

Recall that the ZA-ISS-NLMS algorithm in Equation 6 does not utilize VSS. It is well known that the step size is a critical parameter which determines the estimation performance, convergence speed, and computational cost. Inspirited by the VSS-NLMS algorithm in [11], VSS is introduced to make the step size adaptive to the estimation error to further improve the estimation performance. Based on the previous research [10] and [11], ZA-VSS-NLMS algorithm has the following update equation,

h ˜ n + 1 = h ˜ n +μ n + 1 e n x t x T t x t ρsgn h ˜ n ,
(7)

where μ(n +1) is the VSS which is calculated from the estimation error and the variance of the additive noise. Comparing Equation 7 with Equation 4, it can be found that the step size is different, i.e., step size in Equation 4 is invariant while step size in Equation 7 is adaptively variant. There are two facts about μ(n) and ρ that should be noticed: 1) the variant step-size μ(n) is adopted to speed up the convergence speed in the case of large estimation error, while to ensure the stability in the case of small estimation error; 2) the parameter ρ, which depends on the initial step-size μ and regularization parameter λ, is utilized to exploit channel sparsity effectively. Otherwise, variant parameter ρ(n) = μ(n)λ may cause extracomputational complexity and ineffectiveness use of the channel sparsity.

Optimal step-size μ o (n +1) for the (n +1)-th iteration is derived based on the following assumptions:

  1. (A1):

    Input vector x(t) and the additive noise z(t) are mutually independent at time t.

  2. (A2):

    Input vector x(t) is a stationary sequence of independent zero mean Gaussian random variables with a finite variance s x 2 .

  3. (A3):

    z(t) is an independent zero mean random variables with variance s z 2 .

  4. (A4):

    h ˜ n is independent of x(t).

These assumptions provide a mathematically tractable analysis in the subsequent proposed algorithm. The proposed algorithm in Equation 7 can be rewritten in terms of the estimation error vector v(n) as follows:

v n + 1 =v n μ n + 1 e n x t x T t x t +ρsgn h ˜ n ,
(8)

Taking the expectation on the MSD of h ˜ n , it can be written as:

E v n + 1 2 2 = E v n 2 2 + μ 2 n + 1 E e 2 n / x T t x t + ρ 2 E sgn h ˜ n T sgn h ˜ n 2 μ n + 1 E e n v T n x t / x T t x t + 2 ρ E v T n sgn h ˜ n 2 μ n + 1 ρ E e n x T t sgn h ˜ n / x T t x t D 0 D μ ,
(9)

Based on the assumptions (A1)-(A4), we can get the following results:

E v T n sgn h ˜ n =0,
(10)
E e n x T t sgn h ˜ n / x T t x t =0,
(11)
D 0 =E v n 2 2 + ρ 2 E sgn h ˜ n T sgn h ˜ n ,
(12)
D μ = 2 μ n + 1 E e n v T n x t / x T t x t μ 2 n + 1 E e 2 n / x T t x t .
(13)

According to Equation 9, MSD depends on parameters μ and ρ. However, the optimal value of ρ cannot be directly obtained since it is determined by channel sparsity and the additive noise. In order to find the optimal step-size μ o (n +1), empirical parameter ρ is used to make a fair comparison with the traditional method in Equation 6. When ρ is fixed in Equation 7, finding μ(n +1) becomes a convex problem so that it can maximize D(μ(n +1)), given by

μ o n + 1 =arg max μ o n + 1 D μ n + 1 .
(14)

In other words, to find the optimal step-size μ o (n +1) is equivalent to finding the largest gradient descend from the n-th iteration to the (n +1)-th iteration. By solving the convex problem in Equation 14, the (n +1)-th optimal step-size μ o (n +1) is obtained by:

μ o n + 1 = E e n v T n x t / x T t x t E e 2 n v T n x t / x T t x t = E p o T n + 1 p n + 1 E p o T n + 1 p o n + 1 + σ z 2 T r E 1 / x T t x t ,
(15)

where p o (n +1) x(t)xT(t)v(n)/xT(t)x(t). Obviously, the optimal step size is determined by p(n +1) and the noise variance σ z 2 . Unfortunately however, the optimal vector p o (n +1) depends on the unknown channel vector h and it is not available during adaptive updating process. Based on the assumption (A1), it can be found that:

E x t e n / x T t x t = E { x t ( x T t v n + z t ) / x T t x t } = E x t x T t v n / x T t x t .
(16)

According to Equation 16, an alternative approximate vector p(n +1) by time averaging is given as follows,

p n + 1 =βp n + 1 β x t e n x T t x t ,
(17)

where β [0, 1) is the smoothing factor to control the value of VSS and the estimation error. Note that the VSS will reduce to ISS when β = 0. Therefore, approximate step-size μ(n +1) for ZA-VSS-NLMS is given by:

μ n + 1 = μ max p T n + 1 p n + 1 p T n + 1 p n + 1 + C ,
(18)

where C is a positive threshold parameter satisfying CO 1 / S N R , where SNR is the received signal noise ratio (SNR). To better understand the proposed algorithm in Equation 7, Figure 2 is used to illustrate the two functions: zero attracting (for sparse constraint) and VSS (for convergence speed). According to Equation 18, the range of VSS is given by μ(n +1) (0, μmax), where μmax is the maximal step size. To ensure the stability of the adaptive algorithm, the maximal step size is usually set to be less than 2 [4]. Based on Equation 18, step-size μ for ZA-ISS-NLMS is invariable but the step-size μ(n +1) for ZA-VSS-NLMS is variable as depicted in Figure 3, where the maximal step-size μmax and step-size μ are set as μ = μmax {0.5, 1, 1.5}. From this figure, it can be found that the value of VSS μ (n +1) will decrease as the estimation error decreases and vice versa; on the other hand, ISS is invariant. Specifically, in the case of small step size, high performance can be achieved since small step size ensures the stability of the algorithm; while in the case of large step size, low computation complexity can be achieved since large step size increases the convergence speed. That is to say, as the updating error decreases, ZA-VSS-NLMS reduces its step size adaptively to ensure the algorithm stability as well as to achieve better steady-state estimation performance.

Figure 3
figure 3

ISS is invariable but VSS is variable in different estimation errors.

4 Numerical simulations

To testify the effectiveness of the proposed method, two metrics are adopted, i.e., MSD and BER. Channel estimators h ˜ n are evaluated by the average MSD which is defined as:

AverageMSD h ˜ n =E h h ˜ n 2 2 ,
(19)

where h and h ˜ n are the channel vector and its n-th iterative adaptive channel estimator, respectively. ·2 is the Euclidean norm operator and h 2 2 = i = 1 N h i 2 . System performance is evaluated in terms of BER which adopts different data modulation schemes. The results are averaged over 1,000 independent Monte Carlo (MC) runs. The length of channel vector h is set to be N = 64 and its number of dominant taps is set to be K = 2 and 6, respectively. Each dominant channel tap follows random Gaussian distribution CN 0 , σ h 2 and subjects to a total power constraint E h 2 2 = 1 and positions randomly within the length of h. The received signal-to-noise ratio (SNR) is defined as P 0 / σ n 2 , where P0 is the received power of the pseudo-random noise (PN) sequence for training. Numerical simulation parameters are listed in Table 2.

Table 2 Simulation parameters

Average MSD performance of the proposed method is evaluated at first. K = 2 and 6 are used and the results are shown in Figures 4, 5, 6, 7, 8, and 9 under three SNR regimes, i.e. 5, 10, and 20 dB. The proposed algorithm, ZA-ISS-NLMS, is compared with three existing methods, i.e., ISS-NLMS [4], VSS-NLMS [11], and ZA-ISS-NLMS [9, 10]. It can be observed from Figures 4, 5, 6, 7, 8, and 9 that ZA-VSS-NLMS achieved both faster convergence speed and better MSD performance than ZA-ISS-NLMS. The reason is that VSS-based gradient descend of the proposed algorithm makes a good trade-off between the convergence speed and the MSD performance. In addition, to achieve better steady-state estimation performance, regularization parameter methods for ZA-NLMS-type algorithms are adopted [13, 14] and set to be ρ=0.0015 σ n 2 . In different SNR regimes, ZA-VSS-NLMS always achieves a better estimation performance than ZA-ISS-NLMS. Furthermore, since ZA-VSS-NLMS takes the advantage of the channel sparsity as well, it obtains a better estimation performance than VSS-NLMS, especially in the extreme sparse channel case (e.g., K = 2).

Figure 4
figure 4

Average MSD performance versus received iterative times ( n ).

Figure 5
figure 5

Average MSD performance versus received iterative times ( n ).

Figure 6
figure 6

Average MSD performance versus received iterative times ( n ).

Figure 7
figure 7

Average MSD performance versus received iterative times ( n ).

Figure 8
figure 8

Average MSD performance versus received iterative times ( n ).

Figure 9
figure 9

Average MSD performance versus received iterative times ( n ).

In the next, BER performance using the proposed channel estimator is evaluated. The channel is assumed to be a steady-state sparse channel with number of nonzero taps K = 2 and SNR = 5 dB. Received SNR is defined by E s /N0, where E s is the received signal power and N0 is the noise power. Numerical result is shown in Figures 10 and 11. In Figure 10, multilevel phase shift keying (PSK) modulation, i.e., 8 and 16 PSK are used for data modulation. In Figure 11, multilevel quadrature amplitude modulation (QAM), i.e., 16 and 64 QAM, are used for data modulation. It is observed that the proposed algorithm can achieve a much better BER performance than ISS-NLMS and VSS-NLMS. Although there is no significant performance gain between our proposed algorithm and ISS-ZA-NLMS, fast convergence rate can be achieved by the proposed algorithm.

Figure 10
figure 10

Average BER performance versus received SNR (PSK).

Figure 11
figure 11

Average BER performance versus received SNR (QAM).

Therefore, it has been confirmed that the proposed algorithm can achieve the advantages of good performance and fast convergence speed.

5 Conclusions

Step size is a key parameter for NLMS-based adaptive filtering algorithms to balance the steady-state estimation performance and convergence speed. Either ISS-NLMS or ZA-ISS-NLMS cannot update their step size in the process of adaptive error updating. In this paper, a ZA-VSS-NLMS filtering algorithm was proposed for channel estimation. Unlike the traditional algorithms, the proposed algorithm utilizes VSS which can update the step size adaptively according to the updating error. Therefore, the proposed method can achieve a better steady-state performance while keeping a comparable convergence speed when compared with the existing methods. Simulation results have been presented to confirm the effectiveness of the proposed method in terms of MSD and BER metrics.

References

  1. Adachi F, Tomeba H, Takeda K: Introduction of frequency-domain signal processing to broadband single-carrier transmissions in a wireless channel. IEICE Trans. Commun. 2009, E92-B(no. 9):2789-2808. 10.1587/transcom.E92.B.2789

    Article  Google Scholar 

  2. Adachi F, Kudoh E: New direction of broadband wireless technology. Wirel. Commun. Mob. Comput. 2007, 7(no. 8):969-983. 10.1002/wcm.507

    Article  Google Scholar 

  3. Dai L, Wang Z, Yang Z: Next-generation digital television terrestrial broadcasting systems: Key technologies and research trends. IEEE Commun. Mag. 2012, 50(6):150-158.

    Article  Google Scholar 

  4. Widrow B, Stearns D: Adaptive Signal Processing. Prentice Hall, New Jersey; 1985.

    MATH  Google Scholar 

  5. Herdin M, Bonek E, Member S, Fleury BH, Czink N, Yin X, Ozcelik H: Cluster characteristics in a MIMO indoor propagation environment. IEEE Trans. Wirel. Commun. 2007, 6(no. 4):1465-1475.

    Article  Google Scholar 

  6. Wyne S, Czink N, Karedal J, Almers P, Tufvesson F, Molisch A: A Cluster-Based Analysis of Outdoor-to-Indoor Office MIMO Measurements at 5.2 GHz,” IEEE 64th Vehicular Technology Conference (VTC-Fall). Montreal, Canada; 2006:1-5. doi:10.1109/VTCF.2006.15

    Google Scholar 

  7. Vuokko L, Kolmonen V-M, Salo J, Vainikainen P: Measurement of large-scale cluster power characteristics for Geometric channel models. IEEE Trans. Antennas Propag. 2007, 55(no. 11):3361-3365.

    Article  Google Scholar 

  8. Tibshirani R: Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. 1996, 58(no. 1):267-288.

    MATH  MathSciNet  Google Scholar 

  9. Gui G, Peng W, Adachi F: Improved adaptive sparse channel estimation based on the least mean square algorithm. IEEE Wireless Communications and Networking Conference (WCNC), Shanghai, China 2013, 3130-3134.

    Google Scholar 

  10. Gui G, Adachi F: Improved adaptive sparse channel estimation using least mean square algorithm. EURASIP J. Wirel. Commun. Netw. 2013, 2013(no. 1):1-18. 10.1186/1687-1499-2013-1

    Article  Google Scholar 

  11. Shin H, Sayed AH, Song W: Variable step-size NLMS and affine projection algorithms. IEEE Signal Process. Lett. 2004, 11(no. 2):132-135. 10.1109/LSP.2003.821722

    Article  Google Scholar 

  12. Eweda E, Bershad NJ: Stochastic analysis of a stable normalized least mean fourth algorithm for adaptive noise canceling with a white Gaussian reference. IEEE Trans. Signal Process. 2012, 60(12):6235-6244.

    MathSciNet  Article  Google Scholar 

  13. Huang Z, Gui G, Huang A, Xiang D, Adachi F: Regularization selection methods for LMS-Type sparse multipath channel estimation. The 19th Asia-Pacific Conference on Communications (APCC), Bali Island, Indonesia 2013, 1-5.

    Google Scholar 

  14. Gui G, Mehbodniya A, Adachi F: Least mean square/fourth algorithm for adaptive sparse channel estimation. IEEE International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), London, UK 2013, 1-5.

    Google Scholar 

Download references

Acknowledgements

The authors would like to extend their appreciation to the anonymous reviewers for their constructive comments. This work was supported in part by the Japan Society for the Promotion of Science (JSPS) research activity start-up research grant (No. 26889050), Akita Prefectural University start-up research grant, as well as the National Natural Science Foundation of China under grants (Nos. 61401069, 61261048, and 61201273).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guan Gui.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Gui, G., Peng, W., Xu, L. et al. Variable-step-size based sparse adaptive filtering algorithm for channel estimation in broadband wireless communication systems. J Wireless Com Network 2014, 195 (2014). https://doi.org/10.1186/1687-1499-2014-195

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1687-1499-2014-195

Keywords

  • Sparse channel
  • ZA-NLMS
  • Invariable step size
  • Variable step size
  • ASCE