7.9.1 Linear MMSE Detector and RLS Blind Adaptation RuleConsider the following received signal model: Equation 7.151
where A K , b K and s K denote, respectively, the received amplitude, data bit, and the spreading waveform of the K th user ; i denotes the NBI signal; and is the Gaussian noise. Assume that user 1 is the user of interest, and for convenience we will use the following notations: , and . The weight vector of the linear MMSE detector is given by Equation 7.152
where R r is the autocorrelation matrix of the received discrete signal r : Equation 7.153
The output SINR is given by Equation 7.154
where Equation 7.155
The mean output energy associated with w , defined as the mean-square output value of w applied to r , is Equation 7.156
where the last equality follows from (7.155) and the matrix inversion lemma. The mean-square error (MSE) at the output of w is Equation 7.157
The exponentially windowed RLS algorithm selects the weight vector w [ i ] to minimize the sum of exponentially weighted output energies:
where 0 < l < 1 is a forgetting factor (1 - l << 1). The purpose of l is to ensure that the data in the distant past will be forgotten in order to provide tracking capability in nonstationary environments. The solution to this constrained optimization problem is given by Equation 7.158
where Equation 7.159
A recursive procedure for updating w [ i ] is as follows: Equation 7.160
Equation 7.161
Equation 7.162
Equation 7.163
In what follows we provide a convergence analysis for the algorithm above. In this analysis, we make use of three approximations/assumptions: (a) For large i , R r [ i ] is approximated by its expected value [111, 301]; (b) the input data r [ i ] and the previous weight vector w [ i “1] are assumed to be independent [175]; (c) some fourth-order statistic can be approximated in terms of a second-order statistic [175]. 7.9.2 Convergence of the Mean Weight VectorWe start by deriving an explicit recursive relationship between w [ i ] and w [ i “1]. Denote Equation 7.164
Premultiplying both sides of (7.161) by s T , we have Equation 7.165
From (7.165) we obtain Equation 7.166
where Equation 7.167
Substituting (7.161) and (7.166) into (7.162), we can write Equation 7.168
where Equation 7.169
is the a priori least-squares estimate at time i . It is shown below that Equation 7.170
Equation 7.171
Substituting (7.161) and (7.170) into (7.168), we have Equation 7.172
Premultiplying both sides of (7.172) by R r [ i ], we get Equation 7.173
where we have used (7.159) and (7.169). Let q [ i ] be the weight error vector between the weight vector w [ i ] at time n and the optimal weight vector w : Equation 7.174
Then from (7.173) we can deduce that Equation 7.175
Therefore, Equation 7.176
where Equation 7.177
in which we have used (7.171) and (7.169). It has been shown [111, 301] that for large i , the inverse autocorrelation estimate behaves like a quasi-deterministic quantity when N (1 - l ) << 1. Therefore, for large i , we can replace by its expected value, which is given by [7, 111, 301] Equation 7.178
Using this approximation , we have Equation 7.179
Therefore, for large i , Equation 7.180
where we have used (7.170) and (7.179). For large i , R r [ i ] and R r [ i “1] can be assumed almost equal, and thus approximately [111, 301] Equation 7.181
Substituting (7.181) and (7.180) into (7.176), we then have Equation 7.182
Equation (7.182) is a recursive equation that the weight error vector q [ i ] satisfies for large i . In what follows we assume that the present input r [ i ] and the previous weight error q [ i “1] are independent. In this application of interference suppression, this assumption is satisfied when the interference signal consists of only MAI and white noise. If, in addition, there is NBI present, this assumption is not satisfied but is nevertheless assumed, as is the common practice in the analysis of adaptive algorithms [111, 175, 301]. Taking expectations on both sides of (7.182), we have
where we have used the facts that s T w = s T w [ i ] = 1, s T q [ i ] = s T w [ i ] “ s T w = 0 and Equation 7.183
Therefore, the expected weight error vector always converges to zero, and this convergence is independent of the eigenvalue distribution. Finally, we verify (7.170) and (7.171). Postmultiplying both sides of (7.163) by r [ i ], we have Equation 7.184
On the other hand, (7.160) can be rewritten as Equation 7.185
Equation (7.170) is obtained by comparing (7.184) and (7.185). Multiplying both sides of (7.166) by s T k [ i ], we can write Equation 7.186
and (7.167) can be rewritten as Equation 7.187
Equation (7.171) is obtained comparing (7.186) and (7.187). 7.9.3 Weight Error Correlation MatrixWe proceed to derive a recursive relationship for the time evolution of the correlation matrix of the weight error vector q [ i ], which is the key to analysis of the convergence of the MSE. Let K [ i ] be the weight error correlation matrix at time n . Taking the expectation of the outer product of the weight error vector q [ i ], we get Equation 7.188
We next compute the four expectations appearing on the right-hand side of (7.188). First term Equation 7.189
Equation 7.190
Equation 7.191
Equation 7.192
Equation 7.193
where in (7.189) we have used (7.183); in (7.193) we have used (7.152); in (7.190) and (7.192) we have used the fact that and in (7.191) we have used the following fact, which is derived below: Equation 7.194
Second term Equation 7.195
where we have used (7.183) and the following fact, which is shown below: Equation 7.196
Therefore, the second term is a transient term. Third term The third term is the transpose of the second term, and therefore it is also a transient term. Fourth term Equation 7.197
Equation 7.198
where in (7.198) we have used (7.152), and in (7.197) we have used the following fact, which is derived below: Equation 7.199
where is the mean output energy defined in (7.156). Now combining these four terms in (7.188), we obtain (for large i ) Equation 7.200
Finally, we derive (7.194), (7.196), and (7.199). Derivation of (7.194) We use the notation [ ·] mn to denote the ( m, n )th entry of a matrix and [ ·] k to denote the k th entry of a vector. Then Equation 7.201
Next we use the Gaussian moment factoring theorem to approximate the fourth-order moment introduced in (7.201). The Gaussian moment factoring theorem states that if z 1 , z 2 , z 3 , and z 4 , are four samples of a zero-mean, real Gaussian process, then [175] Equation 7.202
Using this approximation, we proceed with (7.201): Equation 7.203
Therefore,
where in the last equality we used (7.183) and the following fact: Equation 7.204
Derivation of (7.196) Similarly, we use the approximation by the Gaussian moment factoring formula and obtain
since E { q [ i ]} 0. Derivation of (7.199) Using the Gaussian moment factoring formula, we obtain
7.9.4 Convergence of MSENext we consider the convergence of the output MSE. Let denote the mean output energy at time i and [ i ] denote the MSE at time i : Equation 7.205
Equation 7.206
Since [ i ] and differ only by a constant P , we can focus on the behavior of the mean output energy : Equation 7.207
Since E { q [ i } , as i , the last term in (7.207) is a transient term. Therefore, for large , where is the average excess MSE at time i . We are interested in the asymptotic behavior of the excess MSE. Premultiplying both sides of (7.200) by R r and then taking the trace on both sides, we obtain Equation 7.208
Since l 2 + (1- l 2 ) < [ l + (1 - l )] 2 = 1, the term tr{ R r K [ i ]} converges. The steady-state excess mean-square error is then given by Equation 7.209
Again we see that the convergence of the MSE and the steady-state misadjustment are independent of the eigenvalue distribution of the data autocorrelation matrix, in contrast to the situation for the LMS version of the blind adaptive algorithm [183]. 7.9.5 Steady-State SINRWe now consider the steady-state output SINR of the RLS blind adaptive algorithm. At time i the mean output value is Equation 7.210
The variance of the output at time i is Equation 7.211
Let . Substituting (7.209) and (7.156) into (7.207), we get Equation 7.212
Therefore the steady-state SINR is given by Equation 7.213
where SINR * is the optimum SINR value given in (7.154). 7.9.6 Comparison with Training-Based RLS AlgorithmWe now compare the preceding results with the analogous results for the conventional RLS algorithms in which the data symbols b [ i ] are assumed to be known to the receiver. This condition can be achieved by using either a training sequence or decision feedback. In this case, the exponentially windowed RLS algorithm chooses w [ i ] to minimize the cost function Equation 7.214
The RLS adaptation rule in this case is given by [175] Equation 7.215
Equation 7.216
where e p [ i ] is the prediction error at time i and k [ i ] is the Kalman gain vector defined in (7.160). Using the results from [111], we conclude that the mean weight vector w [ i ] converges to w (i.e., E { w [ i ]} w , as i ), where w is the optimal linear MMSE solution: Equation 7.217
The MSE also converges, , as i , where * is the mean-square error of the optimum filter w , given by Equation 7.218
The steady-state excess mean-square error is given by [111] Equation 7.219
where we have used the approximation that , since 1 - l << 1 and N >> 1. Next we consider the steady-state output SINR of this adaptation rule in which the data symbols b [ i ] are known. At time i , the mean output value is Equation 7.220
where the last equality follows from (7.156). The output MSE at time i is Equation 7.221
Therefore, Equation 7.222
Using (7.220) and (7.222), after some manipulation, we have Equation 7.223
Therefore, the output SINR in the steady state is given by Equation 7.224
|