As will be seen below, the maximum- likelihood sequence detector in fading channels typically has prohibitive computational complexity. The EM algorithm is an iterative technique for solving complex maximum-likelihood estimation problems. In this section we discuss sequence detection in fading channels based on the EM algorithm. Both a batch algorithm and a sequential algorithm are discussed. 9.3.1 Expectation-Maximization AlgorithmSuppose that q is a set of parameters to be estimated from some observed data Y . The maximum-likelihood estimate of q is given by Equation 9.15
where p ( Y q ) denotes the probability density of Y with q fixed. In many cases, an explicit expression for the conditional density p ( Y q ) does not exist. In other cases, the maximization problem above is very difficult to solve, even though the conditional density can be expressed explicitly. The EM algorithm [248, 315] is an iterative procedure for solving the ML estimation problem above in many such situations. In the EM algorithm, the observation Y is termed incomplete data . The algorithm postulates that one has access to complete data X , which is such that Y can be obtained through a many-to-one mapping. Typically, the complete data is chosen such that the conditional density p ( X q ) is easy to obtain and maximize over q . Starting from some initial estimate q (0) , the EM algorithm solves the ML estimation problem (9.15) by the following iterative procedure:
It is known that the sequence { q ( j ) } obtained in the EM algorithm above monotonically increases the incomplete-data likelihood function: Equation 9.18
Moreover, if the function Q ( q ; q ') is continuous in both q and q ', all limit points of an EM sequence { q ( j ) } are stationary points of p ( Y q ) (i.e., local maxima or saddle points) and p ( Y q ( j ) ) converges monotonically to p for some stationary point [248, 315]. 9.3.2 EM-Based Receiver in Flat-Fading ChannelsWe consider the following discrete-time flat-fading channel Equation 9.19
where { a [ i ]} is the complex Gaussian fading process, { b [ i ]} is a sequence of transmitted phase-shift-keying (PSK) symbols ( b [ i ] = 1), and { n [ i ]} is a sequence of i.i.d. Gaussian noise samples. Define the following notation:
Then (9.19) can be written as Equation 9.20
Note that both a and n are complex Gaussian vectors: Equation 9.21
Equation 9.22
where E s is the average received signal energy. For mobile fading channels, the normalized M x M autocorrelation matrix has elements given by the Jakes model as Equation 9.23
where B d T is the symbol-rate-normalized Doppler shift and J () is the Bessel function of the first kind and zeroth order. Hence r in (9.20) has the following complex Gaussian distribution: Equation 9.24
and the log-likelihood function of r given B is thus given by Equation 9.25
Note that Equation 9.26
Equation 9.27
where we have used the facts that BB H = B H B = I M and det( B ) det( B H ) = det( BB H ) = 1, since B is a diagonal matrix containing PSK symbols. Hence the ML estimate of b becomes Equation 9.28
The optimal solution involves an exhaustive enumeration of all possible PSK sequences of length M , which is certainly prohibitively complex even for moderate M . The EM algorithm was applied to solve the fading channel detection problem above in [138]. To use the EM algorithm, we define the complete data as consisting of the incomplete data r together with the fading process a [i.e., x = ( r , a )]. Then the log-likelihood function of the complete data is Equation 9.29
Hence the E-step computes the following quantity: Equation 9.30
Since given b = b ( j ) , r and a are jointly Gaussian, we then have Equation 9.31
The maximization step becomes Equation 9.32
An initial estimate of the data symbol sequence b (0) can be obtained with the aid of pilot symbols as follows . Suppose we choose M such that M = NL + 1, where N and L are positive integers. Suppose further that the symbols in positions 0, N , 2 N ,..., ( L -1) N are known. Then the initial channel estimates at these positions are given by Equation 9.33
and the initial channel estimates at other positions are obtained by linear interpolation; that is, Equation 9.34
Substituting the initial channel estimate a (-1) above into (9.32), we obtain the initial symbol estimate b (0) . Finally, we summarize the EM-based pilot-symbol-aided receiver algorithm in flat-fading channels as follows. Algorithm 9.1: [EM algorithm for pilot-symbol-aided receiver in flat-fading channels]
9.3.3 Linear Multiuser Detection in Flat-Fading Synchronous CDMA ChannelsThe EM-based receiver discussed above can easily be applied to synchronous CDMA systems in flat-fading channels [555]. The basic idea is to use a linear multiuser detector (e.g., the decorrelating detector) to separate the multiuser signals and then to employ an EM receiver for each user to demodulate its data. This procedure is discussed briefly next . We consider the following simple K -user synchronous CDMA system signaling through flat-fading channels. The received signal during the i th symbol interval is given by Equation 9.35
where a k [ i ] is the complex fading gain of the k th user's channel during the i th symbol interval; A k is the amplitude of the k th user; b k [ i ] {+1, -1} is the i th bit of the k th user; { s k ( t ), 0 t T } is the unit-energy spreading waveform of the k th user; and n ( t ) is white complex Gaussian noise with power spectral density s 2 . The received signal is correlated with the signature waveform of each user, to obtain the decision statistic: Equation 9.36
where and . On denoting y [ i ] = [ y 1 [ i ] y 2 [ i ] ... y K [ i ]] T , we can write Equation 9.37
where R = [ r jk ], A = diag { A 1 ,..., A K }, F [ i ] = diag { a 1 [ i ], ..., a K [ i ]}, b [ i ] = [ b 1 [ i ] b 2 [ i ] ... b K [ i ]] T , and n [ i ] = [ n 1 [ i ] n 2 [ i ] ... n K [ i ]] T . Note that n [ i ] ~ N c ( , s 2 R ). The multiuser signals y [ i ] in (9.37) can be separated by a linear decorrelator, to obtain Equation 9.38
with u [ i ] ~ N c ( , s 2 R -1 ). We can write (9.38) in scalar form as Equation 9.39
with u k [ i ] ~ N c (0, s 2 [ R -1 ] kk ). We see that for each user, the output of the decorrelating detector (9.39) is of exactly the same form as (9.19). Hence with the aid of pilot symbols, the EM receiver discussed in Section 9.3.2 can be employed to demodulate the k th user's data { b k [ i ]} i , k = 1, ..., K . An alternative suboptimal receiver structure for demodulating the k th user's data uses a Kalman filter to track the fading channel { a k [ i ]} i , based on training symbols or decision feedback [547, 578, 579]. For example, in the simplest setting, the fading coefficients { a k [ i ]} i may be modeled by a second-order autoregressive (AR) process: Equation 9.40
where w [ i ] is a zero-mean white complex Gaussian process. The parameters a 1 and a 2 are chosen to fit the spectrum of the AR process to that of the underlying Rayleigh fading process. On the other hand, a statistically equivalent representation of the linear decorrelator output (9.39) is Equation 9.41
where we have invoked the symmetry of the distribution of u k [ i ]. Based on the state equation (9.40) and the observation equation (9.41), we can use a Kalman filter to track the fading channel coefficients { a k [ i ]} i and subsequentially detect the data symbols. Note that in (9.41), the data symbols { b k [ i ]} i are assumed known, corresponding to the case when they are training symbols. When these symbols are unknown, they are replaced by the detected symbols . Such a decision-directed scheme is subject to error propagation, of course, and thus requires periodic insertion of training symbols. 9.3.4 Sequential EM AlgorithmThe EM algorithm discussed above is a batch algorithm. We next briefly describe a sequential version of the EM algorithm [482, 563]. Suppose that y 1 , y 2 , ... is a sequence of observations with marginal pdf f ( y q ), where q C m is a static parameter vector. A class of sequential estimators derived from the maximum-likelihood principle is given by Equation 9.42
where q ( i ) is the estimate of q at the i th step, P ( y i +1 , q ( i ) ) is an m x m matrix defined later in this section, and Equation 9.43
is the update score (i.e., the gradient of the log-likelihood function). Let H ( y i , q ( i ) ) denote the Hessian matrix of log f ( y i q ( i ) ): Equation 9.44
Let x i denote a "complete" data set related to y i for i = 1,2, .... The complete data set x i is selected in the (sequential) EM algorithm such that y i can be obtained through a many-to-one mapping x i y i , and so that its knowledge makes the estimation problem easy [e.g., the conditional density f ( x i q ) can easily be obtained]. Denote the Fisher information matrices of the data y i and x i , respectively, as
Different versions of sequential estimation algorithms are characterized by different choices of the function P ( y i +1, q ( i ) ) in (9.42), as follows.
|