Session: SPTM-P9
Time: 1:00 - 3:00, Friday, May 11, 2001
Location: Exhibit Hall Area 1
Title: Performance Analysis and Design of Adaptive Filters
Chair: Behrouz Farhang-Boroujeny

1:00, SPTM-P9.1
STOCHASTIC ANALYSIS OF THE AFFINE PROJECTION ALGORITHM FOR GAUSSIAN AUTOREGRESSIVE INPUTS
N. BERSHAD, D. LINEBARGER, S. MCLAUGHLIN
ABSTRACT This paper studies the statistical behavior of the Affine Projection (AP) algorithm for mu = 1 for Gaussian Autoregressive inputs. This work extends the theoretical results of Rupp [3] to the numerical evaluation of the MSE learning curves for the adaptive AP weights. The MSE learning behavior of the AP(P+1) algorithm with an AR(Q) input (Q < P) is shown to be the same as the NLMS algorithm (mu = 1) with a white input with M-P unity eigenvalues and P zero eigenvalues and increased observation noise. Monte Carlo simulations are presented which support the theoretical results.

1:00, SPTM-P9.2
ANALYSIS OF THE ADAPTIVE FILTER ALGORITHM FOR FEEDBACK-TYPE ACTIVE NOISE CONTROL
H. SAKAI, S. MIYAGI
The feedback-type active noise control (ANC) system uses only one microphone to provide necessary signals for adjusting the adaptive filter. Due to the complicated nature of the whole adaptive filter structure there have been no theoretical results about its convergence properties. In this paper, first a stationary point of the adaptive filter using the filtered-X LMS algorithm is obtained by the averaging method combined with the frequency domain technique. Then the local convergence condition is derived. This is a counterpart of the well-known 90 degrees condition for the feedforward-type ANC. Finally, the convergence condition is explicitly given for a simple example and its validity is shown by some simulations.

1:00, SPTM-P9.3
EVALUATION AND DESIGN OF VARIABLE STEP SIZE ADAPTIVE ALGORITHMS
C. LOPES, J. BERMUDEZ
This paper presents a new methodology for evaluation and design of variable step size adaptive algorithms. The new methodology is based on a learning plane, which combines the evolutions of both the step size and the mean square error. It includes both transient and steady-state behaviors and can be used to compare performances of different algorithms against an optimum trajectory in the learning plane. The new technique can also be used for algorithm optimization in system identification applications.

1:00, SPTM-P9.4
CONVERGENCE ANALYSIS OF THE NLMS ALGORITHM WITH M-INDEPENDENT INPUTS
P. SCALART
In most adaptive identification applications, a finite impulse response (FIR) filter is employed with coefficients that are computed using the normalized least mean square (NLMS) algorithm. In this paper, the convergence behavior of the NLMS algorithm is analyzed using a simple model of the input signal vectors. Explicit expressions of the learning curve and misadjustement are derived and compared with those previously established for the NLMS algorithm. Comparisons between theoretical and experimental results are given to validate our approach.

1:00, SPTM-P9.5
LOCAL STABILITY ANALYSIS AND SYSTOLIC IMPLEMENTATION OF A SUBSPACE TRACKING ALGORITHM
F. XU, A. WILLSON
We discuss the DPASTd algorithm for signal subspace tracking. Our analysis shows that, under a sufficient condition on the step size, the DPASTd algorithm is locally stable even though delayed updating is applied. A pipelined realization of the algorithm and the corresponding systolic architecture are also proposed and a method for reciprocal computation is discussed. We also present simulation results to validate the algorithm.

1:00, SPTM-P9.6
MULTIPLE ENVIRONMENT OPTIMAL UPDATE PROFILING FOR STEEPEST DESCENT ALGORITHMS
M. MILISAVLJEVIC
In this paper, the methods for use of prior information about multiple operating environments, in improving adaptive filter convergence properties are discussed. More concretely, the gain selection, profiling and scheduling in steepest descent algorithms are treated in detail. Work presented in this paper is an extension of prior work in [1]. Two flavors of optimization are discussed: average descent rate optimization and maximization of the minimum descent rate. It is demonstrated, just as in the case of single channel optimization, with no additional complexity a substantial increase of convergence rate of steepest descent algorithms can be achieved. Finally, performance of the method is analyzed on the adaptive linear equalizer design for local area networks.

1:00, SPTM-P9.7
STABILITY ANALYSIS OF THE SEQUENTIAL PARTIAL UPDATE ALGORITHM
M. GODAVARTI, A. HERO III
Partial updating of LMS filter coefficients is an effective method for reducing the computational load and the power consumption in adaptive filter implementations. The Sequential Partial Update LMS algorithm is one popular algorithm in this category. In \cite{Douglas} a first order stability analysis of this algorithm was performed on wide sense stationary signals under the restrictive assumption of small step size parameter $\mu$. The necessary and sufficient condition derived on $\mu$ for convergence in the mean was identical to the one for guaranteeing stability in the mean of LMS. In \cite{Godavarti2} first order sufficient conditions were derived for stability without the aforementioned small $\mu$ assumption. The sufficient region of convergence derived was smaller than that of regular LMS. In this paper, we establish that for stationary signals the sequential algorithm converges in mean for the same values of the step size parameter $\mu$ for which the regular LMS does. In other words, we show that the conclusion drawn in \cite{Douglas} holds without the restrictive assumption of small $\mu$. We also derive sufficient conditions for stability on $\mu$ for cyclo-stationary signals.

1:00, SPTM-P9.8
ADAPTATION WITH CONSTANT GAINS: ANALYSIS FOR SLOW VARIATIONS
L. LINDBOM, M. STERNAD, A. AHLEN
Adaptation laws with constant gains, that adjust parameters of linear regression models, are investigated. The class of algorithms includes LMS as its simplest member. Closed-form expressions for the tracking MSE are obtained for parameters described by ARIMA processes. A key element of the analysis is that adaptation algorithms are expressed as linear time-invariant filters, here called learning filters, that work in open loop for slow parameter variations. Performance analysis can then easily be performed for slow variations, and stability is assured by stability of these learning filters. http://www.signal.uu.se/Publications/abstracts/r002.html

1:00, SPTM-P9.9
TRANSIENT ANALYSIS OF ADAPTIVE FILTERS
T. AL-NAFFOURI, A. SAYED
This paper develops a framework for the mean-square analysis of adaptive filters with general data and error nonlinearities. The approach relies on energy conservation arguments and is carried out without restrictions on the probability distribution of the input sequence. In particular, for adaptive filters with diagonal matrix nonlinearities, we provide closed form expressions for the steady state performance and necessary and sufficient conditions for stability. We carry out a similar study for long adaptive filters that employ error nonlinearities relying on a weaker form of the independence assumption. We provide expressions for the steady-state error and bounds on the step size for stability by exploiting the Cramer-Rao bound of the underlying estimation process.

1:00, SPTM-P9.10
MEAN-SQUARE ANALYSIS OF NORMALIZED LEAKY ADAPTIVE FILTERS
A. SAYED, T. AL-NAFFOURI
In this paper, we study leaky adaptive algorithms that employ a general scalar or matrix data nonlinearity. We perform a mean-square analysis of this class of algorithms without imposing restrictions on the probability distribution of the input. In particular, we derive conditions on the step-size for stability, and provide closed form expressions for the steady-state performance.

1:00, SPTM-P9.11
FILTER TRANSITIONS IN ADAPTIVE IIR APPROXIMATE FILTERING
R. ADHIKARY, S. NAWAB
The effects associated with the switching of filter orders in an incrementally adaptive IIR approximate filtering technique are evaluated in the context of speech signals. Through listening experiments, we have found that the perceptual quality of the speech at the filter output is sensitive to the initial conditions used in initiating a transition to a higher filter order. If zero initial conditions are used there is a significant crackling sound due to error bursts at points where switches to higher order filters take place. If output samples from the pre-transition filter are used as initial conditions for the post-transition filter, the amplitude of the error bursts are found to decrease significantly. An analysis is presented to account for these observations.

1:00, SPTM-P9.12
ITERATIVE WIENER DESIGN OF ADAPTATION LAWS WITH CONSTANT GAINS
M. STERNAD, A. AHLEN, L. LINDBOM
We present a method for optimizing adaptation laws that are generalizations of the LMS algorithm. Time-varying parameters of linear regression models are estimated in situations where the regressors are stationary or have slowly time-varying properties. The parameter variations are modeled as ARIMA-processes and the aim is to use such prior information to provide high performance filtering, prediction or fixed lag smoothing estimates for arbitrary lags. The method is based on a novel signal transformation that recasts the algorithm design problem into a Wiener design. http://www.signal.uu.se/Publications/abstracts/r001.html