Top Banner
Robust Techniques for Signal Processing: A Survey In recent years there has been much interest in robustness issues in general and in robust signal processing schemes in particular. Robust schemes are useful in situations where imprecise a priori knowledge of input characteristics makes the sensitivity of perfor- mance to deviations from assumed conditions an important factor in the design of good signal processing schemes. In this survey we discuss the minimax approach for the design of robust methods for signal processing. This has proven to be a very useful approach because it leads to constructive procedures for designing robust schemes. Our emphasis is on the contributions which have been made in robust signal processing, although key results of other robust statistical procedures are also considered. Most of the results we survey have been obtained in the past fifteen years, although some interestingearlier ideas for minimax signal processing are also mentioned. This survey is organized into five main parts, which deal sep- arately with robust linear filters for signal estimation, robust linear filters for signal detection and related applications, nonlinear meth- ods for robust signal detection, nonlinear methods for robust esti- mation, and robustdata quantization. The interrelationships among many of these results are also discussed in the survey. I. INTRODUCTION In recent years there has been a resurgence of interest in what are known as robust methods for statistical signal processing. Such methods are applicable whenever schemes are used to carry out functions such as signal detection, estimation, filtering, and coding, common examples being in radar and sonar signal processing, communication sys- tems, pattern recognition, and speech and image processing. In the early days of development of the body of ideas we now possess for statistical signal processing, the emphasis Manuscript received October 2, 1984. This work was supported by the U.S. Air Force Office of Scientific Research under Grant AFOSR 82-0022, the U.S. Army Research Office under Contract DAAC29-81-K-0062, the Joint Services Electronics Program (U.S. Army, U.S. Navy, U.S. Air Force) under Contract NOOOl4-84-C-0149, theNational Science Foundation under Grant ECS-82-12080, and the U.S. Office of Naval Research under Contracts N00014-80-K-0945 and N00014-81-K-0014. S. A. Kassam is with the Department of Electricat Engineering, University of Pennsylvania, Philadelphia, PA 19104, USA, H. V. Poor is with the Department of Electrical and Computer Engineering and the Coordinated Science Laboratory, University of Illinois at Urbana-Champaign,Urbana, IL 61801, USA. was on the derivation of optimum schemes for use i n specified signal and noise environments. A classic example of this is the matched filter which is optimum for a particu- lar signal and noise model. Because the signals and noise in signal processing applications are usually modeled as ran- dom processes and performance measures therefore usually involve probabilistic quantities (such as mean squared error or probability of error), the theory of statistics has played a fundamental role in the development ofoptimum signal processing techniques. Suppose a signal processing scheme, say a detector for a signal with known waveform in additive noise, is designed to give optimum performance for noise possessing a specific statistical description. For example, one widespread model for noise is that it is a Gaussian process. An important question that arises is, how sensitive is the performance of such an optimum scheme to deviations in the signal and noise characteristics from those for which the scheme is designed? This is an important question because in practice one rarely has perfect knowledge of, say, the noise char- acteristics; the Gaussian or any other specific model is usually a nominal assumption which may at best be ap- proximately valid most of the time. Unfortunately, it turns out that in many cases nominally optimum signal processing schemes can suffer a drastic degradation in performance even for apparently small deviations from nominal assump- tions. It is this basic observation that motivates the search for robust signal processing techniques; that is, techniques with good performance under any nominal conditions and acceptable performance for signal and noise conditions other than the nominal which can range over the whole of allowable classes of possible characteristics. Thus in seek- ing robust methods it is recognized at the outset that a single, precise characterization of signal and noise condi- tions is unrealistic, and so classes of possible signal and noise characterizations are constructed and considered in the design of such methods. To illustrate the above observation with a concrete exam- ple consider further the detection problem mentioned above in discrete time. Thus suppose we have scalar ob- servations X,,X,;..,X,, forming a vector X, which are 0018-9219/85/0300-0433901.00 4985 IEEE PROCEEDINGS OF THE IEEE, VOL. 73. NO 3, MARCH 1985 43 3
50

Robust Techniques for Signal Processing - A Survey

Oct 28, 2014

Download

Documents

201001136
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Robust Techniques for Signal Processing - A Survey

Robust Techniques for Signal Processing: A Survey

In recent years there has been much interest in robustness issues in general and in robust signal processing schemes in particular. Robust schemes are useful in situations where imprecise a priori knowledge of input characteristics makes the sensitivity of perfor- mance to deviations from assumed conditions an important factor in the design of good signal processing schemes. In this survey we discuss the minimax approach for the design of robust methods for signal processing. This has proven to be a very useful approach because it leads to constructive procedures for designing robust schemes. Our emphasis is on the contributions which have been made in robust signal processing, although key results of other robust statistical procedures are also considered. Most of the results we survey have been obtained in the past fifteen years, although some interesting earlier ideas for minimax signal processing are also mentioned.

This survey is organized into five main parts, which deal sep- arately with robust linear filters for signal estimation, robust linear filters for signal detection and related applications, nonlinear meth- ods for robust signal detection, nonlinear methods for robust esti- mation, and robust data quantization. The interrelationships among many of these results are also discussed in the survey.

I . INTRODUCTION

In recent years there has been a resurgence of interest in what are known as robust methods for statistical signal processing. Such methods are applicable whenever schemes are used to carry out functions such as signal detection, estimation, filtering, and coding, common examples being in radar and sonar signal processing, communication sys- tems, pattern recognition, and speech and image processing.

In the early days of development of the body of ideas we now possess for statistical signal processing, the emphasis

Manuscript received October 2 , 1984. This work was supported by the U.S. Air Force Office of Scientific Research under Grant AFOSR 82-0022, the U.S. Army Research Office under Contract DAAC29-81-K-0062, the Joint Services Electronics Program (U.S. Army, U.S. Navy, U.S. Air Force) under Contract NOOOl4-84-C-0149, the National Science Foundation under Grant ECS-82-12080, and the U.S. Office of Naval Research under Contracts N00014-80-K-0945 and N00014-81-K-0014.

S. A. Kassam is with the Department of Electricat Engineering, University of Pennsylvania, Philadelphia, PA 19104, USA,

H. V. Poor i s with the Department of Electrical and Computer Engineering and the Coordinated Science Laboratory, University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA.

was on the derivation of optimum schemes for use in specified signal and noise environments. A classic example of this is the matched filter which is optimum for a particu- lar signal and noise model. Because the signals and noise in signal processing applications are usually modeled as ran- dom processes and performance measures therefore usually involve probabilistic quantities (such as mean squared error or probability of error), the theory of statistics has played a fundamental role in the development of optimum signal processing techniques.

Suppose a signal processing scheme, say a detector for a signal with known waveform in additive noise, is designed to give optimum performance for noise possessing a specific statistical description. For example, one widespread model for noise is that it is a Gaussian process. An important question that arises is, how sensitive is the performance of such an optimum scheme to deviations in the signal and noise characteristics from those for which the scheme is designed? This is an important question because in practice one rarely has perfect knowledge of, say, the noise char- acteristics; the Gaussian or any other specific model is usually a nominal assumption which may at best be ap- proximately valid most of the time. Unfortunately, it turns out that in many cases nominally optimum signal processing schemes can suffer a drastic degradation in performance even for apparently small deviations from nominal assump- tions. It is this basic observation that motivates the search for robust signal processing techniques; that is, techniques with good performance under any nominal conditions and acceptable performance for signal and noise conditions other than the nominal which can range over the whole of allowable classes of possible characteristics. Thus in seek- ing robust methods it is recognized at the outset that a single, precise characterization of signal and noise condi- tions is unrealistic, and so classes of possible signal and noise characterizations are constructed and considered in the design of such methods.

To illustrate the above observation with a concrete exam- ple consider further the detection problem mentioned above in discrete time. Thus suppose we have scalar ob- servations X , , X , ; . . , X , , forming a vector X , which are

0018-9219/85/0300-0433901.00 4 9 8 5 I E E E

P R O C E E D I N G S OF THE IEEE, VOL. 73. NO 3, M A R C H 1985 43 3

Page 2: Robust Techniques for Signal Processing - A Survey

known either to be noise only or to be noise plus a known signal sequence s,, s 2 ; -, s, with positive amplitude 8. We express this situation as a choice between the two hypothe- ses

/ - / , : X j = N j , i=1,2;..,n (I .I a)

and

H 1 : X j = B s j + N,, i=1,2;..,n ( I . Ib)

where the noise components Nj will be assumed to be independent and identically distributed with a common univariate probability density function (pdf) f . The likeli- hood ratio A(X) for the observation vector X i s given in this case by

This ratio can be formed for any particular realization of X provided f is known. It is well-known that a test for H, versus Hq based on the comparison of A(X) to a threshold is optimum according to several criteria. For example, such a test is Neyman-Pearson optimum [ I ] yielding maximum detection power (i.e., minimum “miss” probability) subject to a constraint on the maximum value of the false-alarm probability. Similarly, the test minimizing the Bayes risk for a set of prior probabilities for H, and H,, and the minimax test for a given loss function or payoff matrix with unknown priors are also of this form.

A test based on the comparison of A(X) with a threshold is equivalent to one based on a comparison of the loga- rithm of A(X) with the logarithm of the original threshold. Taking the logarithm on both sides of (1.2), we have

where

If I L ( x ; s,B)I is unbounded as a function of x , the value of logA(X) can be influenced heavily by a single observation component X i for which I L ( X , ; s,,e)l is large. Such a com- ponent can therefore completely override the weight of a possibly large number of other components in the choice between H, and HI. While this effect is certainly accept- able if the model for the noise density function is accurate, it may also be observed because of an occasional com- pletely erroneous measurement which the pdf model f does not take into account. In general, the assumed pdf f describes only an approximate or nominal model. Thus while the actual value of l L ( X , ; s , , O ) l at some observed value X , = x , may not be large relative to that obtained at other observation components, for the assumed model this may happen. To illustrate this, suppose that f is assumed to be Gaussian, in which case [ ( x ; s,e) is linear in x and thus is unbounded. If the true density f has exponential, rather than Gaussian, tails, then the true L ( x ; s,@) is a constant for x in the tails, and so is bounded. For a model specifying exponential tails, an increasing absolute value for an ob- servation component indicates increasing likelihood of one hypothesis over another only up to a ”saturation” value; beyond it, larger absolute values do not indicate larger relative likelihood. If the noise density were truly exponen-

tial (or some other long-tailed pdf), then the performance of the test that is optimum for Gaussian noise could be very poor because of the unexpected number of large noise values.

It would appear, then, that to counter the undesirable sensitivity of the test based on A(X) one should implement a bounded modification i( x; s,e) of the function L ( x; s, 0 ) corresponding to the assumed nominal model. Thus we are led to consider i( x; s, e) of the form

b, L ( x ; s J ) > b

( -a, q x ; s , e ) < - a i ( x ; s , e ) = L ( x ; s , e ) , - a < L ( x ; s , e ) G b (1.5)

where a and b are constants. One can expect that with a and b not too small test performance should degrade only marginally when the assumed model is accurate. On the other hand, the boundedness of i ( x ; s,e) builds in a robustness against the influence of a small number of spurious observations. The size of the interval [ - a, b ] clearly controls the tradeoff between degree of robustness and performance degradation under the assumed model. It is noteworthy that several analytical considerations of robust detection lead to detectors based on functions with the form (1.5), as will be discussed in Section IV.

Often a class of allowable characteristics, say for a noise pdf, is constructed by starting with a nominal characteristic and then including in the class all other characteristics that are “close,” in some well-defined sense, to this nominal one. Then a signal processing scheme that is robust may have performance at the nominal which is not quite as good as the scheme that is optimum for the nominal case, but i t s overall performance with respect to the defined class of characteristics will be good or acceptable. This loose definition of robustness is perfectly reasonable, but it does not provide a systematic approach to obtaining robust schemes. In order to achieve this we must first specify a measure of “overall” performance of a scheme with respect to a class of allowable conditions at the input. One such measure that has been widely used and which leads to interesting and useful results in many situations is the worst case performance of a scheme over a class of input condi- tions. Clearly, if its worst case performance is good we may say that a given scheme is robust. On the other hand, to find such a robust scheme we can look for the scheme that optimizes worst case Performance. This approach leads to what are known as minimax’ robust schemes. Implicit in our association of minimax schemes with robust schemes is the expectation that the worst case performance of a minimax scheme will be acceptably good, being the best that can be achieved. Another expectation one has in de- fining robust schemes in this way is that at any nominal operating point the performance of the minimax scheme will not be very far below that of the nominally optimum scheme, which on the other hand will have much poorer performance away from the nominal point. Fortunately, it does turn out that minimax schemes for the signal processing applications of interest usually have the above

‘A scheme that minimizes the maximum possible value of a loss function is minimax; if performance is measured by a gain function then a maximin scheme would be sought. We shall use the term minimax as a general description for such schemes in all cases.

434 PROCEEDINGS OF THE IEEE, V O L 73. NO. 3, MARCH 1985

Page 3: Robust Techniques for Signal Processing - A Survey

characteristics. They may therefore be said to have a more ”stable” performance than schemes lacking these character- istics (in the literature the terms robust and stable are sometimes used to mean the same thing).

We should emphasize that the classes of allowable char- acteristics one deals with in robust signal processing are generally nonparametric function classes, such as the class of all power spectral density functions with specified total power (area under the function) and which lie between specified upper and lower bounding functions. For uncer- tainties expressed by parametric classes of allowable values for finite-dimensional parameters (such as the mean and variance of a Gaussian pdf) one can of course use minimax designs as well, although alternative parametric approaches of statistical theory can also be applied in such situations.

In this paper we will concentrate on minimax robust schemes. There are useful formulations of robustness other than the minimax one, most notably the stability or qualita- tive robustness ideas introduced by Root [2] in the context of signal detection and by Hampel [3] in the context of parameter estimation. These formulations utilize the idea of robustness as a continuity property of some performance measure as a function of the underlying model, and some brief discussion of these ideas is included here. However, from the viewpoint of design, the minimax approach has had the most impact on robust signal processing schemes. ~ l s o we will not survey adaptive procedures, which may be used as robust schemes when input conditions are not precisely known and may be time-varying. Adaptive procedures, which attempt to learn about input conditions and adjust their specific signal processing structure accord- ingly to maintain good performance, are generally more complex than fixed minimax schemes. Adaptive schemes are more desirable when the a priori uncertainty is so large that the guaranteed level of performance of a minimax scheme would be too poor to be acceptable and when adequate time and data for adapting are available. Con- versely, minimax procedures would be more desirable un- der more constrained uncertainty classes, and especially as robust procedures to guard against excessive performance degradation of nominally optimum schemes for deviations from nominal assumptions. Minimax schemes may be used in conjunction with an adaptive approach, because the learning mechanism in adaptive schemes can never be expected to perform perfectly given any finite time for adaptation to take place. The application of minimax con- cepts to obtain robust versions of optimum adaptive proce- dures has also been considered [4]. While our primary concern here is with minimax robust schemes we will mention other specific techniques whenever it is ap- propriate.

Most of the recent investigations on robust signal processing techniques have been motivated by the works of the statistician Tukey [5] and more so by the seminal 1%4-1%5 results of the statistician Huber [6], [7] on mini- max robust location-parameter estimation and hypothesis testing. There has generally been a tendency to overlook some rather interesting work on minimax procedures which was carried out for signal processing applications in the decade prior to the publication of Huber’s results. In 1954, Zadeh [8] suggested that minimax solutions are the natural choices to use in filtering noisy signals under a priori uncertainties. In [9] Root describes the game-theoretic ap-

proach and its application to obtain minimax decision rules in some communication problems. (These results of Root were originally contained in a 1956 report [IO].) Early con- siderations of minimax schemes for signal processing in- clude those of Yovits and Jackson in 1955 [Il l on signal estimation filters for imprecisely known power spectral density functions and of Nilsson in 1959 [I21 and Zetterberg in 1%2 [I31 on matched filters. We will mention their results again in the following sections. Other early contribu- tions are the 1957 paper of Blachman [14], the 1959 work of Dobrushin [15], and the 1%1 paper of Cadzhiev [16].

The pre-I964 investigations of minimax signal processing schemes tend to be characterized by two attributes. One is that they were generally not concerned directly with pdf variations but rather with power spectral density function or related variations. Secondly, minimax schemes were ad- vocated simply as reasonable approaches when designing systems for operation under conditions at the inputs which could not be determined a priori. Thus the possible non- robustness of optimum schemes for nominal assumptions on the input was not explicitly recognized as an issue.

The term “robust” was first used in describing desirable statistical procedures by Box in 1953 [17]. As we have remarked, minimax robustness of estimation and hypothesis testing schemes was considered by Huber in [6] and [7], and since then a large number of results on minimax and alternative formulations of robustness have been generated in the statistics literature. In a recent paper [I81 Huber has given a most interesting account of some early concerns about robustness of statistical procedures and specific schemes, some of which date back to the last century. Reviews of the more recent techniques of robust statistics have been given by Huber [19], [20], Hampel [21], Bickel [22], and Hogg [23], [24]. Ershov [25] also gives a survey of robust estimation schemes which is quite broad in its scope. A monograph on robust estimation schemes by Andrews et a/. [26] studies the properties of many robust estimates. A collection of chapters edited by Launer and Wilkinson [27] contains some useful expositions. A recent book [28] may be consulted for a more detailed treatment.

This survey will focus specifically on minimax robust signal processing schemes, so that only a small part of the large body of the statistics literature will be mentioned explicitly. Most of the recent developments in robust signal processing have of course been influenced directly by the developments in robust statistics. However, signal proc- essing problems impose their own distinct requirements which are not always standard in problems of statistics. Thus it turns out that some recent developments in robust signal processing have provided new results in robust statis- tics.

Most of the results we survey here are of the post-1965 period. Two of the earliest papers in the signal processing area from this period are those of Wolff and Gastwirth [29] and Martin and Schwartz [30], and they have been responsi- ble for driving much of the subsequent work in robust signal processing. Thus a considerable literature has arisen on robust signal processing just in the last ten to fifteen years.

The statistical descriptions of input conditions in signal processing are usually stated in terms of power spectral density or correlation functions and pdfs. We shall discuss results which have been obtained on minimax robust linear

KASSAM AND POOR: ROBUST TECHNIQUES FOR SIGNAL PROCESSING 435

Page 4: Robust Techniques for Signal Processing - A Survey

filtering for signal estimation (e.g., Wiener filtering) in Section I 1 and for signal detection (eg, matched filtering) in Section Ill. Here the uncertainty classes are for spectral density or correlation functions. In Section IV results on minimax robust nonlinear signal detection schemes for distributional uncertainties are surveyed. Nonlinear parame fer estimation schemes are surveyed more briefly in Section V, since on this topic there is much already available in review form in the statistics literature. Also included in Section V is a brief survey of nonlinear modifications of the Kalman filter for robustness against non-Gaussian pdfs for the observation and process noise components. Section Vi treats the problem of robust quantization of data with unknown statistics, and we close with some concluding remarks in Section VII. Although our survey begins with robust linear filtering, studies on this topic are of more recent vintage than those on nonlinear signal detection and estimation. We feel, however, that the very widespread use of schemes such as Wiener and matched filters in signal processing justifies our beginning with robust versions of such linear processing schemes.

Before we begin let us note some other review, tutorial, and survey articles which are available in the literature. A tutorial on this subject by the authors has been published recently [31]. VandeLinde has given a brief survey in [32]. Ershov [25] and Krasnenker [33] have surveyed nonlinear estimation and detection schemes, respectively. Poor [34] has recently given a more mathematically detailed survey of robust detection schemes. Kleiner, Martin, and Thomson [35] and Martin and Thomson [36] treat the robust estima- tion of power spectral density functions. Robust methods for time series analysis have been considered by Martin in [37], and robust methods for system identification have been described by Poljak and Tsypkin in [38].

As a final introductory comment we should note that the literature in the area of robust statistical methods is vast and broad. Thus although this survey touches on what we feel to be the major contributions in robust signal processing, it is by no means exhaustive. However, the many results and methods that are not discussed here are accessible to the reader through the references provided.

1 1 . ROBUST FILTERS F O R SIGNAL ESTIMATION

One of the most common signal processing tasks arising in applications is that of estimating (e.g., filtering, predic- ting, or smoothing) a signal waveform from a noisy mea- surement. This task arises, for example, in radar and sonar tracking systems, in observers for automatic control sys- tems, in demodulators for analog communication systems, and in medical imaging systems.

Conventional design procedures for optimum signal estimation algorithms often require an exact knowledge of the statistical behavior both of the signal of interest and of the noise corrupting the measurement. For example, in the design of optimum linear estimation algorithms we must know the spectral or autocorrelation properties of the sig- nal and noise in order to specify the optimum procedures, and (as we shall see below) procedures designed to be optimum for a given model can be undesirably sensitive to inaccuracies in the model. As noted in the Introduction, robust procedures can overcome problems arising due to modeling inaccuracy by incorporating modeling uncertainty into the design from the outset.

In this section, we will discuss the design of robust

estimation procedures primarily within the context of the stationary linear (i.e., Wiener-Kolmogorov) estimation problem. Several other signal estimation problems have been treated in the context of robust design, including recursive nonlinear filtering and identification. Results on these problems will be discussed briefly in Section V.

A. The Need for Robustness in Signal Estimation

Consider the observation model

Y( t) = S( t) + N( t), -CO < t < 00 (2.1)

where { S( t); - 00 < t < oo} and { N( t); - oo < t < 00) are real, zero-mean, orthogonal, wide-sense-stationary (wss) random processes representing signal and noise, respec- tively. We assume that {S( t ) ; - 03 t < GO} and { N( t); - w < t < oo} have power spectral densities as and a,, respectively. (Most of these assumptions can be relaxed, as is discussed below.)

Given the observation process { Y( t ) ; - 03 < t < oo} we wish to form an estimate of S( t) of the form

j ( t ) = / m h ( t - T ) Y ( T ) d T -m

where h is the impulse response of a time-invariant linear filter. A common performance criterion for signal estimates is the mean squared error (MSE), which for estimates of the form of (2.2) is given straightforwardly by

where H is the transfer function associated with h (i.e., H i s the Fourier transform of h). If as and @, are known, then the M S E of (2.3) can be minimized over H to find the optimum filter transfer function for linear minimum-MSE estimation. It i s straightforward to show (see, e.g., Thomas [39]) that the minimizing solution is given by

and that the corresponding minimum value of the MSE is given by

ew(@s,@,.,). (2.5)

Suppose that we design a filter H, via (2.4) to be opti- mum for some nominal signal and noise spectral pair (@s,O,@,,O), but that the actual spectra as and @, can range over some classes 9 and N, respectively, of spectra “neighboring” QS,, and An important question that arises in this situation is what is the behavior of the MSE e(aS,@,; H,) as as and @, range over Y and A ” ? For example, it is of interest to know how the quantity

SUP e(@,,@,; H,) (2.6) ( @ 5 , @ N ) ~ ~ X ~

compares with the quantity @,,,; H , ) ew(@~,o,@,,o). The first of these quantities represents the worst performance of the filter H, over the class of possible spectra, whereas the second quantity represents the predic- ted performance assuming the nominal model to be accu-

436 P R O C E E D I N G S OF T H E IEEE, VOL. 73, NO. 3, M A R C H 1985

Page 5: Robust Techniques for Signal Processing - A Survey

rate. A situation in which (2.6) were considerably larger than ew(@s,o,@N,o) would point to a possibly inadequacy of the nominal design H,.

To illustrate the degree to which modeling uncertainty can affect performance, we consider the following example taken from Vastola and Poor [a]. Suppose we have as- sumed a nominal model (@s,o ,@N,o) given by

and

@,,o(o) = 20 4

loo + o 2 ' - w w < W . (2.7b)

Note that these spectra represent first-order wide-sense Markov processes with 3-dB signal bandwidth equal to 1, 3-dB noise bandwidth equal to IO, signal power E { [ S ( t ) ] 2 } = $, and noise power E{ [N( t ) ] ' } = 6. Suppose, how- ever, that all we really know are the total signal power, the total noise power, and the fractional signal and noise powers in the frequency band Io1 Q 1. This knowledge corresponds to the uncertainty classes2

Y= @ - @Jo) do = ps$ ( 51

and I/" QS(o) do = $ 297 i and

where

(2.8a)

(2.8b)

For a given estimation filter, H, and spectral pair (as,@,,,), the signal-to-noise ratio (SNR) at the output of H can be defined by

Output SNR = IOlog,, ( $ /e (QS,aN; H ) ) (2.9)

since the output j( t ) can be written as 3( t ) = S( t ) + (j( t ) - s(t)), and E{[S( t ) ] ' } = $ and

E { ( S ( t ) - j ( t ) ) ' } = e ( a s , Q N ; H ) .

Also, the input SNR is given simply by

Input SNR = lologl, ($14). Using these definitions, Fig. 1 depicts the nominal and worst case performance of the filter Ho designed to be optimum for the nominal spectral pair of (2.7). Note the considerable performance degradation throughout the given

'Note that rational models (such as (2.7)) are often forced upon estimated power spectra, although the actual data only predict fractional powers (such as (2.8)) accurately (Marzetta and Lang [41]).

Fig. 1. Worst case and nominal performance of nominal and trivial filtering for the example in Subsection Il-A.

range of input SNRs. Also depicted in Fig. 1 is the perfor- mance of trivial filtering, which corresponds to all-pass filtering if the input SNR is positive and no-pass filtering if the input S N R is negative. Note that the worst case over (2.8) of the nominally designed filter is uniformly worse than trivial filtering. Thus the nominal filter can actually make the signal noisier than it originally was!

8. Minimax Design of RQbust Filters

The above example illustrates the need for an alternative design philosophy for the stationary linear signal estimation problem for applications in which there is some uncertainty regarding the spectra of interest. In particular, in view of the methods described in the Introduction, we consider as a design philosophy the minimization over H of the worst case performance degradation described by (2.6).3 That is, we consider the design criterion

min( sup e ( a s , Q N ; H ) ) . (2.10)

A solution to (2.10) can be considered to be a robust filter for the uncertainty classes 9 and N ,

To solve this problem, we would like to find a saddle point for the minimax game of (2.10); i.e., we would like to find a spectral pair ( @ s , L , @ N , L ) E Y X N and a filter H, satisfying

H ( C p , , C p , ) € . Y X X

= min e(QS,,,@,,,; H ) . H

(2.11)

Note that the right-hand equality in (2.11) implies that H, is the optimum filter for the pair ( @ s S . L , @ N , L ) [i.e., H R ( o ) = @s,L(o)/(@s,L(o) + @N,L(o)) ] ; thus the determination of a saddle point involves finding a pair ( @ s , L , @ N , L ) which satisfies (2.11) with H, = @ s , L / ( @ s , L + @ N , L ) . The left-hand equality in (2.11) says that H, achieves its worst perfor- mance at the pair of spectra ( @ s , L , @ N , L ) for which it is optimum. This worst performance, e(QS,,,QN,,; HR), is the guaranteed level of performance of the filter H, for the classes of 9 and N .

The problem (2.10) was first posed4 in the context of robustness by Kassam and Lim in [44] wherein a saddle

3An approach to this problem in which the signal is taken to be deterministic with unknown parameters is described by Kurkin and Sidorov in [42].

4Although not in this specific context, the idea of designing filters for uncertain models by using least favorable conditions was pro- posed by Nahi and Weiss in [43].

KASSAM AND P O O R ROBUST TECHNIQUES F O R SIGNAL PROCESSING 43 7

Page 6: Robust Techniques for Signal Processing - A Survey

point solution to (2.10) was’given for the situation in which the spectra are known only to lie within given spectral bands. The problem of (2.10) was considered for general spectral uncertainty classes by Poor [45], and it is shown in [45] that for convex Y a n d X , a spectral pair ( @ s , L , @ N , L ) E Y X N a n d i ts optimum filter H, = @s,L/(@s, + @,,,) form a saddle point for (2.10) if, and only if, the pair ( @ s , L , Q N , L )

is least favorable for Y x X; i.e., if and only if ( @ s , L , @ N , L )

solves

where e, i s the minimum-MSE functional defined by (2.5). The term “least favorable” comes from the fact that ( @ s , L , @ N , L ) is the pair of spectra in Y X JV that correspond to the random processes that are hardest to separate by filtering.

Thus a design procedure for finding a robust filter for given uncertainty classes Y and X i s to solve (2.12) and then to design the optimum filter for the maximizing spec- tral pair. Since the filter design problem is solved by (2.4), the only possible difficulty is in solving (2.12). This problem, however, is straightforward to solve for many uncertainty classes of interest. in particular, the functional ew(QS,QN) can be written as

where C is the convex function C(x) = -(2n)-’x/(I + x). Thus maximizing e, is equivalent to minimizing the func- tional ]C(@s/@N)@N, which is a special case of a general class of divergences or ”distances” between densities (see Ali and Silvey [&I, Csiszar [471). In view of (2.13) least favorables can be interpreted as being the spectra in Y and N whose shapes are “closest together.” Because of this structure, the problem of solving for least favorable spectra for spectral uncertainty classes in which the total signal and noise powers are known and only the spectral shapes are uncertain can be accomplished by analogy with results in robust hypothesis testing. In particular, for a general type of classes with this property, the least favorable spectra are scaled versions of the least favorable probability densities for an analogous robust hypothesis testing problem posed by Huber [7]. This is a useful result because solutions to the robust hypothesis testing problem are known for many uncertainty models of interest. (See Poor [45], [48] for fur- ther details.)

C. Some Useful Models for Spectral Uncertainty

There are a number of useful models for spectral uncer- tainty for which solutions to the robust stationary linear filtering problem can be obtained straightforwardly. The following examples are typical:

Example 7: c-Contaminated Models: One very useful spectral uncertainty model is that given by

(@I@( a) = (1 - €)ao( 0) + €.( 0)

and /-w @(a) d o =/ Qo(o) do) (2.14)

where Qo is a nominal spectrum, u i s an arbitrary and unknown “contaminating” spectrum, and c i s a degree of

W

W - m

uncertainty (between 0 and 1) placed on the nominal model by the designer. This type of model allows for a fairly general type of uncertainty in a nominal spectral model, and it is used frequently to model uncertainty in several contexts.

Example 2: Variational-Neighborhood Models: Another useful model for spectral uncertainty arises from allowing all spectra that vary from a nominal spectrum by no more than some given amount. Using a standard measure of “variational distance” this model becomes

and ’/” @( o) do = $1 (2.15) 2n - m

where a0 and E play the same roles as in the c-con- taminated class and where d = (1/2n)j00,@0(w) do.

Example 3: pPoint Models: The classes of (2.8) are par- ticular examples of a more general type of spectral uncer- tainty class known as ppoint classes. These classes are of the form

(2.16)

where, as before, Qo is a nominal spectrum and 0, ,Q2; - .,a, form a partition of the frequency domain. Note that a ppo in t class consists of all spectra that have a fixed amount of power on each of the spectral regions 0,;. .,0,,. Such a class might arise, for example, when power measurements are taken in a number of frequency bands.

Example 4: Band Models: The first spectral uncertainty model that was considered in the context of robust Wiener filtering consists of the class of those spectral densities (with a given amount of power) that lie in a band bounded above and below by two known functions. This class can be written as

{ @ l L ( o ) < @(o) < u(o), --oo < w < 00

and ’/“ @( o) do = d } (2.17)

where L and U are known functions and where P is fixed. Note that a model such as (2.17) can be used to describe a “confidence region” around an estimated spectrum. Also note that the <-contaminated model of Example 7 is a special case of (2.17) with L = (1 - €)ao and U = 03.

Example 5: Generalized Moment-Constrained Models: As a final example, consider spectral uncertainty classes of the following type:

277

(@I/“ f , ( o ) @ ( o ) d o = c,, k=l ; - . ,n} (2.18) - m

where f,;. ., f, are known functions and C, ; - . ,C , are constants. The quantities If,@, k = l ; . - , n are sometimes known as “generalized moments” of the spectrum 0 corre- sponding to the weightings f,;.., f,. Note that the ppoint class of (2.16) is a special case of (2.18) with f k (w) = 1 for w E 0, and f k ( w ) = 0 for o 6i i l k . More generally, (2.18)

438 PROCEEDINGS OF THE IEEE. VOL. 73, NO. 3, MARCH 1985

Page 7: Robust Techniques for Signal Processing - A Survey

might represent the set of spectra of all processes that yield output power ck when input to a filter with transfer func- tion [2nfk(o)]'/* for k = 1;. ., n. Thus a model such as (2.18) arises when the available information consists of power measurements taken at the outputs of a filter bank.

If 9 and A" are both of the c-contaminated or variational neighborhood type, then it can be shown (see Kassam and Lim [44], Poor and Looze [49]) that the robust filter is of the form

if H,(o) < k' H R ( o ) = H o ( o ) , if k' c H,(o) d k" (2.19) { :::, if H,( w ) > k"

where H, = Os,o/(Os,o + QN,,) is the nominal filter and where k' and k" are two constants depending on the value of c and on the particular model used. This robust transfer function is illustrated in Fig. 2. Note that the effect of

Fig. 2. Typical robust filter characteristic for c-con- taminated or variational neighborhood models for spectral uncertainty.

incorporating the uncertainty into the design is a limiting of the gain of the nominal filter both from above and from below. This solution has a nice intuitive interpretation i f one considers the action of the nominal filter H,. This filter is designed to have near-unity gain in spectral regions where the nominal signal-to-noise power density ratio, O ~ , o ( w ) / O N , o ( w ) , i s large, and to have near zero gain at frequencies where this ratio is small. In other regions, the gain is chosen to balance the effects of signal distortion and noise throughput. The robust filter transfer function reflects similar characteristics but, also, because of the spectral uncertainty, it limits the gain from above to guard against a greater than nominal amount of noise power at the fre- quencies where OsS.o/ONN.O >> 1 and it limits the gain from below to assure that unexpected signal power at frequen- cies where Os,o/ON,O e 1 will not lead to undue distor- tion.

If both signal and noise spectral uncertainty classes are the ppo in t form of (2.16) with common spectral bands Q1; . . ,Q2 , , then the robust filter can be shown to be given by (see Cimini and Kassam [50], Vastola and Poor [51])

(2.20) where

'cfi, /!Frepilency, w

Q2

Fig. 3. Typical robust filter characteristic for ppoint mod- els for spectral uncertainty.

A typical filter of this type is depicted in Fig. 3. Note that this is a zonal filter that can only be implemented ap- proximately for temporal signals; however, in optical filter- ing where the variable t in (2.1) is interpreted as a spatial parameter and w as spatial frequency, this type of filter would be very simple to implement (see Cimini and Kassam [50], [52]). An interesting feature of this model is that the performance of the robust filter is constant over the classes Y and N a n d is given by [50]

for all (as, 0,) E 9' X A". The robust solution for the band model of Example 4 is

given by Kassam and Lim in [44] and its behavior is similar to that for the c-contaminated model (which is a special case). Solutions for generalized moment classes have been given by Breiman in [53], and a particular case of this model will be discussed below. (Of course, the ppoint model is also a special case.) Other models including combinations of the above models (such as the bounded ppoint model [54]) and more general models based on Choquet capacities [ a ] , [55] have also been considered.

To illustrate the potential effectiveness of the robust filter we return to the problem described by the nominals of (2.7) and the uncertainty model of (2.8). This is a ppoint model with a, = [-1,1], Q, = Q;, ps = 1/2, and pN = 0.063. Fig. 4 superimposes the (constant) performance of the robust filter (2.20) for this case onto the nominal and worst case performance curves of the nominal filter from Fig. 1. Note that the performance of the robust filter over the entire

Fig. 4. Performance curves depicting the favorability of the robust filter for the example of Subsection 11-A. and

KASSAM A N D POOR: ROBUST TECHNIQUES F O R SIGNAL PROCESSING 439

Page 8: Robust Techniques for Signal Processing - A Survey

uncertainty class is only slightly degraded from the nominal performance of the nominal filter and is remarkably im- proved over the worst case performance of the nominal filter. This example illustrates fairly dramatically the favora- bility of the minimax design for filtering in uncertain en- vironments.

D. Robust Causal Filtering, Prediction, and Smoothing

The results discussed in the above subsections assume implicitly that the class of allowable estimation filters in- cludes noncausal filters; i.e., h(t - T ) in (2.2) is not neces- sarily zero for T > t. While this assumption is not restrictive for many applications such as those involving spatial filter- ing or enhancement of stored signals, there are also many applications in which causality of the estimation filter is desired for the purposes of real-time processing. To discuss the situation we consider again the observation model given by

v( t ) = S( t ) + N( t ) , -00 c t c 00 (2.22)

where { S ( t ) ; -m < t c a} and { N ( t ) ; -00 < t c 00) satisfy the assumptions made below (2.1). We wish to estimate the signal at time t + A for some fixed X based on observations up to time t; i.e., we wish to consider esti- mates of the form

i ( t + h ) = J f h ( t - T ) Y ( T ) d T . - m

(2.23)

Note that X c 0 in (2.23) corresponds to fixed-lag smooth- ing of the signal, X = 0 corresponds to causal filtering, and A > 0 corresponds to signal prediction.

The M S E associated with the estimate of (2.23) is given by

+ I H(

' e A ( a S ! @ N i H) (2.24)

where H is the transfer function of the filter { h( t); t 2 O}. For fixed as and mN such that the observation spectrum, (as + aN), satisfies the Paley-Wiener condition, the MSE of (2.24) is minimized over all causal filters by the filter with transfer function

Ht( 0) = 1 [ '1 € J A W s ( 0 )

@S( w , + ' N ( 0)1' @S( 0> + @N( +

(2.25)

where the subscript + denotes causal part in an additive spectral decomposition and the superscripts + and - de- note causal and anticausal parts, respectively, in a multi- plicative spectral decomposition (see, e.g., Wong [56]). The minimum M S E is then obtained by combining (2.24) and (2.25) as

e A ( a S r a N ; H t ) ef(@S,@N). (2.26)

As in the noncausal situation discussed above, it is com- monly the case in practice that as and m N are not known exactly but rather are known to fie in some uncertainty classes Y and X of possible signal and noise spectra. In this case we may seek a robust filter for Y and X by

minimizing the worst case error

sup eA( ON; H, (2.27) (@, ,@,)SYXK

over all causal transfer functions H. Although this problem is analytically more difficult than the corresponding non- causal problem, it can be shown for convex Y and JV that, within mild conditions, a saddle point solution for this problem is given by the optimum causal filter HL corre- sponding to the least favorable spectral pair ( @ s , L , @ N , l ) for this causal problem, where ( @ 5 , 1 , ( P N , l ) is defined via

(see Poor [45], Vastola and Poor [55], Franke [57], Vastola [58], Franke and Poor [SS]).

Thus conceptually, the causal robust signal estimation problem is no more difficult than the noncausal one, since one designs a robust filter by first maximizing e:(as,QN) over Y X X and then designing an optimum filter for the maximizing pair via (2.25). However, in the noncausal prob- lem, there is a tractable closed-form expression, namely (2.5), to be maximized to find least favorables, whereas no such general expression exists for the causal problem; i.e., there is no general closed-form expression for e\(Qs,mN) of (2.26). O n the other hand, there are many specific cases of practical interest for which 4(Qs,QN) is known in closed form (see, e.g., Yao [ a ] , Snyders [61], [62D and, moreover, general (but tedious) methods for finding such expressions are available (e.g., [61D. Robustness in several causal filter- ing problems has been considered using these results. Cen- erally speaking, the phenomena observed are more or less the same as for the noncausal case, although, for a given model, nominal causal filters appear to be somewhat less sensitive to uncertainty than nominal noncausal filters are, due to the relatively lower selectivity of causal filters (see, e.g., Vastola and Poor [a]).

Example: Robust Prediction: An interesting example of an application in which the above results can be easily applied is the problem of discrete-time one-step pure pre- d i ~ t i o n . ~ I n particular, suppose we observe a discrete-time signal directly up to some time t; Le., we have

Y ( k ) = S ( k ) , k E { * * . , t - 3 , t - 2 , t - l , t } . (2.29)

Suppose further that we wish to predict the value of S ( k ) at the next sampling instant k = ( t + 1) with a linear predictor

t

i ( f + i ) = 2 h ( t - k ) s ( k ) . (2.30) k - - W

This is the problem of (2.23) with A = 1 and O N ( o ) = 0 for all w .

The minimum-MSE functional 4(aS,aN) for this prob- lem is given by the well-known Kolmogorov-Szego-Krein formula (Hoffman [63])

e!(aS,O) = exp( '/" 2 a -" log@s(o) d u } . (2.31)

Thus in order to design a predictor to be robust over an uncertainty class Y of signal spectra, we choose as, via

'Note that the discrete-time problem i s the special case of the continuous-time one in which the spectra are concentrated in the spectral band IwI 6 n. Thus the above discussion holds for both discrete and continuous time.

440 PROCEEDINGS O F THE IEEE, VOL. 73, NO, 3, MARCH 1965

Page 9: Robust Techniques for Signal Processing - A Survey

@ s , L = arg{ max /" log@s(w) d o ) (2.32) (PSESP -"

and we then design the optimum predictor for this spec- trum. This problem has been considered by Hosoya [64] for the particular case of an €-contaminated spectrum and by Franke [57] and Vastola and Poor [55] for the general case.6 A related problem has been considered by Korobochkin [66].

It i s interesting to note that the spectrum of (2.32) can be interpreted as being that member of 9 which is "closest" to a uniform spectrum (see Poor [67]).' This has a very nice intuitive interpretation since a uniform spectrum corre- sponds to white noise, which is the universal worst case type of signal to predict. (In other words, past and present data are useless in predicting future values of white noise). It is also interesting to note that the quantity

-/ log@s( a) do 297 -" 1 "

is the spectral entropy of the signal process (see, e.g., [a]) , and thus the least favorable spectrum is maxentropic in 9. Also, since the entropy of a process is a measure of its indeterminism, the least favorable spectrum can be thought of as the most indeterministic, a term introduced by Franke W I .

As a specific example, we consider the particular case of an €-contaminated first-order wide-sense-Markov signal; i.e, we have

@,( o) do = I} (2.33) 2.n - m

where

I -$

@ S ! O ( ~ ) = 1 - 2 r c o s ( o ) + r~ ' - . n < w < . n

(2.34)

and where 0 B r 1. The least favorable spectrum for this class is given by (see [67n

@ S , L ( W ) = { ;,- C ) @ ~ , ~ ( W ) , if (I - E ) @ ~ , ~ ( O ) > c' if (I - ~ ) @ ~ , , ( o ) < c'

(2.35) where c' is chosen so that

-1 @s,L(o) do = 1 I =

2.n - m

This spectrum is illustrated in Fig. 5 for the case r = 0.5 and c = 0.25. Note that, as c increases, the peak in the center of the frequency band "melts" into a uniform spectrum in the frequency tails.

'A nonstatistical approach to the minimax prediction problem has been considered by several authors; a recent study by Colubev and Pinsker is found in [65].

'This follows since

where C is the convex function C(x) = -1ogx. Comparing with (2.13) one sees the interpretation of the negative of the spectral entropy of as being a measure of the distance of QS from the uniform spectrum.

2 0 -

L5 - I

1.0 - -

05 -

I P -* f 0 3 9 i 1 , 1

Frequency, w

Fig. 5. Least favorable spectrum for predicting an c-con- taminated wide-sense Markov signal.

Another interesting example of robust prediction comes from consideration of the following signal spectral uncer- tainty classes (see Franke [57]):

(2.36) where c,,. . -, c, is a set of constants. Since

1 /" e'"k@,( w ) do 2.n -"

is the kth-lag autocorrelation (E{S(! )S(!+ k ) } ) of the sig- nal, (2.36) corresponds to the set of all spectra whose first ( m + 1)-lag autocorrelations take the given values c,;. .,c,. Such a model is applicable, for example, when we have measurements of a finite number of autocorrela- tions of the signal process. Note that (2.36) i s an example of the generalized moment constrained model from Example 5 of the preceding subsection.

To find a robust predictor for the class 9 of (2.36) we first look for the least favorable signal spectrum by solving the constrained optimization problem

1 " O,E92.n -" max - / log as( o) do

subject to - /" Zok@s( o) do = c k , I k l = 0 ; . 1 , m . 1

2.n -" (2.37)

Equation (2.37) will be recognized as the well-known maxi- mum entropy spectrum fitting problem, and i ts solution is straightforwardly seen to be given by

I k - 0 I with a, = 1 and with a l ; . . , a M and & satisfying the Yule-Walker equations for the correlations c,; . ., c,,, (see, e.g., [69]). The spectrum ass.[ of (2.38) i s the spectrum of the mth-order autoregression

rn

S ( t ) = d k s ( t + k ) + g r ( t ) , t = 0 , ~ 1 , ~ 2 , ' . ' k - 1

(2.39)

where { c( t)}y- - is a sequence of orthogonal, zero-mean, and unit-variance random variables. Thus the optimum one-step predictor for the least favorable spectrum (2.38) i s

K A S S A M A N D P O O R : R O B U S T TECHNIQUES FOR SIGNAL PROCESSING 441

Page 10: Robust Techniques for Signal Processing - A Survey

given by m

3( t + 1) = dkS( t + 1 + k ) (2.40) k = l

and so the robust one-step predictor for (2.36) is a simple finite-length predictor with coefficients determined from the Yule-Walker equations. That this predictor uses only rn past samples is an intuitively pleasing result, since we have no knowledge of the correlation structure beyond lags of length rn. This result has been generalized to pstep predic- tion by Franke and Poor in [59].

kides and Kassam in [70]. To illustrate the nature of solu- tions to this problem we suppose, for example, that the channel is known to have a linear phase characteristic and that the channel gain l K ( w ) l is known only to l ie between known upper and lower limits L c ( w ) and U,(o), respec- tively. That is, suppose we have

L C ( @ ) Q IK(w)I Q U c ( w ) ( 2 . 4 ) for all frequencies a, but that I K ( w ) l is otherwise unknown. Then, assuming Uc(w) > 0 and Q N ( w ) > 0, i t can be shown [70] that the magnitude of the robust (minimax) equaliza- tion filter H R ( w ) is given by

E. Robust Equalization of an Uncertain Channel

In the observation model of (2.1) it is assumed that the signal we are interested in estimating is corrupted only by the additive orthogonal noise process { N( t ) ; -bo < t < m } . However, in many situations of practical interest we also have linear distortion or spreading of the signal by the observation channel. This situation can be described by an observation model of the form

Y ( t ) = / - k ( f - ~ ) s ( ~ ) d ~ + N ( t ) , - m < t < m --m

(2.41)

where the noise and signal processes satisfy the assump- tions made after (2.1) and where k ( t ) is a channel spread function. The problem of estimating S( t ) from the observa- tion of (2.41) is the general problem of channel equalization (or deconvolution) plus filtering, which arises in many ap- plications such as communications, sonar, seismology, and image processing.

If, as before, we consider signal estimates of the form

i ( t ) =/- h( t -T )Y(T )dT (2.42) - m

then with known as, Q N , and k , the optimum (minimum- MSE) equalization filter is given straightforwardly by the transfer function

where K is the transfer function of the channel and where K* denotes i ts complex conjugate.

In practice, the transfer function of the channel is rarely known precisely. However, to design the optimum equali- zation filter of (2.43) one needs exact knowledge of this channel characteristic. Thus as in the case in which the signal and noise spectra are uncertain, it is necessary to seek an alternative design objective for situations in which the channel characteristic is uncertain. In particular, if we can model the channel as having a transfer function which lies in an uncertainty class X , then an appropriate design criterion might be a minimax M S E formulation where the maximization i s taken over all channels in the class X .

This minimax formulation for equalization of uncertain channels has been proposed and investigated by Mousta-

where

(2.45)

Note that the quantity @ N ( o ) / L ~ ( w ) @ s ( o ) i s a measure of the maximum possible noise-to-signal ratio at frequency o, and A ( w ) is a measure of the uncertainty in the knowledge about the channel. Thus (2.45) implies that if the maximum noise-to-signal ratio at a given frequency is larger than the uncertainty in the channel model, then we use the opti- mum gain for the lower channel, Lc(o), at that frequency. Alternatively, i f the opposite is true for a given frequency, then we simply ignore the noise at that frequency and use the gain prescribed by the inverse average channel. A simi- lar result can be obtained for situations in which the phase of K ( w ) is also unknown.

F. Robust Filtering of Signals in Correlated Noise

Another aspect of the assumptions made above on the model (2.1) that is sometimes violated in practice is that of no correlation between the signal and noise processes. Signal-dependent noise arises in many applications such as radar or sonar, for example, due to phenomena such as multipath and clutter. Often the correlation between signal and noise in such applications is not well modeled, so that robust techniques are useful.

If we assume that the signal and noise processes of (2.1) are jointly wide-sense stationary, then their total correlation picture can be described by the spectral density matrix D given by

where as and aN are as before, QSN i s the cross spectrum between {S( t ) ; - m < t < 03) and { N(t); - m t < m } , and where, as before, the asterisk denotes complex conju- gation.

For a signal estimate of the form

we have that the M S E is given straightforwardly in this case by

442 PROCEEDINGS O F THE IEEE. VOL 73, NO. 3, MARCH 1985

Page 11: Robust Techniques for Signal Processing - A Survey

@S( + @N( w , H w ( w ) = as( w ) + 2 Re { aSN( o)} + QN( w )

(2.50)

and the corresponding minimum value of M S E is given by

e(@,,@,,@s,,; Hw)

A eW(@St@N,@sN). (2.51)

Of course, with @ s N ( w ) identically zero the expressions (2.49)-(2.51) reduce to (2.3)-(2.5).

If the correlation matrix D of (2.47) is not known pre- cisely but, rather, is known only to lie in some class 9 of spectral density matrices, then to seek a robust alternative to the optimum filter of (2.50) we can replace the objective function e(@s,@,,@sN; H) with its supremum over 9. As in the uncorrelated signal and noise case, this yields a mini- max game for designing a robust filter H,. The solution to this problem has been considered by Moustakides and Kassam [71], [72]. Within mild conditions it can be shown that, for convex classes 9, a spectral density matrix Dl E 9 and its corresponding optimum filter from (2.50) (when it is uniquely defined) will be a saddle-point solution to this game if and only if DL is least favorable; i.e., if and only i f

D, = arg ( min ew(@s,@Nt@sN)). (2.52)

To illustrate the possible structure of the least favorable spectral density matrix, it is interesting to consider the case in which the signal and noise spectra are known but the cross spectrum is not known precisely. In particular, sup- pose we can establish the fact that the cross spectrum satisfies the conditions

D E 9

0 d L ( o ) d p S N ( W ) I Q u(o), -00 < 0 < 0O

(2.53)

where L and U are given functions. (Such a model might arise, for example, if a confidence band for the cross spec- trum could be determined via spectrum estimation.) For this model it can be shown (see Moustakides and Kassam 1721) that the least favorable cross spectrum is given by

i -L (o) , if B ( w ) d L ( o )

@sN,L(w) = - I3(0), if L ( m ) d B ( 0 ) Q U ( o ) - U ( o ) , if U ( w ) d B ( o )

(2.54)

where I3 is defined by

B ( w ) A min{@s(w),@N(w)}. (2.55)

Thus at a given frequency, whether the worst case involves minimum cross-spectral density, maximum cross-spectral density, or something in between, depends on the relation- ship among L , 13, and U at that frequency. If, for example,

nothing is known about the cross spectrum, then all we can say is that

o d l @ ~ ~ ( w ) I d [ @ ~ ( ~ ) @ N ( ~ ) ] 1 ’ 2 , - 0 O < w < m

(2.56)

where the right-hand inequality follows from the required nonnegative definiteness of D(o). Since min { a , b } d

for all a >, 0 and b 2 0, it follows straightforwardly from (2.54) that in this case we have

@ s N , L ( ~ ) = -min{@s(w),@N(w)}, -00 < w < 00 . (2.57)

Equation (2.57) together with the optimum-filter expression of (2.50) gives that the robust filter for completely unknown cross correlation is given by the somewhat surprising result

Other results for different bounded uncertainty classes are giveq in [71].

G. Uncertainty Classes of Spectral Measures

In the above discussion, we have considered several aspects of the basic problem of robust linear filtering of stationary signals in additive stationary noise. In particular, we have discussed the basic robust filtering problem in a noncausal framework, and we have also discussed the treat- ment of filter causality, equalization, and cross correlation between signal and noise. One issue which has not been discussed is the treatment of stationary processes that do not necessarily have spectral densities but, rather, have associated spectral measures (or, equivalently, spectral dis- tributions). This situation arises in practice primarily when there are pure harmonics in the signal and/or noise. For example, in a communications receiver, one might have processes that nominally have spectral densities but also contain pure-harmonic uncertainties caused by sources such as line hum or tone jamming. To treat the robustness problem in this more general context requires a measure- theoretic formulation of the filtering problem. This issue has been considered by Poor [48] for the case of noncausal filtering and by Vastola and Poor [55] and Franke and Poor 1591 for the case of causal filtering. The results obtained for this situation are quite similar to those for the case in which all processes concerned have spectral densities, with the additional advantage that quite general results concerning the existence of least favorable spectral measures can be obtained.

In particular, suppose we have the observation model of (2.1) in which the signal and noise processes are real, zero-mean, orthogonal, wide-sense stationary, and quad- ratic-mean continuous. Then for an estimate of the form of (2.2) we can always write the M S E as

E ( I W - m 2 )

= /Oa 11 - H(w)12m, (do) - m

(2.59)

443 KASSAM AND P O O R : R O B U S T T E C H N I Q U E S FOR SIGNAL PROCESSING

Page 12: Robust Techniques for Signal Processing - A Survey

where m, and mN are the spectral measures associated (via Bochner’s Theorem [56n with the processes { S( t); - m t < m} and { N( t ) ; - m < t < m}, respectively. The transfer function that minimizes (2.59) for fixed m, and mN is given by

H w ( w ) = d( m, + mN) ( w )

where the differentiation in (2.60) denotes the Radon- Nikodym derivative. Alternatively, if m, and mN are known only to lie in classes -k; and X , respectively, then we would like to solve the minimax problem

min sup e( m,, m,; H ) ) . (2.61)

An interesting class of such problems arises when total powers in the signal and noise processes are both known (i.e., when ms(( - m, m)) 2nP, and mN(( - 00, 00)) ZnP, are constant on -k; and X). In this case, solutions to (2.61) can be characterized for a general type of uncertainty class studied by Huber and Strassen in [73]. These classes are of the form

4= { m E d I m ( B ) g v ( B ) ,

H (cms,mN)=&x&

for all 6 E B and m(R) = v(R)} (2.62)

where R denotes the set of real numbers, B denotes the a field of Bore1 sets in R,Adenotes the set of all spectral measures on (R,S?), and v is a 2-alternating (Choquet) capacity. A 2-alternating capacity in this context is a set function mapping B to R with the following properties: v ( + ) = 0, v(R) < m, A c 6 - v ( A ) < v(B), v is continu- ous from below and is continuous from above on closed sets, and v( A U B ) + v( A n B ) g v( A ) + v( B ) for all A and B in S?. Examples of classes of the type of (2.62) are given in [73]-[76] and include such common uncertainty models as the r-contaminated class, the variational neighborhood, the band model, and others.

Classes of the form of (2.62) are useful for the problem of (2.61) because the solution to (2.61) for the situation in which = 4s and = 4N for two 2-alternating capac- ities q and v, is given by (see Poor [MI)

where n = dq/dvN is a Radon-Nikodym derivative be- tween % and v,. (Note that H, from (2.60) could be written as H,(o) = X(o)/(X(o) + 1) where X = dm,/ dm,). Thus for uncertainty classes generated by 2-alternat- ing capacities the general solution to (2.61) is characterized. Similar results for causal problems are discussed in [55].

H. Robust Interpolation

A problem related to robust filtering is that of robust signal interpolation. In the signal interpolation problem we have a discrete-time signal S ( k ) , k = 0, *I, f2;.., which we observe for all k # 0. Our objective is to estimate S(0) from { S ( k ) ; k = *I, *2, ... } with a linear estimate

i(0) = h( k ) S ( k ) . (2.64) k # O

The mean-square estimation error incurred by using (2.64)

is given straightforwardly by

E( I i(0) - S(0) 1’) = 2a 11 - H( a) 12@,( o) do

e( as; H ) (2.65)

where H is the transfer function of the filter sequence - . . ,h( -2) ,h( - l ) ,O,h(+ l ) ,h(+2) ; . . and where @, is

the spectrum of the signal. For fixed a,, the MSE of (2.65) is minimized by the interpolator with transfer function

/-/,(a) =I -- [ - (I/@,( 0)) d w l - ’ . @,(o) 2~ --I

(Note that the zeroth Fourier coefficient of H, equals zero, as is necessary for H, to be a valid interpolator.) The minimum possible value of the mean-square interpolation error for a given signal spectrum @, is given by

e(@,; H,) = [L/“ 2n (l/@,(u)) d w 1 - l e , (@,) .

(2.67)

Suppose we design an interpolator based on an assumed signal spectrum but that the true spectrum is a,. The resulting mean-square interpolation error is given by insert- ing the optimum interpolator (2.66) for into (2.65), in which case we have the mismatch error

Note that the numerator integrand @5/(@,,o)2 implies a high degree of sensitivity of the interpolation error to large spectral components in the actual spectrum that are not predicted by the nominal model. Thus as with the other linear-estimation problems discussed in the preceding sec- tions, it is desirable to robustify an interpolator against uncertainty in @,. This can be done by embedding @, in a spectral uncertainty class 9 and replacing the minimization of e(@,; H) with the minimization of its worst case value over 9. This problem was posed and solved by Taniguchi in [77] for the case in which 9 is an r-contaminated model, and the solution to the general case was characterized by Kassam in [78].

In [78] it is shown that, within mild conditions on 9, a spectrum @,, E 9 and its corresponding optimum interpo- lator from (2.66) will be minimax robust over SP if and only i f @,,L is least favorable; i.e., if and only if

a,, = arg max e,( as) a S € 9

(2.69)

where e, is from (2.67). Moreover, note that equiv- alently minimizes

/ w C( @,( 0)) dw 2 n --I

where C is the convex function C(;) = l /x . Thus as with the filtering and prediction problem discussed above, least favorables for many normalized uncertainty classes can be found by applying analogous results from Huber’s robust hypothesis-testing formulation (see Kassam [78] for details).

444 PROCEEDINGS O F THE IEEE. VOL. 73. NO. 3, MARCH 1985

Page 13: Robust Techniques for Signal Processing - A Survey

I. Robust Linear Filtering for Vector Observations

Many problems of practical interest involve the vector version of the observation model (2.1). Although conceptu- ally similar t o the scalar problem, vector filtering problems often present practical difficulties that are usually overcome by the imposition of specific structural models such as a finite-state signal model. These difficulties are not allevia- ted in the robust versions of vector filtering problems, and so the types of uncertainty models for which minimax robust filters can be developed are usually more structured than in the single variable case. Developments for the vector problem analogous to those for the scalar problem discussed in the preceding subsection are found in [79]-[83]. As an example of the type of uncertainty class that can be treated in this context, Chen and Kassam [81] consider an uncertainty set of spectral density matrices which share a common (constant) eigenstructure but whose eigenspectra lie in band models.

Another set of structural assumptions that allow for the treatment not only of the vector case but also of time-vary- ing situations is the usual state-space signal model

S( t ) = C( t ) X ( t )

X( t ) = A( t ) X ( t ) + v( t ) , t > to (2.70) where E { v( t ) v( s)} = Q( t ) 8 ( t - s), { v( t ) } is independent of { N( t ) ) , and A( t ) and C( t ) are matrices of appropriate dimensions. We assume that the observation noise { N ( t ) } has correlation E{ N( t )N(s ) } = R ( t ) 8 ( t - s). The best lin- ear estimator of X ( t ) from { Y ( T ) ; to < T < t } is given by the well-known Kalman-Bucy filter

j?( t ) = A( t ) i ( t ) + K( t ) [ V( t ) - c( t ) i( t ) ] (2.71)

where K( t ) is the Kalman gain matrix determined from the error covariance P( t ) = COV(X( t ) - k( t ) ) .

Although uncertainty can arise within the model (2.70) in several ways, there are two basic types of uncertainty that can affect the performance of linear estimators. One of these is uncertainty in the second-order statistical proper- ties of the noise (and initial condition), such as the matrices Q and R and the assumptions that { v( t ) } and { N( t ) } are white and uncorrelated. The other type of uncertainty is uncertainty in the dynamical model itself; i.e., uncertainty in the A and C matrices. Problems of the first type have been studied by a number of investigators including D’Appolito and Hutchinson [79]. VandeLinde [MI, Morris [85], Poor and Looze [%I, and Verdu and Poor [87]. The basic result for minimax design within this type of uncer- tainty is that the linear structure is preserved in the minimax solution and the corresponding gain matrix is chosen to be optimum for a least favorable set of second-order statistics. These results are thus of the same type as those discussed in the preceding subsections. Problems of the second type, however, have received less attention in this context, and have been treated only recently in a paper by Martin and Mintz [ a ] . As one might expect, the effects of uncertainties in the dynamical structure of the model (the A matrix) are quite different than those of uncertainty in the noise statis- tics. In particular, the minimax solutions for uncertain A matrix are not pure strategies (i.e., they are not simply of the form (2.71)) but rather are mixed strategies-randomiza- tions among several filters of the type (2.71). Another ap- proach to this problem has been proposed by Cusak [89].

1 1 1 . ROBUST FILTERS F O R SIGNAL DETECTION AND RELATED APPLICATIONS

One of the most pervasive of functions that signal processing schemes are required to carry out i s that of detecting a signal of a generally known type in noisy observations. Obvious examples of applications in which signal detection is required are provided by radar (detection of echo pulses) and sonar (detection of a random signal present in an array of hydrophones). Numerous other appli- cations may be listed; for example, detection of specified two-level pulse-code sequences in communication systems and detection of abnormal patterns in medical imaging.

In the classical theory of signal detection one starts with specific statistical or deterministic models for the signal and observation process, and proceeds to obtain a detector that has optimum performance under an appropriate detection performance measure such as detection probability or out- put SNR. In this section we will restrict attention to the class of linear detectors or filters and to the output signal- to-noise (SNR) as the measure of a linear filter’s detection performance. The design of optimum detectors under these two restrictions has been carried out for a large number of special applications. There are several reasons for the widespread acceptance of these design restrictions. One is that optimization of the linear filter design to maximize output SNR is usually a simple mathematical problem and leads to explicit solutions, and the implementation of a linear filter is usually straightforward. Another reason is that in many cases the statistical models for the noise and any random signal are often stated in terms of only their corre- lation functions or power spectral densities (that is, their secondorder statistical properties). In such cases, one sim- ply does not have enough statistical information, such as the parametric form of the probability-density functions, to allow optimization to be attempted over a larger class of detectors. If the noise process can be assumed to be Gaussian then it it possible to show that the optimum detector maximizing the detection probability in the detec- tion of a deterministic signal is indeed the optimum linear filter maximizing the output SNR; this of course is the ubiquitous matched filter of detection theory.

In general, the matched-filter specification depends on the exact form of the deterministic signal for which it maximizes i ts output SNR, and on the exact noise autocor- relation function or power spectral density. Since these quantities are rarely known exactly, the need for applying robust techniques may arise naturally in the matched filter- ing problem. One interesting extreme case occurs when the nominal assumptions on signal and noise characteristics are such as to result in a singular detection problem. This means that under nominal conditions the output SNR is infinitely large, implying perfect detection is possible. Con- sider, for example, an ideal low-pass nominal signal sinc(w,t) whose Fourier transform S ( w ) i s shown in Fig. 6, and let the noise have a nominal power spectral density O N ( w ) which is the triangular function of Fig. 7. The matched-filter frequency response S * ( w ) / @ . , ( w ) i s shown in Fig. 8; it increases without bound as w approaches +ao. The output SNR under this situation is

KASSAM AND POOR. ROBUST TECHNIQUES FOR SIGNAL PROCESSING 445

Page 14: Robust Techniques for Signal Processing - A Survey

Fig. 6. Ideal low-pass nominal signal characteristic S(o).

- WO 0 WO

Fig. 7. Triangular noise power spectral density Q N ( o )

- WO 0 WO

Fig. 8. Mached filter frequency response for ideal low-pass signal and triangular noise power spectral density.

But suppose that the signal deviates very slightly from the nominal and becomes sinc([q, - r]t) where c i s small and positive. The output SNR using the original matched filter now drops to zero, or - co decibels! In this situation where the signal bandwidth may not be precisely q,, it would be better to design the matched filter for the smallest band- width that may be encountered. The resulting filter will then perform well for the minimum bandwidth signal and will be fairly insensitive to deviations of the signal band- width from q,. This basic idea is expanded upon in Subsec- tions Ill-A and Ill-C, where we consider minimax robust matched filters which maximize the worst case perfor- mance over a given pair of classes for the allowable signal and noise characteristics. Also a property of such robust matched filters in most cases is that their performance is relatively stable, or not too variable, over the allowable classes of characteristics. The issues of stability and singular- ity in certain matched filtering problems have been consid- ered by several authors, most notably Root [2] and Kailath [%I. The minimax robust matched filter has also been described as an optimally stable matched filter.

While the simple example we have considered above to illustrate the possibility of extreme sensitivity of perfor- mance of a nominal matched filter assumed that only the value of one parameter (the signal bandwidth) was variable, in general we deal with nonparametric classes of functions in modeling the uncertainties in signal and noise character- istics. The results and interpretations we survey in this

section differ from those in the previous section on robust filters for signal estimation primarily because our criterion of performance here is the output SNR. This leads to mathematical approaches that are conveniently unified using Hilbert spaces. In addition, unlike the Wiener filtering problem, no direct correspondence exists between these results on minimax robust matched filters and results on robust hypothesis tests, although some indirect connections do exist.

In Subsection Ill-B we will give some interesting special applications of the general results on robust matched filters in spatial array processing and time-delay estimation. We will close this section with a more mathematical discussion in Subsection Ill-C of the general Hilbert space approach for formulating robust linear detection problems; this ap- proach provides a common framework for a variety of such problems.

A. Robust Matched Filters

As an example of a system using a matched filter for signal detection, consider the structure of the receiver for a pulse train in which a given pulse shape p( t ) occurs in the i t h position of the train with some amplitude a,, i = 1,2;-., m. In a pulsed radar system these might represent the echo pulses from a target in some fixed range gate, with the amplitude factors a, produced by the beam pattern of the receiving antenna as i ts main beam scans past the target position. The noisy received waveform may be described by the equation

rn

v( t ) = e d i p j ( t ) + N( t ) , o G t G mT (3.2) i = l

where p,( t ) = p( t - [ i - 11 T), the basic pulse p( t ) delayed by ( i - l ) T units of time, Tis the pulse repetition period, 8 i s the overall pulse-train amplitude, and N ( t ) is random noise. Note that the above observation model can also describe the received signal in a binary signaling scheme. For example, we may take m = 1 and 8 = $, 8 = 8, as being two possible values of 8, with eo = 0 and 8, # 0 corresponding to the case of on-off keying; or, with m > 1 (3.2) may describe a coded waveform such as are used in direct-sequence spread-spectrum modulation, with p( t ) being the "chip" waveform and a1,a,; ..,a,,, being a pseu- donoise code sequence. For the general problem of testing 8 = 8, versus 8 = 8, in the observation model (3.2) when N ( t ) is Gaussian with power spectral density aN(u), an optimum receiver structure is shown in Fig. 9. This receiver is an implementation of the likelihood-ratio test for 8 = 8, versus 8 = in the case of Gaussian noise. More generally, the output of the matched filter at the end of each pulse interval has the maximum SNR obtainable from a linear filter. The frequency response /-/,(a) of the matched filter is

. I oj

Fig. 9. Structure of optimum receiver for pulse train in Gaussian noise.

I

L---------i

446 P R O C E E D I N G S OF THE IEEE, VOL. 7 3 , NO. 3, MARCH 1985

Page 15: Robust Techniques for Signal Processing - A Survey

where P ( w ) is the Fourier transform of the pulse p( t ) . We shall now focus on the design of the matched filter when P ( o ) and Q N ( w ) are not precisely known. In the next section we will survey techniques for modifying the correla- tor detector following the matched filter in Fig. 9 for situa- tions where the noise at its input cannot be assumed to be Gaussian.

The general expression for the SNR at time T at the output of a filter with frequency response H ( o ) when the input is Bp( t ) + N( t ) , is

This SNR is maximized by H ( o ) = H,(w) of (3.3). The synthesis of H M ( w ) in this case requires an exact knowl- edge of P ( w ) and Q N ( w ) . Suppose, however, that the received pulse p ( t ) i s a possibly distorted version of a nominal pulse &(f), as illustrated in Fig. 10. A good mea-

makes it too sensitive to deviations in the pulse characteris- tics, the actual pulse possibly having lower energy at these frequencies. Thus the filter (3.7) is robust from an intuitive viewpoint. Another interpretation of (3.7) is that signal uncertainty is translated into an additional white-noise component in the noise spectrum. It is interesting to note that in the case of white noise with the above signal uncertainty the filter matched to the nominal pulse is itself the robust filter, since any absolute gain factor is irrelevant in matched filtering. This result agrees well with behavior observed in practice [93]. It is also interesting to note that the additional white-noise component which appears un- der signal uncertainty in this approach ensures that the detection is not singular, regardless of how artificial the nominal model is.

Alternatively, suppose we assume that the pulse shape, say h( f), and the total noise power uk are known, but that the true noise spectrum is known only to lie in a class V, of spectra bounded by known upper and lower bounds U,(w) and LN(w) , respectively. That is, suppose we assume that the spectral shape of the noise is constrained only by the band model

Tlme,t - Fig. 10. Non-nominal pulse shapes p( t )

sure of the degree of pulse distortion is the integrated squared difference between p( f) and &( t ) or, equivalently (via Parceval's relation), the integrated squared difference between their Fourier transforms, P ( o ) and Po(w). Thus a useful signal uncertainty model for this matched filtering problem is to assume that the received pulse p( f) i s known only to lie in the class V, which is the class of pulses p( t ) with Fourier transforms P ( w ) satisfying

Here P,(o) i s the Fourier transform of the nominal pulse, and A determines the degree of uncertainty or possible distortion in p ( t ) .

The solution H = H, to the problem

max min SNR (3.6)

where SNR is from (3.4) is given by Kuznetsov [91] and Kassam, Lim and Cimini [92] as

H P € W p

where the positive constant 00 depends on A and is an increasing function of A . Equations (3.3) and (3.7) imply that the robust filter for (3.5) is forced to have less gain than it would otherwise have at frequencies where the noise power is low. It is the higher gain of the optimum filter for Po(w) [that is, Pb(w)e - j "T /QN(w) ] at these frequencies which

and

- / Q N ( a) do = u i 2 8 - m

I * (3.8b)

This class, which is the same as that considered as Example 4 in Subsection 11-C, i s illustrated in Fig. 11. In this case the

Spectral Denslty

t

Allowable No~seSpectra

Frequency, w

Fig. 11. Bounded noise power spectral density class

robust filter can again be given an interesting interpretation. It turns out that the robust filter H,(o) in this case is the matched filter for detecting &(f) in noise whose spectrum tends to be as close in its shape as possible to the shape of IPo(w)l. Specifically, the results in [9l], [92] show that the robust filter solving

max min SNR (3.9) H @,E%',

i s given by

(3.10)

KASSAM AND POOR ROBUST TECHNIQUES F O R SIGNAL PROCESSING 447

Page 16: Robust Techniques for Signal Processing - A Survey

Here k is a constant determined by the requirement that @N,L(o) must be in ‘3, and therefore must have total power uh, which i s assumed to be known. It is clear that @ N , L ( o ) tries to follow as closely as possible a scaled version of IPo(o) l . This phenomenon is illustrated in Fig. 12. An illustration of the magnitude of H R ( u ) for the example of Fig. 12 is shown in Fig. 13. This figure also shows what the optimum filter magnitude would be if a “nominal”

Frequency, w - Fig. 12. Illustration of least favorable noise spectral density of (3.11).

Frequency, w - Fig. 13. Illustration of IHn(u)1, robust matched filter ampli- tude response, for a bounded noise power spectral density class.

noise spectrum having the characteristic (c/2)[L,(u) + U,(u)] (that is, the normalized center of the band) is used in the design of the filter, where c is a normalizing constant chosen to get the correct power u&. The implication for the robust filter i s that its magnitude characteristic is flatter than it would be with any other @,(a) in 9,. This means that extremes of gain and attenuation do not appear, mak- ing the filter less sensitive to the presence of additional noise power at relatively high-gain frequencies and to a reduction in noise power at relatively high-attenuation fre- quencies. The performance of the robust filter may thus be said to have been stabilized. Of course, the advantage gained in using a robust filter in any particular case depends on how extreme the frequency response of the nominally optimum filter is.

By combining and extending the above results it is possi- ble to obtain the robust filter for simultaneous uncertainties about the pulse shape and the noise spectrum. This has been done in [92]. The extension of this spectral-domain formulation of the matched-filtering problem to the case of discrete-time observation processes is quite straightforward.

It i s interesting to note that an explicit consideration of a game-theoretic approach in the design of a matched filter was first considered by Nilsson [I21 and Zetterberg [13]. In

[13], in particular, Zetterberg considered uncertainty in the noise spectrum which is assumed to have a known white- noise component and another component which has fixed total power but is otherwise unknown. He obtained a result which i s similar to that of (3.11). Although Zetterberg did not allow uncertainty in the signal, it is noteworthy that he assumed a fixed white component for the noise spectrum. We have seen above that a white-noise component for the noise spectrum is prescribed for the L, signal uncertainty class of (3.5). Other related studies are also discussed in [13]. More recent work of this nature has also been reported by Rodionov [94], Cahn [95], and Turin [%I.

The multi-input matched filtering problem under signal and noise uncertainty in the frequency domain has recently been considered by Chen and Kassam [97], [%I. An m-com- ponent signal vector s ( t ) may be modeled to belong to a generalized L , uncertainty class that is a neighborhood of some nominal signal vector so( t ) . In [97], [ X ] the class was defined in terms of a set of constants A , , A 2 ; . . , A q ; these were now the bounds on the allowable values of q sums of component-wise integrated square differences between the vectors s( t ) and s o ( t ) . The noise power spectral density matrix @,(u) was assumed to lie in a generalized bounded spectrum class. Specifically, from the decomposition @,(a) = P,(u)A,(u)PZ(u) where P , ( u ) is a unitary matrix and A N ( o ) i s diagonal, one may obtain an uncertainty class by requiring the components of hN(a) to have known upper and lower bounds. In addition, a generalized noise power constraint was imposed which required that the sums of integrals of the components of A N ( a ) for r sets in some given partition of these components be equal to known values Q,, Q 2 ; . . , Q r . The results given in [98] for this multi-input matched filtering problem form a useful exten- sion of the scalar results we have mentioned above.

The frequency-domain formulation of the matched filter- ing problem is useful for large (infinite) observation inter- vals. A direct time-domain approach in modeling the signal and noise uncertainties is desirable when the observation interval is some finite or semi-infinite interval [ to, t , ] with one or both end-points finite. In this case, the matched filter weighting function (impulse response) is the solution to an integral equation. In the finiteinterval discrete-time case the SNR functional and the equation for optimum filter weighting function are given in terms of matrix and vector quantities. Thus for this situation the robust matched filtering problem is amenable to simpler mathematical anal- ysis. By viewing the finite-length discrete-time matched filtering problem as a multi-input (time-domain) problem it is seen that the ideas for the models for signal and noise quantities in the multi-input frequency-domain case are directly applicable. This approach has been followed in [97], [98] where the signal vector s = (s,, s,; .., s,) of samples s, and the noise covariance matrix R, are modeled as belonging to generalized d, and bounded eigenvalue un- certainty classes, respectively. A slightly different situation is considered by Kuznetsov in [99]. In [IOO] a more general approach has been taken by Verdh and Poor, and three signal vector uncertainty classes are considered (!,, t,, and dm classes defined as neighborhoods around a nominal signal vector). A general class of noise covariance matrices, defined in terms of matrix norms, has also been studied in [IOO]. The general approach in [IOO] is based on a Hilbert space formulation of the matched filtering problem, which

448 PROCEEDINGS O F THE IEEE, VOL. 73, NO. 3, MARCH 1985

Page 17: Robust Techniques for Signal Processing - A Survey

we shall discuss in Subsection Ill-C. There we will also see how the time-domain continuous-time version of the prob- lem may be approached. In fact, the finite-time-interval continuous-time robust matched filtering problem was originally considered by Kuznetsov [I011 even before his frequency-domain problem formulated in [9l] . In [ lol l Kuznetsov used a signal uncertainty class in which the possible deviation 8 ( t ) of a signal waveform s ( t ) from a nominal signal s o ( t ) i s bounded in its integrated squared error over the interval. Further aspects of this problem have been considered more recently by Burnashev [102]. A simi- lar model was also used for the uncertain noise covariance function, and in addition uncertainty in the mean value of the noise was also taken into account. Subsequently Aleyner [I031 modified the signal uncertainty class by imposing a signal-energy equality constraint, a result which was later extended to the envelope uncertainty case for bandpass signals in [104].

A variant of the robust matched filtering problem is that of optimal nominal-signal selection to obtain the best pos- sible minimax performance. This situation has recently been considered by Verdir and Poor in [I051 for the finite-length discrete-time situation. There the actual signal s is assumed to be a nominal signal so with an additive vector 8 which can belong to some class of possible vectors. For given so the robust matched filter can be obtained and the minimax performance level can be determined. With so allowed to be any nominal vector with fixed total energy, the choice of so maximizing this guaranteed performance level is ob- tained in [I051 for signal uncertainty 8 in t,, t,, and t, neighborhoods of the zero vector. While the result for the e, uncertainty class indicates that so should be the mini- mum-eigenvalue eigenvector of the noise covariance matrix (a similar result was obtained independently by Kuznetsov [%I), this classical solution does not hold for the other uncertainty classes.

Before considering two specific applications in the next subsection, we mention that recently a maximin sonar sys- tem design problem has been considered by Vastola, Farnbach, and Schwartz i n [ l a ] , in which the signal and detector pair are picked to maximize the worst case perfor- mance of the system over a class of possible reverberation scattering functions.

B. Two Examples of Robust Multisensor Systems

We will now describe briefly two specific applications of these results on robust matched filters. Our first application is a narrow-band spatial array system in which the individ- ual sensors or antenna outputs are given complex (ampli- tude and phase) weights to detect signals arriving from any particular direction. While spatial matched filtering has pre- viously been considered for this case [107], here we will be concerned with uncertainties in some of the signal and noise characteristics. It has been found that the use of minimax robust weights leads to another significant ad- vantage in such a system. The second application we will consider is that of a system for estimating the time delay between the random signal arriving at two sensors, with the signal observed in independent additive noise at each sensor. While this is not directly a signal-detection problem involving matched filters, we find that in at least one approach to optimum system design the mathematical anal-

ysis is closely related to that performed in obtaining the matched filter in a deterministic signal-detection problem.

I ) Robust Spatial Arrays for Signal Detection and Loca- tion: Consider a /-element narrow-band linear array, il- lustrated in Fig. 14. The signals received at the array ele- ments from any source are time-delayed (or phase-shifted) versions of the source signal. Let x i be the position of the

FOf-feld swrce 0 1 8, rmvelength k

I ' freanncv

i &ray outpu1

Fig. 14. Narrow-band spatial array for signal detection and location.

i th array element, measured from an arbitrary origin. Let X be the wavelength of the source signal, and suppose a far-field source is in some direction 0 from the array broad- side.

To detect a signal from the direction 8 , a phased-array system uses a set of complex weights or phase shifts h,,(B) = exp(-j2.rrxj sinO/X) to "line-up" in phase the signals received at each element. More generally, when sources of interference are present in specified directions and the observations at each element are noisy, one may design the weights h j ( 8 ) to maximize the SNR at the array output due to a signal in direction 8. Now it can be shown that the generalized / X 1 noise covariance matrix R , for K inter- fering sources at locations el, 0,; . ., 8, is, element-wise,

K RN( i , e) q8,, + R k & 2 n ( x , - x ~ ) s i n e k / X (3.12)

k = l

where M.: i s the white-noise level at the i th element, S,, is the Kronecker delta, and R, is the power of the kth interfering source. For a source in direction 0 the nominal signal complex envelope at the i th element is

so, - - e /2nx , slne/X (3.13)

assuming a normalization to unit signal amplitude at each array element. The matched-filter array weight vector maxi- mizing output SNR is

ho( e) = RN1e( e) (3.14)

where so(0) = [sol(@), sO,(8);. ., s0,(8)]'. This weight vec- tor becomes the phased-array weight vector with compo- nents h,,(B) = exp( - j 2 n x , sin @/X) (with amplitudes nor- malized to unity) when there are no interfering sources present. For the general case of R , given by (3.12), the spatial matched filter will have unequal amplitude weights

KASSAM A N D POOR. ROBUST TECHNIQUES F O R SIGNAL PROCESSING 449

Page 18: Robust Techniques for Signal Processing - A Survey

and a set’ of phase weights different from that for the phased-array case.

Our interest is in the case where there is some uncer- tainty in the characterization (3.13) of the signal compo- nents. This may arise because of imperfect propagation characteristics, element position uncertainties, or element gain variations. To model such uncertainties we use the e, class of possible signal vectors s(e) which satisfy

While the covariance matrix R, may also be taken to be uncertain, we will assume here that it is known exactly. If the are all equal to some common value W it can be shown that the eigenvectors of R, are independent of W. Moreover, (/ - K ) eigenvalues of R, are equal to Wand the others are of the form W + uj where the ui are inde- pendent of W. Such properties of R, may be used in enlarging the class of allowable noise covariance matrices, although we shall not do so here.

General results for finite-length discrete-time matched filters discussed below can be applied here to obtain the robust spatial matched-filter weights as

= ( R, + e) (3.16)

where u,, i s a constant depending on A which effectively adds uncorrelated noise components to the array elements, and s,(B) is the least favorable signal vector.

Table 1 shows numerical values for the output SNRs using the nominally optimum weights and the minimax robust weights for A = 2 and 4 for the following system:

0 signal direction = 0’ / number of array elements = 12

K number of interferers = I R , power of interference = 1.0 0, direction of interference = -60’

white-noise levels, given by

element spacing = x/2

(0.25, i = 1,2,8,12 i = 3,4,9,10,11

2.0 i = 5 \O.l; i = 6,7.

Table 1 Signal-to-Noise Ratio of Array Output Using Nominal and Robust Matched-Filter Weights

As expected, the robust weights hR(B) give a worst case performance (for s , ( B ) ) which is better than the corre- sponding performance of the nominally optimum weights h,(B) = R;’@(e). Note that the SNR is a function of the filter weights, the signal vector and the noise covariance matrix R,, and that for A = 0 the vector hR(0) is the nominally optimum vector ho(B).

The performance shown in Table 1 does not show a dramatic advantage (or disadvantage) obtained by using the robust spatial filter, in terms of SNR performance. However, there is an important aspect of performance of such an array system which is not captured by examining the output ‘ SNR in the correct “look” direction 8 only. As the array scans for targets by adjusting its weights for different angles 8 the output SNR becomes a function of the “look” angle 0 for any fixed source direction 8 = 0,. Alternatively, a beam pattern for the array may be defined as

(3.17)

the array output signal magnitude for a nominal signal in direction 0 with array weighted to ”look” in the direction e,. Ideally B(e,e , ) should have only a narrow peak at 0 = e,, as a function of 0 for fixed e,.

Figs. 15 and 16 show (f.or the numerical example giving Table 1) that the robust spatial matched filter weights produce normalized beam patterns that have a distinct

0

-10

g-20

0 I m-30

-40

-50-

sin e Fig. 15. Beam patterns of nominal ( A = 0) and robust 2) matched filter spatial arrays.

1 , k,?: , o f , IT,“:, , , , , ‘1 , ols , 1 -50-1 -08 -OB -a4 -0.2 o 0.2 0.4 o

sin e Fig. 15. Beam patterns of nominal ( A = 0) and robust 2) matched filter spatial arrays.

Sin 6 Fig. 16. Beam patterns of nominal ( A = 0) and robust ( A = 4) matched filter spatial arrays.

advantage over the pattern of the nominally optimum array. The main beams of the robust array weights are narrower and their close-in sidelobes are appreciably below those of the nominally optimum array, a factor important in array design. Thus the use of a robust weighting scheme pro- duces a ”side” benefit which probably exceeds in impor- tance i ts original justification of maintaining the output

450 PROCEEDINGS O F THE IEEE. VOL. 73, NO. 3. MARCH 1985

Page 19: Robust Techniques for Signal Processing - A Survey

SNR under uncertainties. The robust weights differ from the nominally optimum weights as a result of the addition of the matrix a,/ to R,, and this suggests that as a general procedure a consideration of different values of a, in array weight design can lead to ”optimum” tradeoffs between output SNR and beam characteristics under nominal condi- tions.

Notice that the use of A values such as A = 2 implies that allowance is made for up to two “dead” elements in the array. As an extension of the signal class modeled by (3.15), one can consider a generalized class defined in terms of several tolerances A,,A2; .., A q for q groups of sensors. Further aspects of the performance of a robust spatial array system are discussed in [97], [%I. Before leaving this subject we note that recently Ahmed and Evans [I081 have consid- ered robust narrow-band array processing under uncertain- ties, although their definition of robustness is based on the notion of an acceptable set of performances rather than on optimizing worst case performance.

2) Robust Eckart Filters for Time-Delay Estimation: The estimation of time delay between signals arriving at two spatially separated sensors is of interest in applications such as seismology and sonar. In such systems, the time-delay measurement gives information about the direction of arrival of a wide-band source. Let Y,( t ) and VI( t ) be the outputs of two sensors, these signals being described by

v,( t ) = 5( t ) + Nd t ) (3 .I 8a) V2( t ) = S( t - D ) + N2( t ) . (3.18b)

Here S ( t ) i s a random source signal and N,(t), N2( t ) are uncorrelated additive-noise processes at the sensors. The basic technique for estimating the unknown relative delay D is to cross-correlate Y,( t ) and V 2 ( t ) and to use as an estimate of D that time argument which gives the maxi- mum value of the cross-correlation function.

To improve the estimation process one can use a filter to weight the cross-spectrum estimate. Various criteria for optimum choice of the filter have been proposed [109]. One particular performance measure is the output SNR at the correct time delay; for weak signals (low-input SNR) this output SNR for long averaging time is [I091

where W(w) is the real filter frequency response, @,(a) i s the power spectral density of the zero-mean and stationary signal process, and Q(w) = @$(w), the square of the noise power spectral density at each sensor. The noise processes are here assumed to be zero-mean and stationary with identical power spectral densities, and signal and noise processes are assumed to be uncorrelated. The filter func- tion W,(o) maximizing the above SNR is given by

(3.20)

which is the Eckart filter. Comparison of (3.3) and (3.4) with (3.20) and (3.19) shows the correspondence between the matched-filter maximizing output SNR in the detection of a deterministic signal and the Eckart filter maximizing output

SNR at the correct time delay for long observation time under weak signal conditions in time-delay estimation. The main difference lies in the fact that aS(w) is a nonnegative power spectral density whereas P ( w ) in matched filtering was a Fourier transform of a finite-energy signal.

The implementation of an Eckart filter requires knowl- edge of Qs(o) and Q(o). These are often not precisely known but may be modeled as belonging to uncertainty classes such as those considered for power spectral densi- ties in Section II. As an example, consider the ppo in t classes for both aS(o) and Q(w), which specify the frac- tions of the total powers of @,(a) and Q(w) in specified intervals partitioning the entire frequency spectrum. Such information may be obtained from measurements at the output of a simple cross correlator when signal is present, and from output power measurements in the frequency bands of interest under noise-only conditions. Since our earlier results on the minimax robust matched filter were obtained specifically for the L , signal uncertainty class, these results are not directly applicable here. However, it is quite simple to show [I101 that for the time-delay estima- tion problem where SNR of (3.19) is the performance mea- sure of interest, the minimax robust Eckart filter for ppoint classes for QS(w) and Q(o) has a piecewise-constant frequency response. The values of the constant levels of the robust Eckart filter frequency response W,,(o) are de- termined simply as the ratios of the given fractional powers in @,(w) and Q(u) in the common frequency bands de- fining the p-point classes (as in Example 3 of Subsection 11-C).

As an example, assume that the nominal signal power spectral density Qs,o(w) and the nominal noise power spec- tral density C J ~ , ~ ( W ) are given by

@N,O( = @5,0( (3.21) and

For A = 1 (input SNR = 0 dB) the nominally optimum Eckart filter W,,(w) results in an output SNR of 277 or 8 dB. Now Qs,o(o) and @&,o(o) with a = 1/3 are members of particu- lar ppo in t classes with two frequency bands [0,1.61] and [1.61, w). (The value 1.61 of the frequency-band edges is the one which maximizes the performance of a two-level piecewise-constant filter under nominal conditions.) Sup- pose the noise power spectral density actually is a constant c on [O,w,] and zero outside. Now we may pick the values of c and o, to make @$(w) a member of the ppoint class to which Q$,o(u) belongs. In this case we find (with A = 1, as before) that the input SNR is now - 3 dB. The robust Eckart filter for this situation i s the optimum two-level piecewise-constant filter for the nominal situation. I ts SNR in the nominal case is 2.76 or 4.4 dB, and it maintains this value of the SNR for all Q(u) in the ppoint class defined above. The nominally optimum Eckart filter’s output SNR drops to the value of 0.01 or -20 dB when the noise has the above ideal low-pass spectrum in the ppoint class. Simulation results [I101 confirm that the robust two-level piecewise-constant Eckart filter works quite effectively to

KASSAM AND P O O R ROBUST TECHNIQUES F O R SIGNAL PROCESSING 451

Page 20: Robust Techniques for Signal Processing - A Survey

obtain a good estimate of D for all Q(o) in the defined ppo in t class, while the nominally optimum filter is quite sensitive to deviations form the nominal assumption.

More general results for other convex classes of total- power-constrained allowable power spectral densities as(@) and Q ( w ) = @k(o) have been discussed in [IIO]. Although the SNR performance criterion of (3.19) is very similar to the SNR measure of (3.4) for deterministic signal detection, the minimax robust Eckart filter is the optimum filter for a least favorable pair @ s , L ( w ) and QL(u) which is found exactly as in the case of robust Wiener filtering in Section II. This follows from the general observation made in Section I1 concerning least favorable pairs whenever the performance measure is of a type expressible as a "distance" measure between aS(u) and Q(w) for convex classes of power spectral density or probability density functions [Il l] .

Before leaving this particular application let us note one further interesting feature of the robust Eckart filter. As in the case of the robust spatial matched filter we have con- sidered above, there is another aspect of the performance of a time-delay estimation system that we have not consid- ered explicitly. This is the variance of the time-delay esti- mate. It turns out that the robust Eckart filter has the additional advantage of generally producing time-delay estimates with considerably lower estimation variances than the optimum Eckart filter under nonnominal conditions, with almost similar performances for both under nominal conditions. This phenomenon is discussed further in [IIO].

C. General Formulation of Robust Matched Filtering Problems in Hilbert Space

The results we have discussed so far in this section serve to illustrate the ideas behind robust matched filtering and their possible applications, and can be extended in several ways as we have indicated. In this subsection we will consider a general formulation of the robust matched filter- ing problem which has recently been developed in [112], using a Hilbert-space framework. Most matched-filtering situations (e.g., continuous-time/discrete-time, one- dimensional/multidimensional, time-domain/frequency- domain) can be fit into a single general Hilbert-space setting which is convenient for studying robustness in all of these problems simultaneously. In particular, suppose X is a separable Hilbert space (e.g., I , or R") with inner product ( 1 , .) and let 2 denote a set of bounded non- negative linear operators (e.g., integral operators, matrices, or spectral operators) mapping X to itself. A matched-filtering problem on X involves three quantities: a signal quantity s E X (eg, a signal spectrum or wave- form); a noise quantity n E 4 (e.g., a noise spectrum, autocorrelation function, or covariance matrix); and a filter quantity h EX (e.g., a filter transfer function or impulse response). The design criterion for the matched filtering problem is based on a functional p: X X 2% 24 R de- fined by

and representing an SNR. Note that, for properly defined X , most conventional matched filtering formulations fit this model (see, for example, Thomas [39]). The example described in Subsection Ill-A corresponds to the particular case in which X i s complex 12( - 00, w), i.e., X i s the set

of complex-valued functions f satisfying

' -m

and 4 corresponds to a set of positive, symmetric, real- valued functions. The signal quantity s is identified with the shifted-pulse Fourier transform P(w>e'"'; the filter quantity h i s identified with H * ( w ) , the complex conjugate of the filter transfer function; the noise quantity n corresponds to the noise spectrum aN(w); and for an element n E 2 and any f E X , the operation nf i s defined by ( n f ) ( w ) = n ( w ) f ( w ) . Note that the inner product on I,( - 00, 00) is defined by

(f,g) =lJx f*(w)g(w)do. 2n -" Examples of other matched filtering problems that fit this general Hilbert-space formulation are discussed below.

Within this general Hilbert-space formulation, for fixed signal s and noise quantity n the matched filter (i.e., the element h, E X that maximizes p ) i s given by any solution to the equation nh, = s. i f n is invertible then h, is n - l s and the maximum value of p is given by (s, n-'s). If, on the other hand, s and n are known only to be within classes Y and JV of signal and noise quantities, respectively, then we can consider the alternative design criterion (as in (3.6) and (3.9))

,ax( inf p ( h ; s , n ) ) . (3.24) h E X ( s, n) E 9 X X

It can be shown (see [112, Lemma I n that if 9 and X are convex, then a pair (s,, n,) E Y X X and it optimum filter h, (satisfying nLhR = s t ) is a saddlepoint solution for (3.24) if and only if the following inequality holds for all (s, n) E Y X X :

Moreover, i f nL is invertible then this occurs if and only i f ( s L , nL) i s least favorable for matched filtering for Y and M, i.e., i f and only if

(s,,n;'s,) = min (supp(h;s,n)). (3.26) ( s , n ) E y x X h c X

Expression (3.25) provides a means for checking potential sol- utions to (3.24), and (3.26) provides a means for searching for such solutions since the expression in brackets in (3.26) i s known in closed-form for many situations of interest.

By using the above results, solutions to the general robust matched filtering problem of (3.24) have been obtained for generalizations of the uncertainty classes gP and VN of (3.5) and (3.8a and b), as well as for other uncertainty classes of interest. For example, consider the situation in which the noise quantity is known to be some fixed no E 2, but the signal quantity s is known only to be in the class Y: E 2' defined by

(3.27)

where so is a known nominal signal, A is a fixed positive number representing the degree of uncertainty in the sig- nal, and llxll denotes the norm of x E X defined by (1x11 = [ ( x , ~ ) ] ' ' ~ . For this problem it can be shown (see [112, Theorem I]) that a saddlepoint solution to (3.24) i s given by (h,; s L , no) where h, i s given by

h, = ( n o + qJ)-ls, (3.28)

452 PROCEEDING O F THE IEEE. VOL. 73, NO. 3, MARCH 1985

Page 21: Robust Techniques for Signal Processing - A Survey

with I being the identity mapping from 2‘ to itself and with a, being the positive solution to

u i l l h R l 1 2 = A (3.29) and where

sL = S, - u0hR. (3.30)

Note that, for the specific spectral-domain example dis- cussed in Subsection Ill-A, the identity operator is repre- sented by the unit white-noise spectrum aN(w) = 1 and so (3.28) corresponds to (3.7) with no and so being repre- sented by a N ( w ) and fO(w)e+IuT, respectively. (Recall that, for this case, h i s identified with HI(w)) . It is interesting to note that, as in the spectral case, the identity operator / generally describes a white-noise process, so that (3.28) indicates that the type of uncertainty described by y: has an effect on the design equivalent to that of adding white noise of spectral height ’JO.

The result regarding the band model of (3.8a and b) can also be extended to general Hilbert spaces. In particular, consider the situation in which the signal is known, say so, and in which all of the noise operators n E M can be represented by spectral components { n(w); w E Q} for some set a. This situation arises in stationary models such as that yielding (3.8a and b), in which case n ( w ) can be taken to be the power spectral density at the frequency w and Q i s R (it could similarly be R”). Alternatively, for other models, this situation arises when all of the members of M share a common eigenstructure and the set { n(o ) ; w E a } is the eigenspectrum of the operator n. For example, such a model is generated in L, if the all noise autocorrelation functions in M have Karhunen-Loeve expansions in the same eigenfunctions. In this latter case, would be the set of positive integers, and n ( w ) would be the oth eigenvalue of n.

In this general setting the band model (3.8a and b) becomes

sp, = { n(n’( w ) < n( w ) < n”( w ) ,

V U E Q and t r { n } = c } (3.31)

where n’ and n” are known functions and t r { .} denotes the trace operation. Note that for power spectral densities the trace is just

and for discrete eigenspectra the trace is

c 4 ; ) . m

i-1

(More generally, we can write

t r { n ) = j n ( o ) p ( d 4 D

for some measure p on 3.) Assuming that the signal so also has the spectral representation { S o ( w ) ; w E a } in terms of the same eigenstructure as the members of Nn, it can be shown using (3.25) (see [112]) that the robust filter forNn of (3.31) is the optimum filter for so and a least favorable noise operator whose spectrum is

n L ( w ) = max{ n’(w),min{ ~ ’ ~ ~ , ( o ) l , n ” ( o ) ) }

(3.32)

where k i s chosen so that tr { n L } = c. Note that (3.32) i s

identical to (3.11) with nL = @ N , L , L, = n‘, U, = n“, and p0 = 5,. This result can be combined straightforwardly with (3.28)-(3.30) to give the robust solution for uncertainty in both signal and noise (see Kassam, Lim, and Cimini [92] and Poor [I1 21).

Example: Binary Communication: To illustrate the gen- erality of the above result consider the problem of anti- podal signaling in additive Gaussian noise as described by the following pair of statistical hypotheses:

versus

H,: Y( t ) = N( t ) + s( r ) , o G t G T (3.33)

where { ~ ( t ) ; 0 < t < T } is a known (i.e., deterministic) square-integrable signal waveform and { N( t); 0 < t d T } i s a zero-mean Gaussian noise process with autocorrelation function { RN( t,u); 0 < t < T , O < u < T } . Assuming that H, and H1 are equally likely, the Bayes optimum (i.e., minimum-probability-of-error) receiver for (3.33) is of the form (see, e g , Helstrom [113])

(3.34)

where sgn { . } denotes the algebraic sign of the argument; + = + I and + = -1 denote the acceptance of hypotheses Hl and H,, respectively; y = { y( t) ; 0 < t < T } denotes the observed realization of the random process { Y( t) ; 0 t < T } , and where { h( T, t ) ; 0 t < T } denotes the impulse response of a linear filter. The probability of error associ- ated with the receiver of (3.34) is given by

f‘ = 1 - @([SNR]”’) (3.35)

where SNR denotes the signal-to-noise ratio at the output of the filter at time Tand is given by

Here @ denotes the standard (unit) Gaussian probability distribution function. The optimum receiver for (3.33) i s given by (3.34) with h( T , t ) being the solution to the in- tegral equation

This problem fits within the above Hilbert-space formula- tion with &’ being real L,[O, TI, s = { s ( t ) ; o < t < T } , h = { h ( T , t ) ; O < t < T } , and n = {R,(t,u); ( t , u ) E ~ ~ , T 1 2 } . W i t h a = { a ( t ) ; O ~ t ~ ~ } a n d b = { b ( t ) ; O ~ t ~ T } in 2, we have

( a , b) = j r a ( t)b( t ) dt 0

and the operation na is defined by

SO, p(h; s, n) is given by (3.36), and (3.37) is the equation nh = s. For this model, the signal uncertainty class y: of (3.27) i s given by

(3.38)

KASSAM A N D P O O R : ROBUST TECHNIQUES F O R SIGNAL PROCESSING 453

Page 22: Robust Techniques for Signal Processing - A Survey

where so = {so( t ) ; 0 Q t Q T } is the known transmitted signal. This model represents a general model for types of receiver and channel distortion that are difficult to model parametrically. Further justification for this model is found in Slepian’s notion of indistinguishable signals [114].

From (3.28) and (3.29) it follows that the robust receiver of the form (3.34) for uncertain signal distortion described by (3.38) is given by the impulse response solving the equation

j o r R , ( f , u ) h , ( T , u ) d u + g h , ( T , f ) = s , ( f ) , o < t < T

(3.39)

which is a Fredholm equation of the second kind. The constant u, i s specified by (3.39) and the condition

t$/\h,( T,t)I2 dt = A . 0

The performance of specific solutions to (3.39) i s discussed in [112]. It is interesting to note again that the effect of the uncertainty modeled by (3.38) i s to introduce a white-noise ”floor” of height u,. This device is used in standard treat- ments (see, e.g., Van Trees [I]) in order to circumvent singularity problems arising in the solution to (3.37) for the case of continuous R,. That this phenomenon arises natu- rally here gives additional justification for considering minimax design within (3.38). Equation (3.39) and the con- dition following it for the robust filter was originally de- rived in [IOI]. A modified condition for a signal class with an energy equality constraint is given in [103].

To further illustrate the general Hilbert-space results, consider the discrete-time observation models under the two hypotheses

H,: L; = N; - s i , i = 1 ,2 ; . . ,m

and

HI: L; = N, + s,, i = 1 ,2 ; . . ,m (3.40)

where N = (N, ; . . , N,)’ is a random vector having zero mean and nonsingular covariance matrix R, and where s = ( 5’;. ., s,)’ is a known signal vector. Here, the output at time m of a linear filter with impulse response {h; , i = 0,1 ; . . ,m - I } is given by hTs where h, = h,-,, and the output SNR is ( l i5 ) ’ /hTRNh. The matched filter for known R, and s is given by RG’s. Of course, this fits the above Hilbert-space formulation with &’= R” (i.e., ( a , b ) = aTb), s = s, h = h, and with the operator n represented by the matrix R,. A band model class JV of operators such as (3.31) occurs here in the case in which all members of JV have the same orthonormal eigenvectors 9, y; . - , v , in which case (3.31) i s given by

R,IR,y = A j y and X; Q X; Q X,!,

forall i=l;..,m,and EX,= rn

i-1

Thus applying (3.32), the least favorable covariance matrix is given (using its spectral representation) by

rn

RN,L = c XL,,v;YT (3.42) i-1

with

A’; I K ’ p ; A’, k-lp;, A‘; Q k-lp; < X; (3.43)

x:l ‘ A’,! < k-lp;

where p; = IsTyI and where k is chosen so that m

A L , ; = c. i = l

Note that (3.42) must be nonsingular and thus the robust filter is described by h, = R ; l L ~ . This result is a special case of more general results found in Chen and Kassam [97], [%I.

The discrete-time model of (3.40) can also be used to illustrate least favorable signals and noise operators for uncertainty models of interest other than those of (3.27) and (3.31). This problem has been treated in [ IO] in some detail. For example, suppose the noise covariance is known to be given by diag { uf, uj’; . ., u:} (corresponding to un- correlated samples), and the signal is known only to differ in total absolute distortion by no more than A from some nominal signal so, i.e., we assume the signal lies in the class

rn 3’ = s E R” 1s; - SO,,[ < A . (3.44) ( Ij=l 1

Then, it can be shown [ IO] using (3.25) that the robust filter h, is given by

where h, is the nominal filter and where c > 0 satisfies the equation

rn u: min {O,lho,,l - c } = A .

i-1

Thus in this case the robust matched filter is a clipped version of the nominal filter. Similarly if R, is diagonal and s lies in a class

(3

then the robust filter becomes (see [loo])

( 50,; - A)/$, A < SO,; - A < so,;< A (3.47)

(so,, + A)/$, so.; < - A .

This filter is the optimum filter for a least favorable signal in Sq“ that is as near zero as possible (in terms of the norm max,,l,2,,,,,,IsjD. Extensions of these results can be found in 11 0 1 .

The above treatment makes clear the rather general appli- cability of the Hilbert-space framework for formulating and solving robust linear detection problems. Recently it has been shown [I151 that a similar Hilbert-space framework can also be used for robust linear estimation problems (which we surveyed in Section 11); for details the reader is referred to [115]. We shall now proceed, in the next section, to consider robust nonlinear signal detection schemes pri- marily designed to protect against uncertainties in the am- plitude statistics of noise.

454 PROCEEDINGS O F THE IEEE, VOL. 73. NO. 3, MARCH 1985

Page 23: Robust Techniques for Signal Processing - A Survey

Iv. NONLINEAR METHODS F O R ROBUST SIGNAL DETECTION justification it provided of the use of detectors based on

In the previous two sections we focused on the design of robust linear filters for signal estimation and detection, for situations in which there was uncertainty about the spectral densities or correlation functions of the signal and noise. The performance measures considered there were the M S E and the SNR, which did not involve the exact functional forms of the signal and noise pdfs.

In this section we will survey results on robust signal detection which pertain specifically to robustness when the detection performance is measured by characteristics di- rectly related to the probability of detection or error prob- abilities instead of SNRs. Although in most situations an assumption that the noise is Gaussian gives a direct rela- tionship between the SNR and such detection performance measures, this is not true in the case of non-Gaussian noise. In almost all such cases it is possible to obtain explicit results on the structures of minimax robust detectors (gen- erally nonlinear) only when the noise processes are se- quences of independent random variables so that only univariate pdf uncertainties need be considered. In the first two sections we were able to avoid considerations of pdf uncertainties because the performance measures (MSE and SNR) depended only on second-moment characteristics. For the detection performance measures we use here, consider- ation of the correlation functions and more generally of pdfs beyond those of first order is avoided by the assump- tion of independence. For the pulse-train detection prob- lem discussed at the beginning of Section Ill, this means that a robust detection scheme can be arrived at by decov pling the treatments of spectral density and probability density uncertainties. That this approach is necessary is not surprising, considering the well-recognized difficulties of dealing in general with non-Gaussian random processes. There are available, nonetheless, some recent results on robust detection in correlated noise, under certain con- straints, and we will discuss these later in this section.

For the most part, then, we will be considering here various nonparametric classes of univariate noise pdfs ex- pressing different types and extents of uncertainty about the exact noise pdf. We will consider in particular the canonical detection problems of known low-pass signals in additive noise, deterministic bandpass signals in bandpass additive noise, and random signals in additive noise. Even with the simplification due to the white noise assumption, explicit solutions for minimax robust detectors are generally possible only under a further restriction to consideration of the local or weak-signal case. We will begin, however, with a description of the results of Huber, and other related results, for a general hypothesis testing problem. As we have mentioned in the Introduction, the 1964 and 1965 results of Huber [6], [7] have greatly influenced and moti- vated much of the subsequent work on robust signal processing schemes.

A. Robust Hypothesis Testing

In 1965 Huber [7] published an explicit solution for the robust test for a binary hypothesis testing problem related to the signal-detection problem we discussed in the Intro- duction. The significance of this result lies not so much in the solution it provided for the hypothesis testing problem considered by Huber, as much as in the mathematical

bounded functions such as i ( x ; s, 0 ) of (1.5). Let X = (Xl, X,; . . ,X, , ) be a vector of independent and

identically distributed (iid) observation components. Under a hypothesis H, (null hypothesis) let the common pdf of the X i be f,, and under an alternative hypothesis HI let the common pdf be f,. The requirement is to construct a test for H, versus HI based on the observations X , when f, and f, are not specified completely. The approach taken by Huber in [7] was to define first classes of allowable pdfs under the null and alternative hypotheses. The classes con- sidered in [7] were obtained as neighborhood classes con- taining in each case a nominal density function and density functions in i ts vicinity. One such pair of neighborhood classes which is popular in robustness studies is the pair of c-contamination classes, for which under H,,j = 0,1, the class of allowable density functions is

( r;O,c,) = { f l f = (1 - c j ) co + € ; h i } , j = 0,l.

(4.1) Here $' is the nominal density function under hypothesis H,, quantity c, in [0,1) is the maximum degree of con- tamination for $', and h, i s any density function. These classes are, of course, the same (aside from normalization) as the spectral uncertainty classes of Example 1 in Subsec- tion Il-C. The hypothesis testing problem is therefore that of choosing between the two hypotheses?

H,: the pdf of the X , is any pdf f, in So,

H, : the pdf of the X ; is any pdf f, in 9,. (4.2a)

(4.2b) Huber then sought the least favorable pair of probability densities in 9, x SI, which is defined on the basis of a risk function R( f ,+) . In the risk function, $ denotes the test for H, versus HI which rejects Hj in favor of HI-, with probability # ( X ) when X = (Xl , X , ; . ., X , ) is observed: the X ; being iid with density function f in 9, or 9,.

Consider, for example, the minimum probability of error criterion for which we would have

R( f ,+ ) = E , { +I( X ) } , if f E 9. (4.3)

Then for ( f,, f,) E 9, X 9, the probability of error (assum- ing equally likely priors) is

and this is minimized when $I is a test based on a compari- son of the likelihood ratio

n

A ( x ) n f l ( X i > / S ( X i ) i-1

with a threshold value of unity. The least favorable pair (9,,q1) in 9, X F , for this problem is that pair for which the corresponding test +q based on its likelihood ratio satisfies, for all (6, f,) E 9, X 9,

'The arguments for Fo,Fl will be dropped when no confusion

'Note that + ' ( X ) + # ( X ) = 1. can result.

KASSAM AND POOR: ROBUST TECHNIQUES F O R SIGNAL PROCESSING 455

Page 24: Robust Techniques for Signal Processing - A Survey

One of the major results in [7] is that for the c-contamina- tion classes a least favorable pair (q,, q,) exists satisfying

R ( f , + q ) d R ( c ~ ~ , + ~ ) , for f ~ q , j = 0 , 1 (4.6)

where (Pq is any test for 9, versus q1 based on a compari- son of the likelihood ratio with a threshold (a "probability ratio test"). This clearly implies that (4.5) i s true and also gives similar results for other risk-based criteria. The pair (qO,q1) and +q for minimum error probability form a sad- dlepoint for the error probability functional, that is

f e ( f O t f l ; + q ) 6 pe(q0 t91 ;+q ) f e ( q O , % ; + ) (4.7)

for any (h, f,) E So X Sl and any test +. The test +q is called a robust test for f E So versus f E sf; it minimizes over all tests the supremum (least upper bound) of the error probability over all pairs in So x SI. Indeed +q is robust, in view of (4.6), for other risk-based criteria such as the Ney- man-Pearson criterion. For this criterion we can define the threshold for +q in such a way that for a design value a of the false alarm probability

R ( f , a ) d R ( q 0 , $ ) = a , f E 6 (4.8)

For the r-contamination class Huber's solution for the with R( f ,+) as in (4.3).

least favorable pair turns out to be

I c" L ( 1 - c O ) f f ( x), otherwise

(4 .sa)

\?(I - cl) ho( x), otherwise

where c' < c" are nonnegative numbers such that qo and q, are pdfs. The proof of the existence of such a pair in So X PI where Po and Sl are disjoint (the case of inter- est, since otherwise qo = q,), and the proof that (4.6) holds, can be found in [7]. Note that the likelihood ratio h q ( x ) = 9,(x)/q0(x) for a single observation for the least favorable pair is a "censored" version of the nominal one,

bc" , c" d X,( x)

Xq(x) = bXo(x), c' < Xo(x) < c" (4.10)

[bF, X,(x) d c'

where b = (1 - cl)/(l - c 0 ) and ho(x) = ff(x)/r,"(x). In [7] the least favorable pair of density functions was also

obtained for another uncertainty model for the density functions under H, and HI. This was defined by the total- variation classes of probability densities which may be expressed as

~ ( < o , c ) = ( f ~ ~ ~ f ( x ) - < o ( x ) l d x ~ ~ i . (4.11)

This is analogous to Example 2 of Subsection Il-C. In a later paper [73] i t was shown by Huber and Strassen that least favorable pairs of probability measures can be found for probability measure classes which are defined as being

bounded by 2-alternating capacities, which are generalized notions of measure as discussed in Subsection 11-G. Other specific uncertainty models have been considered by Khalfina and Khalfin [116], Rieder [74], and Bednarski [75].The <-contaminated classes of nominal densities have remained the most widely used uncertainty models, partly because of their earlier introduction, but primarily because it is possible to justify their use, in many cases, from physical considerations (see, for example, Trunk [117]). In modeling impulsive noise, for example, the presence of a small proportion (whose maximum value is E) of impulsive components in a background of, say, Gaussian noise can be modeled to have an c-contaminated nominal Gaussian pdf. Any uncertainty about the exact pdf of the impulsive com- ponents then leads directly to use of an c-contaminated uncertainty class, perhaps with a side constraint on the nature of the contaminating pdf h. Models of this type have been studied by Miller and Thomas [118].

One other specific uncertainty class of pdfs that has been used more recently is that analogous to the spectral band model discussed in Subsection 11-C; this model is justifiable from physical considerations in some applications. It may be viewed as being more general than the c-contamination class and allows the results obtained for the c-contamina- tion classes to be extended. Specifically the bounded classes q( f i L , C U ) considered in [I191 are defined by

q(CL,Cu) = { f l $ d f d f j u } , i = O , I , (4.12)

so that allowable density functions in q are those bounded by given nonnegative functions C1 and CU which make the 7 non-empty. (These are band models analogous to those defined in Example 4 of Subsection Il-C.) These classes reduce to the c-contamination classes of (4.1) for C1 = (1 - E,)$' and fiu = a. It is not true, however, that the classes (4.12) are obtained from (4.1) by imposing an upper bound on the contamination densities hi. This is clear from the fact that even if we set f l1 = (1 - c j ) G o , the resulting "nominal" density Go may not belong to q( C1, <,,). in [I191 the least favorable pair in the above classes has been found, and the likelihood ratio for the least favorable pair is also essentially a censored version of either f l l / f & or f lu /ku . The bounded classes of pdfs defined by the band model in (4.12) can be viewed as arising naturally in situations where the pdfs under the hypotheses are estimated from training data, in which case the bands ( C1, $,) are confidence bands.

It should be clear that in the varlous uncertainty models considered above for the densities f of each independent component X , of the observation sequence X , there was no necessary restriction that X ; be one-dimensional and that f be univariate. It is possible, for instance, to treat an observation sequence of n one-dimensional components as a single n-dimensional observation, and characterize i ts multivariate density by one of the above classes. Essentially, this approach was taken in [I201 and [I211 by Kuznetsov, who developed results for the characteristics of the robust test for hypotheses described by the bounded classes of (4.12). Since the likelihood ratio in a threshold comparison test may be replaced by any function which yields the same critical region and threshold equality region, simplifications may be made to results of the type in (4.10). In (4.10), for example, for a single observation test with the threshold T

in between bc" and bc', one gets an equivalent test if simply bX,(x) is compared to T . Note that bh,(x) =

456 PROCEEDINGS O F THE IEEE, VOL. 73, N O 3, MARCH 1965

Page 25: Robust Techniques for Signal Processing - A Survey

f , ,(x)/f , ,(x) when the r-contamination class is viewed as a special case of the bounded class.

In [I201 and [I211 Kuznetsov shows that the robust tests are obtained by threshold comparisons using one of the functions f ,Jf, , , f,,/fOu, f lL/ fOL, or f,Jf&,. As an interest- ing application, consider the detection problem (1.1) where 6s = e( s,, s2; . ., s,) is not precisely known, but is known to belong to a neighborhood of a nominal sequence Bos0. Since so is in R” one might define the neighborhood as a spheroid of radius A centered on Boso, as was done in Section Ill. Let the N, be iid and zero-mean, unit-variance Gaussian random variables. A useful bounded class of den- sity functions for X under HI is then defined in terms of the maximum-likelihood and minimum-likelihood esti- mates of Bs,. The robust detector is based on a combina- tion of square-law and linear processing of X; details are given in [121]. The results in [I201 and [I211 on binary hypotheses have recently been extended to multiple hy- pothesis testing problems in [122].

We have remarked that it is possible in general to solve for least favorable pairs of probability measures for other pairs of probability measure classes which are defined to contain measures bounded by lalternating capacities [73]. Specific examples of other pairs of classes for which explicit least favorable pairs are available are the ppoint classes (discussed in Subsection It-C in defining spectral density classes; these become classes of pdfs when the power is fixed at unity), and the bounded ppoint classes [54]. Let us also reiterate the interesting connection between the robust binary hypothesis testing problems for such classes of pdfs and the corresponding robust linear filtering problems with M S E s as performance criteria. For many of the robust linear filtering problems with fixed-power spectral density uncer- tainty classes, the least favorable pairs of pdfs for the robust hypothesis testing problems defined on corresponding unit-power normalized spectral-density classes produce di- rectly the least favorable spectra [55]. In general, one may interpret the least favorable pair of pdfs in Po xF, as being that pair which has the minimum distance between i ts components. Indeed, strong connections exist between distance minimization and least favorable pdf pairs [ I l l ] , [I191 under some general restrictions. It is this fact that results in the close relationship between the solutions for the least favorable pairs in the robust linear filtering and corresponding hypothesis testing problems. Similar con- siderations arise in robustness problems formulated in terms of Chernoff bounds [123], [I241 and for point-process ob- servations [ I 251.

The problem of robust sequential hypothesis testing was also considered by Huber in his first paper on robust hypothesis testing [7]. More recently this work was ex- tended by Shultheiss and Wolcin in [126], which also dealt with the c-contamination model for noise pdf classes under the null and alternative hypotheses. Numerical performance results from simulation experiments and computations are also given in [126].

6. Robust Detection of Known LowPass Signals

Consider again the signal detection problem described by the hypothesis testing situation of (1.1). Here the signal sequence is known to within an overall amplitude factor, that is, the s,, i = 1,2,.. ., n are known. The observations X ,

described by H1 in (l.lb) may be considered to arise as a result of sampling a continuous-time additive mixture of a low-pass or baseband signal waveform 6s(t) and a sta- tionary noise process. If the noise bandwidth is large rela- tive to that of the signal in such a situation it becomes reasonable to assume a sampling rate which results in the noise components N;, i = 1,2;. e , n, under H, or H, being statistically independent random variables with some com- mon univariate pdf f . Another common situation in which the above observation model is appropriate is that arising in the detection of a pulse-train, as described in Section Ill.

In the absence of precise knowledge about the noise pdf f , one can now attempt to obtain a robust detection scheme for a class of possible noise pdfs generated by a model such as the c-contamination model used by Huber. But it should now be apparent that Huber’s solution which we discussed in Subsection IV-A above does not directly apply to our known signal in additive noise detection problem. The least favorable pair of pdfs given by (4.9a) and (4.9b) was ob- tained under the assumption that the pdfs f, €3, and f, €9, are chosen independently; this means that the contaminating pdfs h, and h, in the c-contamination mod- els were not constrained to be related to each other in any particular way. In the known-signal detection problem we should require h, to be a translated or shifted version of h,. On the other hand, Huber’s approach in which indepen- dent contaminating densities are allowed under the two hypotheses can be used to obtain a conservative solution to the robust detection problem.

One of the first attempts to adapt Huber’s approach and results on robust hypothesis testing for signal-detection problems was reported in 1971 by Martin and Schwartz [30]. Martin and Schwartz were interested in the signal detection problem described by (l.l), in which the observations X , under H, are independent but not necessarily identically distributed. They first showed that Huber’s result for the €-contaminated classes extends directly to the time-varying problem, where for each component X , of X the nominal densities f:, f t and contamination degrees rOi,rl, may be different. For the detection of a known signal in nearly Gaussian noise they set 6;. = 9, the zero-mean Gaussian density with unit-variance, and f $ ( x ) = + ( x - 6 5 , ) . They took E,, = c to be sufficiently small for a given 6 so that the resulting r-contamination classes F, and 3, are disjoint for each i, which yielded a symmetric censoring or limiter characteristic. The structure of the resulting correlator- limiter detector is shown in Fig. 17. The quantities ai in Fig. 17 are related to the degree of limiting and can be solved for from an implicit equation. Although this detector re- quires knowledge of the value of 6 for implementation, a lower bound is shown in [30] for the detector power func- tion when it is designed for a specific set of values of the parameters.

Fig. 17. Correlator-limiter robust detector for a known sig- nal in <-contaminated Gaussian noise.

K A S S A M A N D P O O R R O B U S T T E C H N I Q U E S F O R S I G N A L P R O C E S S I N G 45 7

Page 26: Robust Techniques for Signal Processing - A Survey

In addition to the fact that there is only one independent class of pdfs, the noise pdfs, in the known-signal detection problem, a further limitation of the above approach to obtaining a robust detector is that the signal amplitude 8 needs to be known. In Huber's approach, the uncertainty classes of pdfs are usually defined as expansions of corre- sponding single nominal pdfs, so that the resulting robust test does not generally give a test which is uniformly robust in the strict minimax sense for, say, a range of nominal alternative hypotheses. As a specific example, consider the likelihood ratio of the robust test given by (4.10). Here c' and c" depend on the density f f . Thus in [30] where these results were applied to the known-signal detection prob- lem, the f t were defined for a particular signal strength 8 . While for this problem it is possible to find a lower bound for the robust detector power function when it is designed for a particular 8 , only for that 8 is the detector implement- ing a "minimax" test for the detection problem. Finally, the robust tests and least favorable pairs were necessarily ob- tainable only when Sr , , s f were not only disjoint but had a finite "separation"; thus the case of vanishing signal strength in weak-signal detection cannot be considered.

These considerations point to the desirability of another formulation for robust hypothesis testing problems of signal detection. One such alternative approach which has been quite fruitful considers the asymptotic case of weak signals and large sample sizes ( 8 + 0, n + 00). It is interesting to note that the basis for this alternative approach has been the theory of robust estimation of a location parameter which will be discussed briefly in Section V below.

Asymptotically Robust Detection: For the known-signal detection problem of (1.1) consider the log-likelihood ratio defined by (1.3) and (1.4). This generally depends on 8 , but for the locally optimum (LO) detector which maximizes the slope of the power (detection probability) function at 8 = 0, subject to a false-alarm probability bound, it is well known that the test statistic for given noise pdf f should be [127], [I 281

(4.13)

This LO test statistic is a special case of the generalized correlator (CC) statistic

n

TGC(X) = c a ; / ( x ; ) (4.14) i=l

where a , ; . . ,an is a correlation sequence and C i s a non- linearity. The class of GC statistics with ai = s; is then a natural class of candidate test statistics to which attention may be restricted in obtaining a robust detector for known si but for noise pdf f not precisely known.

For a class So of noise pdfs one would like to be able to obtain the minimax robust detector from amongst the class of GC detectors (with ai = s i ) , with slope of the detector power function as the performance criterion. For fixed finite sample size n this is, unfortunately, extremely difficult in general. However, in the asymptotic case n -+ 03 (and 8 + 0), under mild regularity conditions yielding asymptoti- cally normal distributions for the test statistics, the problem reduces to a consideration of a more tractable performance measure called the efficacy. The efficacy B of a detector

based on a test statistic T ( X ) may be defined as [I271 2

B= l im - 1 [ ; E W ) I 8 } I e-o ] . (4.15) n-rm n var{ T ( X ) l 8 = 0}

It also turns out that maximizing the efficacy with T ( X ) of the form

n

L ( x ; ; Si) i-1

obtained by dropping the undesirable 8-dependence in (1.3), leads one to the LO statistic of (4.13) because the slope of the power function can be related directly to the efficacy. Without the limit ( n -+ 03) in (4.15) the quantity is sometimes called the differential or incremental SNR.

For the test statistic of (4.14) with a, = si the efficacy becomes

within mild regularity conditions on f and C and with the assumption that C ( X j ) has zero mean value under the noise-only hypothesis. We may also assume without loss of generality that

that is, that the signal has unit average power when i ts amplitude is unity. A significant observation about the efficacy of (4.16) with this normalization is that the recipro- cal of this normalized efficacy is exactly the asymptotic variance of an M-estimate for the signal amplitude 8 of a constant signal (5, = 1). An M-estimate 8 of 8 is in general that quantity solving

n

C( x; - e ) = 0. (4.17) i=l

(Note that 4 is the sample mean when C ( x ) = x . ) The asymptotic variance of an M-estimate using nonlinearity t' i s the reciprocal of the normalized efficacy of (4.16) under some regularity conditions.

A brief discussion of robust M-estimation of a location parameter when the pdf f of the additive noise is not precisely known is given in Section V. Here we shall note only that in 1964 Huber [6] found the least favorable pdf in the class of <-contaminated noise pdfs for the robust M- estimation problem, and hence obtained the corresponding optimum M-estimate as the robust M-estimate. It should now be clear (since minimizing the asymptotic variance in M-estimation is equivalent to maximizing the efficacy for our detection problem) that the least favorable pdf found by Huber will also make its sequence of LO detectors (for n = 1,2, . . . ) an asymptotically robust sequence of detec- tors, providing a saddlepoint value for the game in which the efficacy is the performance function. In fact, one can get a stronger result where the false-alarm probability can simultaneously be bounded. These results for asymptoti- cally robust detection were first extended from Huber's estimation results, by Martin and Schwartz [30] and later were expanded upon by Kassam and Thomas [129].

458 PROCEEDINGS OF THE IEEE. VOL 73. NO. 3. MARCH 1985

Page 27: Robust Techniques for Signal Processing - A Survey

The general result in [30] and [I291 may be summarized as follows. Let 9 be the class of all detectors of asymptotic size (i.e., false-alarm probability) a for our hypothesis test- ing problem with f E So(g,c) as defined in (4.1). Now g is assumed to be a symmetric nominal density function that is strongly unimodal (Le., - logg is a convex function), and is twice differentiable, and the contamination h is symmet- ric and bounded but otherwise arbitrary. Let B,,(elf) be the power function of a detector D with f E So(g,c) the noise density function. Our false-alarm probability constraint i s that for each” D E 9

l im BD(Olf) d a, all f ESo(g,c). (4.18)

The asymptotically most robust detector D, E 9 is then defined as the locally optimum detector for the least favora- ble f, ESo(g,c) such that

n+ 00

l im 1 , all fEgFg(g,c), (4.19) rl- 0O Bb,(’l fR)

in addition to (4.18) and

The DR satisfying (4.18)-(4.20) exists for a > a(g,c), a lower bound depending on g and c , and is the locally optimum detector of asymptotic size a for fR E Fo(g,c) given by Huber’s exponential-tailed density

fR( = { (1 - c)g(x), 1x1 < a (1 - Og(a)exp[-b(lxl - a ) ] , 1x1 > a

(4.21)

where a, b satisfy the equation

1: g( x) dx + 2g( a) /b = (1 - c)-’ (4.22) a

and

The lower bound on a i s given explicitly as a function of g and c in [129].11 For the unit-variance Gaussian nominal density this bound is no less than 0.158, which is obtained when c 0. Note that the robust detector is based on the test statistic TGc of (4.14) with C= C, = - f ; / f , , which is given by

The threshold for the robust detector can be set by consid- ering the normalized test statistic

and basing the computation of false-alarm probability on

’‘More explicitly, each D E 9 is an infinite sequence ( D , , D2, . . . ) of detectors, one for each sample size n. The func- tional dependence of the test statistics on X is the same for all members of the sequence.

”A method for relaxing this restriction is discussed in [130].

the asymptotically normal distribution of this finite-variance statistic.

The resulting generalized correlator detector may be de- scribed as a limiter-correlator detector, and this is one of the canonical structures of robust detection theory. Note that when g is zero-mean, unit-variance Gaussian we have -g’( x)/g( x) = x, and !, becomes the “amplifier-limiter” or “soft-limiter” nonlinearity. The structure of the limiter-correlator detector which is asymptotically robust for known-signal detection in c-contaminated Gaussian noise i s shown in Fig. 18. Note that this particular GC

r---------I - I CorreIota !

.&,.soft Llmlter - I ol=s, I . , L -_-_-_-__ J

Fig. 18. Limiter-correlator robust detector for a weak signal in €-contaminated Gaussian noise.

detector can be used in place of the linear correlator detector acting on the matched filter outputs in Fig. 9, especially when the matched filter output noise compo- nents can be modeled as being €-contaminated Gaussian random variables with unknown contaminating pdfs.

further details of the above results are given in [30] and [129], including comparison with other detectors on the basis of asymptotic relative efficiency (ARE), and a discus- sion of the robustness property of the simple sign detector, which does an extreme form of limiting on the observa- tions. A numerical study of the performance characteristics of various limiter-correlator detector nonlinearities for the additive known-signal detection problem has been made by Miller and Thomas in [118]. The noise density in [I181 was taken to be the c-contaminated Gaussian nominal with contaminating densities of the impulsive-noise type mod- eled by exponential (Laplace) and generalized Cauchy func- tions. The nonlinearities considered were various multilevel approximations of the canonical ”amplifier-limiter,” in- cluding the hard-limiter and the noise-blanker. The perfor- mance was characterized in terms of asymptotic relative efficiencies with respect to the conventional linear-correla- tor detector.

The growing interest in detection systems using quan- tized data led to a consideration in [I301 of the above known-signal in additive-noise problem with the require- ment that the detector characteristic ! in (4.14) be an m-level piecewise-constant quantizer characteristic. It has been shown in [I301 that the most robust quantizer-detec- tor in the sense of (4.18)-(4.20) for the c-contaminated noise density class So(g,c) of (4.1), and for the detection problem (l . l) , i s again the locally optimum quantizer-de- tector for the least favorable density of (4.21).

Robust Detectors Based on M-Estimators: The detectors we have considered so far in this section have all had the general structure of Neyman-Pearson optimum detectors, which are based on a comparison of the likelihood ratio to a threshold. The test statistics of our detectors have been of the form

2 < , ( X i ) i = l

K A S S A M A N D P O O R : R O B U S T TECHNIQUES FOR SIGNAL PROCESSING 459

Page 28: Robust Techniques for Signal Processing - A Survey

which is that of the statistic of (1.3). A different approach for robust detection of additive known signals was pro- posed by El-Sawy and VandeLinde in [131]. In their ap- proach, the test statistic is simply an M-estimate of the signal strength parameter 8 , and their robust detector test statistic is the robust M-estimate of 8 for the same class of noise densities. To understand the motivation for this ap- proach, and its difference from the previous one and the possible advantages of the resulting detectors, let us reex- amine briefly the above results on asymptotically robust known-signal detection. It is clear that, as for any consistent test, the robust detector based on a test statistic of the form of (4.14) has a power function BD(81f) which approaches unity as n approaches 00 for each 8 > 0, assuming that

i 5f/n i-1

approaches a positive value and ai = s,. Thus the practical interpretation of (4.19) is that for large enough n for which use of the central limit theorem is reasonably well justified, the slope of the power function at 8 = 0 for D, may be considered to be minimized by f, in Fo(g,c).

In [I311 i t is pointed out that this condition does not guarantee that for each f €F0(g,c), we will have BD,(81f) > BD,(8l fR) in some interval (0, 8,) . A simple example of a density in 4ro(g,c) for which &,(elf) < BD,(elfR) when 8 =- 0 proves this. The lack of a strict inequality in (4.19), which allows this to happen, led to the alternative ap- proach in [131].

For the hypothesis testing problem of (l.l), note that for a finite-power signal (for which

n

Iim sf/. = c n-m . 1 - 1

a finite positive value) the signal energy as n + 00 becomes infinite under the alternative hypothesis H,. If the hypothe- sis testing problem statement is slightly modified by re- placing the amplitude 8 with v/n1l2 for some constant v, the total signal energy in any sample of size n remains finite, and the limiting energy is v2, assuming C = 1 without loss of generality. In a practical sense, suppose an observa- tion of length N is given for which the known signal sequence (s1,s2;..,s,,,) has amplitude 8, and for which

N

Sf/N C,. i-1

For N large enough, the results to follow can be applied with v = el[ NC,,,]l/’.

The minimax robust detector for this known-signal detec- tion problem was obtained from amongst a class of M-de- rectors in [131]. This is a class of detectors for which the test statistic is an M-estimate (as in (4.17)) of the signal strength 8. Let Y be the class of M-estimator functions L which satisfy convexity, symmetry, monotonicity, and differentia- bility requirements, in addition to mild requirements on the moments cd the associated random variables when the noise density function f is a member of a general class 9 of symmetric densities. The M-estimator b of 8 , based on n observations and a function L , is the value of 8 minimizing

n

L ( x; - Os;) i-1

so that n

b = arg min I ( X , - es,). (4.25)

This is a generalization for nonconstant signals of Huber’s definition. Note that (4.17) for s, = 1, all i, is obtained from this with / ( x ) = dL(x)/dx.

Let the asymptotic detection power as a function of v for an M-detector based on a function L be denoted as BL(vlf). This power function obviously depends on the noise den- sity function f €9. The analysis in [I311 establishes the following results. Let U2( f, L ) be the asymptotic variance of ri”*(b - e), which is asymptotically normally distributed with mean zero for f E F and L E 9. Suppose there exists a density f, EF such that L, = -log fK is in 9, and U2( f, L , ) Q U2( f,, L, ) for all f E . F . Let the M-detector based on L , have a threshold y. Then for v 2 y , we have

0 i-1

Bl,(olf) 6 BL,(OIfRR)! (4.26)

and

In fact, since the M-detector based on -log f, is asymptot- ically equivalent in performance to the detector based on the likelihood ratio for f, when fR is the noise density function, the maximum over Y in (4.27) can be replaced by the supremum over all detectors with size equal to that of the M-detector based on L,.

The main requirement in the above is that v not be less than y . This implies that for a given v, the design false-alarm probability cannot be too small. The requirement v > y can be met by making the sample size n sufficiently large in any given situation. Note, however, that the saddlepoint condi- tions (4.27) and (4.26) do not hold together for a given sample size and false-alarm probability constraint, in an interval around v = 0. It has been shown in [I311 that for such cases the robustness criterion based on asymptotic slope of the power function (as in (4.19) and (4.20)) also makes the M-detector based on L, the most robust detec- tor, but again subject to the same restrictions on the mini- mum value for the false-alarm probability.

If the saddlepoint solution (f , , L, ) exists it may be ob- tained by first minimizing the Fisher information function for location, I ( f ) , over all f E 9. The Fisher Information function for location is the function

I ( f ) = lrn ( f’( X))*/f( x ) dx. (4.28) - m

It is no accident that / ( f ) is the maximum value of the normalized efficacy in (4.16), obtained for / ( x ) = - f’(x)/f(x), the LO nonlinearity. The minimizing density is f, and the function L , = -log f, in Y gives the saddle- point pair ( f , , L,) . In [I311 the family of symmetric density functions with a known probability p assigned to an inter- val ( - a , a ) was explicitly considered as an example. For this ppo in t class of densities the least favorable density f, was again shown to have exponential behavior outside ( - a , a ) , and the robust M-detector characteristic L, had derivative t‘, which was a constant outside ( - a , a ) . Some numerical performance comparisons of asymptotic performance and finite-sample simulation results are also given in [I311 for a particular example.

The significance of the function dL(x)/dx = / ( x ) in M- detection is that under appropriate regularity conditions,

460 P R O C E E D I N G S O F THE IEEE, VOL. 73, NO. 3. M A R C H 1985

Page 29: Robust Techniques for Signal Processing - A Survey

the test statistic 8 must satisfy n

S i t ( X , - 8 S i ) = 0. i-1

(4.29)

(This is the same as (4.17) for s, = 1 , all i.) The variance of the statistic 8 i s the reciprocal of the efficacy of the generalized correlator detector based on the test statistic of (4.14), with ai = s,. Note that the mean of 8 i s 0. Since t‘ i s a monotonic increasing function of i ts argument for L in the class 9, we have a simple way of implementing such a detector when s, all have the same sign. In this case, the sign of

n c S A X , - us,) i-1

directly indicates if 8 is above or below the threshold y . The robust detection of known signals using an M-esti-

mate has also been extended in [I321 to the sequential binary signaling problem. Here the two hypotheses H, and HI are defined by

where the density function of noise components N, is a member f of some class 9 of symmetric densities. Under the same mild restrictions on the class Y of allowable M-detector characteristics L as in the nonsequential case, a robustness property is established in [I321 for the sequential M-detector (MS-detector) based on the sequence of robust M-estimates 8 defined by the same characteristic L, as in the robust nonsequential case. Thus when f = f, for this robust scheme, the probabilities of error are upper bounds on the probabilities of error for arbitrary f E 9, and the same holds for “normalized” expected sample sizes, in the limiting case when lel - $1 4 0 and sample sizes are large, for which Gaussian approximations to distributions for 8 can be used. In addition, it has been shown that the pair ( f,, L,) also forms a saddlepoint for performance measured as a risk function which is a linear combination of the error probabilities and the “normalized” expected sample sizes, for one set of weighting or cost coefficients. The “normal- ization” of the expected sample sizes is required to obtain well-defined quantities, since the results are valid asymptot- ically when actual sample sizes become infinitely large.

An alternative scheme for robust sequential testing in the situation described by (4.30) was also considered in [132]. This scheme is the robust stochastic approximation sequen- tial (SAS) receiver. Again the test statistic is directly based on the sequence of robust fixed-sample stochastic ap- proximation estimates of the location parameter 0 = e,, or el; this estimation scheme will be discussed briefly in Sec- tion V. It has been shown that the robust MS- and SAS-de- tectors have identical asymptotic characteristics. The SAS- detector does not store past observations, as it uses a recursive algorithm for updating the estimate of 0 as new observations arrive. On the other hand, the MS-detector converges to i ts asymptotic performance faster and does not need an initial estimate to obtain 8.

Another class of robust estimates of mathematical statis- tics i s the class of L-estimates, which are estimates formed as linear combinations of order statistics. One example is the median, which also happens to be an M-estimate. The a-trimmed mean i s another example of an L-estimate and

will be discussed in Section V. The robustness of these L-estimates in estimation has prompted investigation of their use in signal detection; notably Trunk and George [117], [133], [I341 have investigated the use of some simple L-estimates in signal detection problems.

The results we have discussed in this subsection show quite clearly the central role in robust signal detection of robust estimation theory, particularly of the original results of Huber, in addition to that of Huber’s results on robust hypothesis testing. A different formulation of the additive- signal robust detection problem in bounded-amplitude noise was considered by Morris [135]-[137]; this work is related to the early work of Root [9] mentioned in the Introduction. We now go on to consider, in somewhat less detail, the other two canonical signal detection problems of random signals and of bandpass signals in additive noise.

C. Robust Detection of Random Signals

In many applications involving the detection of random signals the observations are obtained simultaneously from a number of sensors forming an array. The detection problem then often becomes that of detecting a random signal common to each sensor in the presence of noise processes uncorrelated with each other and with the signal. An exam- ple of such an application is a hydrophone array in a passive sonar system used to detect the presence of, and to locate, sources of random signals. For each search direction, rela- tive time delays are imparted to the sensor outputs to equalize the propagation delays to each sensor for a poten- tial signal from that direction. This is followed by a detector for a common random signal.

A special case of interest is that in which the sensor array is composed of two elements. In this situation, the sampled observation vectors X, = (Xl1, X l Z ; . . , ( y l n ) and X , = (X, , , X,,; . a , can be described by

X,i=f?S,+N/i , i = 1 , 2 , - . . , n , j = 1 , 2 (4.31)

where the signal amplitude 0 is zero under the noise-only null hypothesis and 0 # 0 under the alternative hypothesis. We assume here that the Ni = (Ng, Nj2; .., Njn) are inde- pendent sequences of zero-mean iid noise components with common pdf f , and S = (S,,$;..,S,) i s an indepen- dent signal sequence of iid zero-mean components, whose variance may be taken to be unity without loss of gener- ality.

The presence of a signal in the above model (e # 0) causes the output pairs of observations ( X l i , X , ; ) to be positively correlated, in addition to increasing the power level received at each sensor. The increase in the corre- lation value from zero is in particular due to the presence of the common random signal. Thus it is quite reasonable to restrict attention to the class of generalized cross-correla tion (CCC) detectors based on test statistics of the type

n

TGCC(XlfX2) = c t ( , % ) t ( X , ; ) (4.32) i-I

where t‘ i s the detector weighting function. For detection of weak signals ( e 2 4 0) it can be shown easily that detec- tion efficacy is maximized with ! ( x ) = - f ’ ( x ) / f ( x ) , which is also the LO nonlinearity for known-signal detection.

In one of the earliest published studies of robust detec- tion of random signals, Wolff and Gastwirth [29] examined several specific simple nonlinearities t for robustness of

KASSAM AND POOR. ROBUST TECHNIQUES F O R SIGNAL PROCESSlNC 461

Page 30: Robust Techniques for Signal Processing - A Survey

performance when f is not known. A finite class of pdfs comprised of the Gaussian, logistic, and the pdf for a particular t-distribution was considered in [29]. The nonlin- earities G considered in [29] were the simple three- and four-level symmetric quantizer and the “amplifier-limiter” or “soft-limiter” continuous function which is linear in an interval including the origin but constant outside that inter- val. The criterion of performance used in [29] was effec- tively an asymptotic relative efficienc): it was the ratio of the efficacy for noise density f obtained with weighting function C and the optimum efficacy for noise density f. In the definition of efficacy for this problem, 0’ i s now the signal strength parameter.

In [I381 the detection of a random signal common to an array of receivers was considered as an extension of the two-input case. Again, no attempt was made to include a bound on the false-alarm probability as part of the perfor- mance criterion, so that the assumption was that detector thresholds could be adapted to get the correct false-alarm probability for any noise density. The performance criterion was the detection efficacy, which as we have remarked before for the known-signal case is directly related to the slope of the power function at 0 = 0. For the correlator-array structure (generalization of (4.32) by a sec- ond sum over all pairs of sensors) an interesting saddle- point robustness result was shown. This was that for the €-contaminated class Fo(g,c) describing the independent noise component at each receiver across the array, with the same restrictions on it as in the known signal in additive noise case, the minimax robust cross-correlator nonlinearity t‘, is exactly the same as C, of (4.24) for the known signal case. Of course, the result here was only possible in terms of the weaker efficacy criterion above. Fig. 19 shows the

i” ri

‘fy

u Threshdd

+ Decism

e = o or e*o Fig. 19. Generalized cross-correlation robust detector for a weak random signal.

structure of a two-element random signal detector using an “amplifier-limiter” CCC nonlinearity d .

In 11381 other variations of the c-contamination class for this multi-input problem were also considered. Modifica-

tions of locally optimum detectors (rather than cross-corre- lators) for common random-signal robust detection were also investigated, although explicit minimax robustness re- sults could be deduced only for the special case of a contaminated double-exponential noise density, for which a hard-limiter based polarity coincidence array detector [I391 is most robust. The polarity coincidence array detector i s based on use of the two-level sign function for C, and is well known as a nonparametric detector with a fixed false- alarm probability for zero-median noise pdfs.

The class of detectors for single-input additive random signal contains no counterpart of the cross-correlator struc- ture of the multi-input problem. Development of robust detectors for this problem has paralleled that for the single-input known-signal problem. The first instance of this is due to Martin and McGath [I401 in which the optimum quadratic detector for Gaussian statistics was modified, from intuitive considerations, into the limiter- quadratic detector. Numerical asymptotic relative efficiency computations for Gaussian-mixture noise pdfs (€-con- taminated nominal Gaussian with larger variance Gaussian contaminating density) and finite-sample detection power and false-alarm probability computations verify the ex- pected robustness of the limiter structure. The subsequent results in [I381 extended such structures to the multi-input or array case.

Explicit minimax results for the single-input additive ran- dom-signal detection problem paralleling those for known-signal detection have been obtained more recently in [141]. Two statistical models for this problem have been investigated in [141]. In one model, the alternative hypothe- sis of signal presence is defined to produce additive signal and noise components in the observations, (i.e. (4.31) for j = I) . In the second, scale-change, model the alternative hypothesis is defined by the condition that the pdf of the iid observation components X ; i s f ( x / o ) / o , where f is the null-hypothesis observation (noise) pdf. It had been shown earlier [I421 that for the additive noise model the locally optimum detector is based on the generalized energy (GE) detector test statistic

(4.33)

The LO statistic for the scale-change model is also given in [I 421.

From knowledge of the previous results on known-signal robust detection one might conjecture that a minimax re- sult should be obtainable for the robust detector formed by introducing hard-limiting beyond some argument value a into the nonlinearity f ” / f of (4.33). Indeed, consider the exponential-tails least favorable density of (4.21), with a, b defined as in (4.22) and a modified (4.23), namely

(4.34)

For this noise density the nonlinearity f ” ( x ) / f ( x ) becomes

Consider the c-contamination model S0(g,c) for the noise density f , as described in our summary of known-signal results. In addition to requiring a further smoothness prop- erty for g (since here $‘ is involved), a more restrictive

462 P R O C E E D I N G S OF T H E IEEE, VOL. 73, N O . 3, M A R C H 1985

Page 31: Robust Techniques for Signal Processing - A Survey

further condition on the allowable contamination h is im- posed. This is that allowable contamination densities are zero in ( - a , a). With these conditions, results exactly paral- leling those for asymptotically robust detection of known signals in additive noise are obtained in [141]. An interest- ing feature of the solution, arising from the restriction on allowable h, is that (4.19) holds with equality over the allowable f and the result is valid for all values of the size a. Numerical asymptotic relative efficiency comparisons for general limiter nonlinearities are also given in [141].

A very similar result is established in [I411 for the scale- change model. Here again for a similarly restricted version of the c-contamination class, the density of (4.21) i s again shown to be least favorable. A noteworthy point about this solution is that as in the case of asymptotically robust known signal detection, this scale-change robust solution is directly related to Huber's results on robust M-estimation of a scale parameter [6]. Thus we see once again a close relationship between robust estimation and asymptotically robust detection problems.

D. Robust Detection of Bandpass Signals in Bandpass Noise

Bandpass signals are commonly encountered in applica- tions such as radar and communication systems, and tech- niques for their detection in bandpass noise with impre- cisely known statistical characterizations are therefore of practical interest.

Let us first consider briefly one bandpass known-signal detection problem for which an asymptotically robust de- tector may be defined using ideas very similar to those used for low-pass known-signal and completely random-signal robust detection problems. An observed continuous-time waveform is now described by

X ( t ) = 6 v( t ) cos [ wot + +( t ) ] + N( t ) (4.36)

where the low-pass signal amplitude and phase compo- nents v( t ) and +( t ) , respectively, and the frequency q, are known, the overall signal amplitude 0 being either 0 (noise only) or having some positive value (alternative hypothesis). The bandpass noise N( t ) may also be expressed in terms of its in-phase and quadrature components N,( t ) and N,(t). For a detector operating on sampled values of the in-phase and quadrature components of the observation X ( t ) the input data may then be represented as a vector ,i = X , + j x , where the components X,, and X,,, i = 1 ,2 , . . ., n, of X , and X Q , respectively, are

X, ; = eS,; + N,; (4.37a)

xQI = es,, + NQ;. (4.37b)

Here the s,,, sol, N,,, and NQi are samples of s,( t ) = v( t ) cos +( t ) , so( t ) = - v( t ) sin +( t ) , N,( t ) , and No( t ) , re- spectively. We make the assumption that the i; = X, , + i%,, are independent for i = 1,2; .., n, implying a restriction on the sampling rate.

If the noise components N,, and NQ, are restricted to be independent then this problem is essentially the same as the low-pass known-signal detection problem. A more gen- eral assumption is that the joint pdf f,, of N,; and N,, has circular symmetry, so that

f/Q( u t .) = c( I.-=. (4.38)

For Gaussian bandpass noise both circular symmetry and independence of the in-phase and quadrature noise com-

ponents is obtained. Under the circular symmetry assump- tion the LO detector uses a generalized narrowband corre- lator (CNC) test statistic

n

TCNC( k) = t'( R ; ) Re { S, ,kf} (4.39)

where the R ; = I i i l are the observation envelopes with t' the LO envelope weighting function [I431

i=l

and where S i = s,, + is,;. Note that f ( r ) = 2nrc(r) is the pdf of the R; . For Gaus-

sian noise f i s the Rayleigh pdf and the LO function of (4.40) i s a constant resulting in the linear narrow-band correlator (LNC) detector. In general, if the pdf f i s some known function 2arg(r) the LO detector can be obtained. In [I441 the c-contamination model for f was considered, with conditions on the nominal function g similar to those for known-signal asymptotically robust detection. As in the case of random-signal detection, however, the class of contaminating pdfs 2arh(r) was restricted to produce tail- contamination only, that is, the h were zero in some inter- val at the origin. With these restrictions, the asymptotically robust detector was shown in [I441 to be a CNC detector which is LO for a least favorable pdf f, having an exponen- tially decaying tail. The weighting function t'= t', in the test statistic TCNC(i) of this robust detector is

for a depending on c and g. Notice that the function rt ' ,(r) i s a constant for r 2 a; this function may be interpreted as providing the weighting applied to the hard-limiter narrow-band correlator (HNC) terms Re { S j ( i ? / R j ) } , the , k j / R j having unit amplitude value. For a nominal Gaussian case, the function -g ' ( r ) /g( r ) is linear in r .

Fig. 20 shows the structure of the asymptotically robust detector for known bandpass signal under the above as-

Llmftw

Fig. 20. Generalized narrow-band correlator robust detec- tor for a weak bandpass signal.

sumptions. Note that the HNC detector is a special case of this structure and may be viewed as the robust detector for a nominal g function which is exponential, or as an extreme case which actually possesses a constant false-alarm prob- ability and functions as a nonparametric detector [145]. Further interesting properties of the robust detector em- ploying t', of (4.41) are discussed in [144].

So far we have been considering the case of a coherent

KASSAM A N D P O O R . ROBUST TECHNIQUES F O R SIGNAL PROCESSING 463

Page 32: Robust Techniques for Signal Processing - A Survey

bandpass signal. In the incoherent case there is an ad- ditional phase term 'k, uniformly distributed on [0,2.n), so that the signal is v( t ) cos[w,t + +( t ) + *I. The asymptoti- cally optimum detector test statistic now becomes

l?GNc(i)12 = I C(Ri)Zikt (4.42)

where d is the optimum nonlinearity of (4.40). It turns out that the efficacy for such a square-law quadrature GNC detector is directly related to that of the GNC detector for coherent signals, and thus d = dR of (4.41) also results in an asymptotically robust detector in the incoherent case.

In [I461 El-Sawy considered a special case of the incoher- ent signal detection problem in which v ( t ) , the low-pass signal amplitude, was a constant, and + ( t ) , the low-pass signal phase term, was - ~ / 2 (or, equivalently, zero). The in-phase and quadrature noise samples were assumed to be independent and a robust detector was obtained which used the squares of robust M-detector outputs for the separate in-phase and quadrature observations. Specific re- sults were obtained for the ppoint classes of univariate noise pdfs. A more general problem involving signals with random parameters has been considered recently by Kelly in [147].

Another important variant of the observation model of (4.36) is obtained by allowing a random signal amplitude 0 = A in addition to the random phase 9. The usual as- sumption for the signal is that A and * are independent, with A being a Rayleigh random variable and \k being uniformly distributed on [0,2v). In [I481 the performance characteristics of a particular modification of the optimum detector for Gaussian noise are considered. The modifica- tion consists of replacing the squarers of the optimum detector with limiter-squarers for robust performance. Both the single bandpass-pulse detection situation and that of multiple independent bandpass pulses is considered. In the latter case, a binary integration or double-threshold detec- tor is considered. The structure of this detector is shown in Fig. 21. The numerically computed performance characteris-

12 i - 1

v ( t ~ u n ( w ~ t + + ( t ) )

Fig. 2l. Robust detector for independent random ampli- tude and phase bandpass pulses in additive noise.

tics indicate that, for the independent €-contaminated in- phase and quadrature bandpass matched filter output noise components that were assumed, the limiter-squarer struc- ture is very effective in guarding against drastic perfor- mance deterioration in heavy-tailed non-Gaussian con- tamination. The structure considered in [I481 was suggested by the many formal results which we have discussed that show the robustness of limiter-type structures in different situations. In fact, the problem considered in [I481 may be considered as being a bandpass version of a random signal detection problem, for which robust detectors have been

considered in [I401 and [141]. We should note finally that Martin and Schwartz also considered in their study of robust detection [30] the detection of multiple independent pulses. The robust structure they suggested used a limiter function on the sampled in-phase and quadrature observations in- side a digital or discrete-time implementation of the quadrature single-pulse matched filters, followed by a square-law envelope detector.

We have described in these last three subsections the main results on asymptotically minimax robust detection for three canonical signal detection situations. A general theory for robust detectors in the asymptotic case (vanishing signal strengths, infinitely large sample sizes) is discussed by Wolff and Sullivan in [149]. Here asymptotic normality of statistics of the form of (4.14) i s studied for four detection problems; namely, known signal in additive noise, two random signal problems (one with a scale-change in the noise density, another with an additive random signal, under the alterna- tive hypothesis), and envelope detection of a narrow-band signal. In addition, the general characteristics of the robust solutions are considered, which are not restricted only to the univariate case, and a few explicit uncertainty models and robust detector structures are discussed. This work essentially considers only a performance measure related to the efficacy in the known signal case, and defines a general notion of Fisher's information; simultaneous consideration of a false-alarm probability constraint is not attempted in [I 491.

E. Extensions and Other Results

While we have surveyed the basic results of robust hy- pothesis testing and robust signal detection in the above subsections, several further results in this area deserve some comment. We have seen that explicit minimax robust struc- tures can be derived under some rather specific and some- times restrictive assumptions. In this subsection we will also mention some work that has been directed at easing some of these constraining assumptions.

Signal Uncertainty: All the results that we have dis- cussed in this section were obtained for robust detection when only the noise pdfs are not precisely known. Con- sider, for example, the known-signal detection problem of Subsection IV-6. In our discussion of the asymptotically robust detector of the generalized correlator type with test statistic of the form (4.14), the choice a, = s i for the coeffi- cients in the test statistic was the obvious one for known- signal detection. But it is also possible to consider here uncertainty about the exact values of the signal compo- nents s,. Notice that this problem was considered in Sec- tion Ill for robust matched filters and SNR performance; however, here the CC detector is not necessarily linear and we have been concerned with weak-signal detection per- formance.

In [I501 Kuznetsov considered this signal uncertainty problem, using the differential SNR (see discussion follow- ing (4.15)) as a performance measure. For known noise pdf and e, signal uncertainty class the result is quite similar to that for the robust linear matched filter of Section Ill, because the differential SNR is after all a performance measure derived from the SNR. The random-signal detec- tion problem with signal covariance matrix uncertainty is also considered in [ISO]. For joint uncertainty in signal

4.64 PROCEEDINGS O F THE IEEE. VOL. 73, NO. 3, MARCH 1935

Page 33: Robust Techniques for Signal Processing - A Survey

characteristics and noise pdfs solutions for the minimax robust structures are difficult to obtain in the general case. The special case where the noise is Gaussian with uncertain covariance matrices for the signal and noise has been treated in [150], with the detector restricted to employ a combina- tion of linear and quadratic statistics. A similar problem involving deterministic signals and linear detectors has been considered by Kuznetsov in [99], as mentioned in Section Ill.

Serial Dependence and Asymmetry: Although a consid- eration of the asymptotic performance of detectors allows appealing minimax robust detector structures to be derived, a major limitation of the schemes we have described is that the asymptotic robustness property can be attributed to them only under the assumption of independence of the detector input samples. This assumption is generally dif- ficult to get around, serial data dependence introducing considerable complications in the analysis. One of the problems, of course, is the definition of appropriate statisti- cal models and uncertainty classes. Recently, however, some progress has been made in the specification of asymptoti- cally minimax robust detectors for serially dependent data samples.

In [I511 Poor considers the case of weak serial depen- dence of data in a constant-signal M-detection problem. The dependence structure considered is that obtained from a moving-average model for the noise components. In particular, the noise model

(4.43)

is considered where the { Y } are iid with uncertain margin- al pdfs. For p = 0 the approaches described previously are applicable; in [I511 the case of weak nonzero correlation is considered. For the €-contaminated pdf classes, it is shown there that the robust M-detector nonlinearity is the limiter function which is robust for p = 0, corrected by a linear term. Thus for the nominal Gaussian case a nonlinearity t', of the form illustrated in Fig. 22 i s suggested; more detailed

Fig. 22. M-detector nonlinearity for robust detection of a signal in <-contaminated, weakly correlated, Gaussian noise.

considerations modify this so that the nonlinearity remains bounded. These results are similar to those derived earlier by Portnoy for robust estimation with dependent data [152], [I 531.

More recently Moustakides and Thomas [I541 have con- sidered a less structured dependence assumption for the known-signal detection problem. The additive noise se- quence was assumed here to be +-mixing, which includes the case of stationary Markov noise sequences. For the c-contamination model for the univariate noise pdf model, with attention confined to the class of generalized correla-

tor detectors, and with efficacy as the performance criterion and an explicit requirement of false-alarm probability con- trol, the form of the robust detector nonlinearity was de- rived. The very interesting conclusion was that, subject to some regularity conditions, the robust detector nonlinearity is a null-zone modification of the independent-data robust detector nonlinearity for the same class of univariate noise pdfs. For the nominal Gaussian noise pdf the form of this nonlinearity i s shown in Fig. 23. It is interesting to note that

iR"'

Fig. 23. Generalized correlator detector nonlinearity for robust detection of a signal in <-contaminated correlated Gaussian noise.

a simple three-level nonlinearity which approximates this characteristic can be used to provide nonparametric perfor- mance for symmetric noise pdfs [155]. The result of Moustakides and Thomas represents a major breakthrough in the extension of asymptotically robust correlator-detec- tor structures for dependent data, and similar results can now be sought for random- and bandpass-signal detection problems.

Another aspect of robust detection in dependent data has been considered recently by Martin [156], in which autoregressive dependence and a regression signal model which includes the known signal and random-phase band- pass signal cases are assumed. He suggests that, in this situation, robust M-detection is best accomplished by first prewhitening the noise and then applying an M-estimate to obtain the test statistic.

The asymptotic theory of robust detection, fundamentally related to Huber's results on robust parameter estimation as it is, is limited by the same factors that have constrained the applicability of robust parameter estimation theory. In ad- dition to serial independence of data one other assumption which has been generally required, at least in the widely used c-contamination model for uncertain pdfs, is symme try of the pdfs of the noise. In estimation theory this results in unbiased estimates for which the variances can be writ- ten down as second moments. Recently, an attempt has been made in [I571 to apply to robust detection of known signals ideas formulated for robust M-estimation for classes of density functions allowing tail asymmetry. Specifically, [I571 considers both the generalized correlator structure of (4.14) and the M-estimate structure (4.29) for detection test statistics, and develops the form of the robust detection function t', in (4.14) and (4.29) for classes of c-con- taminated nominal densities with arbitrary behaviors out- side a central interval. The development is based on the work by Collins on the corresponding M-estimation prob- lem [158], and shows the robustness of detector functions which redescend to zero and remain zero outside the interval of symmetry.

Two general limitations of minimax robust detection the-

KASSAM A N D P O O R . R O B U S T T E C H N I Q U E S F O R SIGNAL PROCESSING 465

Page 34: Robust Techniques for Signal Processing - A Survey

ory should by now be clear. These are that a considerable part of the theory that has been developed is asymptotic theory, and that it is currently unable to deal directly with continuous-time observation processes. It should be kept in mind that here our concern is with uncertainty in the noise pdfs, and not just with power spectral density uncertainties and linear schemes which we discussed in Section ill. Care has to be taken in applying asymptotic theory in practical situations involving a finite number of observations, since in some cases predicted asymptotic performance is ap- proached rather slowly. The Gaussian approximation for test statistic distributions in setting threshold values should be applied with care, even if the structure of the robust test statistic appears to have strong justification. in addition, another aspect of asymptotic formulations is that they are based on the assumption that signal strengths are approach- ing zero. Thus the effect of a nonzero signal and a finite sample size together could result in deviations in perfor- mance from theoretical predictions, even for large samples.’* Since closed-form analytical results are rarely feasible, about the only alternative left is numerical computations and simulations to verify the actual performance of such robust schemes in general.

Continuous-time results are very difficult to come by because of convergence issues and also because of the difficulty in specifying models to define simple classes of allowable random processes which are physically meaning- ful. Of course, if only second-moment characteristics are modeled and exact probability distribution functions are irrelevant, as in maximizing SNR at the output of a linear detector, robust filters such as the robust matched filter can be obtained. Although [I601 introduces classes of “mixture” or contaminated random processes, there has been no application made of this in signal detection with continu- ous-time observations. However, Kelly and Root [161], [I621 have recently combined some of Root’s original stability ideas with a minimax-type design philosophy to develop robust schemes for continuous-time processes.

Adaptive and Nonparametric Detectors: in a general sense, a detector can be said to be robust if it has good (close-to-optimal) detection performance under nominal conditions and if it also maintains an acceptable level of performance when the noise statistics deviate, within some allowable class, from the nominal. Our focus in this paper has been on fixed detectors designed to provide minimax detection schemes, which generally possess the above two characteristics. Another approach which can be taken when the noise environment statistical characterization is impre- cisely known, or is largely unknown, is to use adaptive procedures. in most adaptive procedures one generaljy be- gins with a specific test statistic structure in which some parameters are free to be set and updated as functions of previous inputs, which might include separate training data. in addition, the threshold is also usually free to be set adaptively. We have mentioned earlier that even for “fixed” minimax robust detectors the thresholds may have to be set adaptively to maintain false-alarm probability requirements whenever the performance criterion does not explicitly include such requirements. Since our focus here has been on minimax robust fixed-test-statistic structures we shall

’’For a recent contribution on nonlocal asymptotic robustness see Moustakides [159].

not try to give here an exhaustive survey of adaptive robust detectors, but will mention only some of the main recent contributions.

One general structure in the known-signal detection problem is obtained by requiring the function f‘ in the generalized correlator test statistic of (4.14) to be an m-level quantizer characteristic. Such a quantizer-correlator detec- tor may be viewed as partitioning the observations X ; into m subsets or intervals to each of which a distinct level is assigned. For a given noise pdf the asymptotically optimum quantizer characteristic can be found for this and similar detection problems [163]. if, however, the noise pdf is not known one may use estimated values of these optimum parameters. Alternatively, the quantizer breakpoints may be required to fall at the quantiles of the noise distribution, which can be estimated and updated. Adaptive m-interval partition detection schemes based on this idea have been described by Kurz in [I641 and for sequential detection by Dwyer in [165]. Other studies of this nature can be found in

Recognizing that contamination of a Gaussian noise den- sity function by an impulsive-noise component, and heavy-tailed densities in general, are reasonable models for random noise and interference in several applications [118], [I711 adaptive structures are considered by Modestino and Ningo in [143], [I721 for good performance over such classes of noise. The simple structure considered in [I721 for the detection of a known signal in additive noise is illustrated in Fig. 24. i t is motivated by the fact that for Gaussian noise

[I 661-[I 701.

H a d Llrnlter

Fig. 24. Adaptive detector for signals in a mixture of Gaussian and impulsive noise.

the linear-correlator test statistic is optimum, whereas for heavy-tailed noise pdfs often used to model impulsive noise the sign-correlator test statistic performs very well. The mixture test statistic considered in [I721 is T&(X) of (4.14) with a, = s, and

(4.44) where the free parameter y is set adaptively. Modestino in [I 721 discusses a stochastic approximation technique maxi- mizing the SNR for setting the value of y . A similar robust detector structure based on adaptively forming an optimum test statistic combination is described in [I431 for bandpass signal detection. An alternative detector described by Mil- stein, Schilling, and Wolf in [I731 uses extreme-value theory to obtain the proper threshold setting in an otherwise fixed matched-filter envelope detector structure.

Although adaptive schemes can be useful in situations where the noise statistics are unknown or are nonsta- tionary, the implementation of efficient adaptive schemes

466 P R O C E E D I N G S OF THE IEEE. VOL. 73, NO, 3, M A R C H 1985

Page 35: Robust Techniques for Signal Processing - A Survey

can add in a major way to the complexity of a detector. Furthermore, other considerations such as speed of con- vergence may limit their applicability.

In nonparametric detection the main concern is that the probability of false alarm remain bounded by some design maximum value for broad classes of noise statistics, for example for all univariate noise pdfs which are symmetric about the origin. Once again it is possible to consider adaptive structures for nonparametric detection, which are mainly of the adaptive threshold type, although fixed-struc- ture and fixed-threshold nonparametric schemes are the most common. For a given class of noise pdfs there gener- ally exist several possible nonparametric detectors, and in choosing between alternatives one usually considers detec- tion performance at some nominal cases within the class. The most common nonparametric detection schemes are those based on signs and ranks of observations, including multilevel (quantization) versions of sign-based schemes. It often turns out that nonparametric detectors are robust, in the more general sense, in their detection performance over the classes of noise pdfs for which they are designed. Note, however, that the primary aim of nonparametric schemes is to keep the false-alarm probability bounded by any desired value, whereas robust schemes are required to additionally exhibit good detection performance for the whole class instead of at some nominal operating points only. In this regard the reader i s referred to some comments by Huber in [20]. For survey papers, a bibliography, and details of non- parametric detectibn schemes we refer the reader to [I741 and [175].

In conclusion, we point out a recent Russian survey paper by Krasnenker, available in English translation [33] which is on the subject of robust detection. Another similar survey by Ershov [25] is concerned with robust estimation, al- though it also covers some detection and hypothesis testing results. VandeLinde [32] gives a short survey of robust tech- niques in communications which covers robust detection. Finally, Poor has given a more recent survey, emphasizing some of the mathematical details, in [34].

v. NONLINEAR METHODS F O R ROBUST ESTIMATION

In Sections II and Ill we discussed robust linear methods for estimation and detection within uncertain second-order models. In Section IV we saw that, when the uncertainty is described in terms of the distributional model rather than the second-order model, nonlinear methods are called for to provide robustness in signal detection. Similar considera- tions arise in problems of estimation in uncertain distribu- tional models, and in this section we discuss some of the main issues arising therein. Since many of these issues have been surveyed and unified elsewhere (see, e.g., the book by Huber [28], and the surveys by Martin [37], Ershov [25], and Poljak and Tsypkin [38]), we touch only briefly on the essential ideas of this area.

The large majority of work in this area has been con- cerned with robust nonlinear parameter estimation rather than with robust nonlinear filtering. In Subsection V-A we outline basic methods for (minimax) robust parameter estimation by considering the important special case of robust estimation of signal amplitude. Some of these meth- ods have already been mentioned briefly in Section IV, since they are closely related to robust detection. In Subsec-

tion V-B we consider robustness in other estimation con- texts including system identification and Kalman filtering. We also give brief mention to problems of robustness in other methods of time-series analysis.

A. Robust Estimation of Signal Amplitude

Many signal processing applications arising in practice fall within the category of parameter estimation. A common example is the estimation of the amplitude of a signal embedded in additive noise. A standard model for this particular situation is that we have observations X = ( X , , . . . , X,) given by

X ; = N,+es, , ; = I , . . . , n (5 .I)

where ( N , ; . . , N,) is an iid noise sequence with symmetric marginal pdf f , (s,;.., 5,) is a known signal waveform, and the signal amplitude 8 is unknown and is to be estimated.

The particular situation of (5.1) in which the signal is constant (s, = s2 = . .. = s, = 1) is the location estima- tion problem of statistical inference, and it was this prob- lem that was studied in the seminal paper of Tukey [5] that demonstrated the lack of robustness of the classical estima- tors for this parameter, and in the pioneering 1964 paper of Huber [6] that established the importance of minimax meth- odology for robust estimation. In the following discussion of the basic results in this area, we will consider this constant signal case unless otherwise noted. Modifications for time-varying signals will be mentioned where ap- propriate.

Three classical estimators for the location parameter of (5.1) with constant s, are the sample mean

- 1 ” X = - - C X ,

i-I

the sample median, med { X , ; . . , X , } , and the maximum- likelihood estimate (MLE)

The sample mean and the M L E coincide for the case in which f is a Gaussian density

f( x ) = - 1 e - x 2 / 2 0 2

G U

and the sample median and the M L E coincide when f is a Laplacian (double-exponential) density

All three of these estimators are examples of a more general class of estimators, known as M-estimators, proposed by Huber in [6]. As discussed in Section IV, this class of estimators consists of those estimators of the form

I n

where L i s a function determining the estimator. The sam- ple mean corresponds to the choice L ( x ) = x 2 , the sample median corrsponds to L ( x ) = ( X I , and the M L E corresponds to L ( x ) = -log f ( x ) . Note that 4, of (5.3) is the estimate of 0 that best “fits” the data when errors, X ; - 8 , ( X ) , are weighted with the function L .

K A S S A M A N D POOR: ROBUST TECHNIQUES FOR SIGNAL PROCESSING 467

Page 36: Robust Techniques for Signal Processing - A Survey

Assuming that L is convex, symmetric about the origin, and sufficiently regular, the M-estimate based on I is consistent and has the property that 6 ( 8 , ( X ) - 6 ) is asymptotically Gaussian with zero mean and variance V(d, f ) where

/e2 f V ( d , f ) A 7 \ 7 (5.4)

with d = I’ (see [6] for details). Thus estimators of this type can be compared on the basis of their asymptotic variances V(d, f ) . For now we will call the function d the influence curve of 8, although this term will be given a slightly different meaning below. For fixed f = f,,, the Schwarz inequality implies that

V ( d , f , , ) 2 V(d0,fO) =- I( f,,) (5.5)

I

where do = - c/f,,, and where I( f ) = /( f ’ ) 2 / f i s Fisher’s information for location. (Note that the inequality V ( t , f,) 2 I/ /( h) is a special case of the Cramer-Rao lower bound.) This do corresponds to Lo(x) = -log &(x); so, assuming - log 6 is convex, the most efficient M-estimate of location is the maximum-likelihood estimate, a well-known result.

If, as in the signal-detection problems discussed in Sec- tion IV, f is not known precisely but rather is known only to be in some class 9 of bounded symmetric pdfs, then it is possible that the performance of an improperly designed location estimator can be quite poor. For example, suppose we consider the r-contaminated Gaussian class

.F= { f l f = (1 - c ) + + ~ h , h E X } (5.6)

where

X is the class of all bounded symmetric pdfs, and E E [0, I]. The optimum estimator based on the nominal model + is the sample mean, which corresponds to /(x) = /,,(x) x and so has asymptotic variance

V( dm,,, f ) = /O0 x2f( x) dx = (1 - E ) + €1“ x2h( x) dx. -m -m

(5.7) For c > 0 the asymptotic variance (5.7) can be arbitrarily large since h is arbitrary, and so the sample mean is a very nonrobust estimator of location for this type of uncertainty.

The basic problem with the sample mean is that it has an unbounded influence curve so that too many large observa- tions (or outliers) can destroy i ts efficiency. This could be corrected by employing the sample median, which has the influence curve !,,(x) A sgn(x). For the sample median the asymptotic variance (assuming f is continuous at the origin) is

V( dm,, f ) = 1/4f2(0) d 1 II

4(1 - t)’+’(O) 2(1 - E)’

- -

So, the sample median is certainly more robust than the sample mean in this case; however, i ts efficiency relative to the sample mean is only 64 percent at the nominal model. Ideally, we would like an estimator that is almost as effi- cient as the sample mean at the nominal and has the robustness of the sample median away from the nominal. It turns out that these goals can be achieved as we see below.

To correct the possible nonrobustness of classical estima- tors of location, Huber proposed the design of location estimators using a minimax formulation, viz.

min max v( d , f ) . c f € F

(5.9)

Within some assumptions on 9, the solution to (5.9) is given by the influence curve dR = - fL/fR where f, i s a least favorable density for location estimation defined by

(5.10)

The pair ( d R , f R ) is a saddlepoint solution for (5.9) provid- ed -log f, i s convex; i.e., under the convexity of - log f,, (e,, f R ) satisfies

v( I f , d v( / R , f R ) d v( e , f R ) (5.11)

for all f E 9 and all d . Note that (5.11) implies that, not only does d, have minimax asymptotic variance but also its variance is upperbounded over 9 by its variance when f, is the true density. This means that the performance of the estimator will never be worse than V(d,, f,) over F. Also note that, since I( f ) is Fisher’s information for location, f, has the interpretation of being the density for which the observations are least informative about 6.

For the particular case of €-contaminated Gaussian data (i.e., 9 from (5.6)), the least favorable density turns out to be given by (4.21) with g = + as discussed in Section IV. In this case, dR i s a soft-limiter

where a i s from (4.22) and (4.23), and so !,(x) d a’, and !;(x) = 1 for 1x1 < a and /:(x) = 0 for 1x1 > a. Thus the numerator of V(d,, f ) is

ti( x) f( x) dx = (1 - €)jm !;(x)+( x) dx - m

+ E j-m /;(x) h( x) dx m

d (1 - c ) j m d;(x)+(x) dx + Ea’ -m

(5.13) ’-m

and in the denominator we have

I-r, /;(x) f( x) dx = / a f( x) dx - a

= (I - c)/:$J( x) dx + h( x) dx - a

= j- d;( X) fR ( X) dx. (5.14) m

m

From (5.13) and (5.14) the left-hand inequality of the sad- dlepoint condition (5.11) is readily seen to hold in this case. (The right-hand inequality of (5.10) is just a restatement of ( 5 3 ) Similarly, for the r-contamination model with + re- placed by any density g (with - logg convex) a saddlepoint solution to (5.9) i s given by (- fL/ fR, f,) with f, from (4.21).

Note that the influence curve (5.12) of the robust estima- tor for r-contaminated Gaussian noise combines features of the influence curves for the sample mean and the sample

468 PROCEEDINGS OF THE IEEE, VOL. 73, N O 3, MARCH 1965

Page 37: Robust Techniques for Signal Processing - A Survey

influence curves for robust estimation in these models are shown in Fig. 25(b) and (c), respectively. Appropriate mod- ifications of the influence curve for robust estimation in dependent noise and for asymmetric noise have been con- sidered by Portnoy [152], [I531 and Collins [158], respec- tively. Also, the estimate of (5.3) i s straightforwardly mod- ified to account for the time varying signal s,; . ., s,, via

Assuming - n

this estimate has the same properties as in the constant-sig- nal case except that the asymptotic variance is V(!, f ) / C . Thus the minimax robustness problem for the time-varying case is identical to that for the constant-signal case [131].13

From a practical viewpoint, the robust M-estimator of signal amplitude has the disadvantage of being a "batch" procedure; i.e., all data must be stored in order to minimize

n

median (see Fig. 25(a)). It has the boundedness of e,,, but it has the linear shape of e, , for 1x1 d a. (This eliminates the sensitivity of e,,,, to the value of f at zero.) For the case E = 0.1, the limiting point a is approximately 1.1 and the efficiency of the robust estimator at the nominal model relative to the sample mean is 92 percent. Moreover, the worst case variance of fK over 9 is V(!,, f,) G 1.5 com- pared with

sup v( dm,,, f ) 2 1.9. f€S

Thus we see that this estimator does indeed achieve the goals set forth for robust estimation.

Robust M-estimators can be found for uncertainty classes other than the €-contaminated class by minimizing Fisher's information over the uncertainty class and then choosing the estimator accordingly. It is not necessary that the uncer- tainty class contain only distributions with density func- tions, although the least favorable will always correspond to a continuous distribution. A variational method for mini- mizing Fisher's information i s discussed in [28, ch. 51. Exam- ples of useful uncertainty classes for which solutions are known are ppo in t classes, which consist of the set of all noise distributions that place a fixed amount of probability on a given interval [I761 and the class of noise distribution functions F that satisfy

sup I F ( x ) -@(,)I d € - - m < x 4 m

where @ is the unit Gaussian distribution function [6].

i - I

iteratively. This disadvantage can be overcome by consider- ing a class of recursive estimators of the stochastic ap- proximation type. Robustness theory for this type of estima- tor was first considered by Martin [I781 and these ideas were developed further by Martin and Masreliez [I761 and by Price and VandeLinde [179].

To discuss these results, we first consider a generalization of the minimax problem of (5.9)

min sup v( 8 , f ) (5.15)

where 8 is the class of all asymptotically unbiased estima- tors of B and where v(8, f ) is the asymptotic variance of 8. A sufficient condition for 8, to solve (5.15) is that it, together with some f, €9, satisfy the saddlepoint condi- tion

aEe fEF

for all 8 E 8 and f €9. It is possible that (5.9) and (5.15) have different solutions since, in (5.9), consideration is restricted to M-estimates. However, since the Cramer-Rao bound implies v(8, f,) 2 I / / ( f,) = V(!,, f R ) for all 8 E 8, we see that the minimax robust M-estimator also satisfies (5.16) and so i s in fact a minimax robust estimator among all asymptotically unbiased estimators. However, there is also another saddlepoint (8,, f,) for (5.15) in which 8, is not the M L E for f, but rather is a recursive estimator.

In particular, suppose 9, f,, and !, are as in the robust M-estimation formulation and let 8 F ( X ) = 8,, be defined for each n by

(5.17)

with do arbitrary. Note that 89 i s computed recursively (i.e., at the i th sampling instant 8, i s computed from and %,). Then under regularity conditions the asymptotic

I3The problem in which s1;.., s, is also unknown and to be estimated has also been considered in several studies (see for example, Tsybakov [177]).

KASSAM AND POOR: ROBUST TECHNIQUES F O R SIGNAL PROCESSING 469

Page 38: Robust Techniques for Signal Processing - A Survey

variance of 8p for a given noise density f is given by [I791

Note that ~(82, f,) = I//( f,) so that 8? is asymptotically optimum for f,. Moreover, it can also be shown that

. ( d ? , f ) d v ( 8 ? , f , ) (5.19)

for all f E F, so that (e?, f,) is also a saddlepoint solution to (5.15). Thus we have the interesting conclusion that the recursive estimator of (5.17) is minimax robust over 9 just as is the M-estimator based on C,. However, although both estimators have the same worst case performance, I//( f,), one can see that their performances ( V(C,, f ) and .(e?, f ) ) generally differ for given f # f,. For the c-contaminated Gaussian case it is easily shown that ~(82, f ) > V(C,, f ) whenever

/:ah( x ) d x > 0.

Thus for the computational advantage of the recursive structure one pays the penalty of some lost efficiency when worst case conditions are not present. However, for exam- ple at c = 0.1 this lost efficiency is negligible (< 1 percent) at the nominal density +.

The appropriate modification of the algorithm of (5.17) for the time-varying case is

' i = ' ; - l + s ~ c R ( X ; - 8 i - , s ; ) / [ ' ( f R ) ~ , r ~ ] ~

i = 1 , 2 ; . . , n . (5.20)

This estimator has not been studied analytically; however, simulations for the case of a sinusoidal signal in c-con- taminated Gaussian noise indicate that the performance of (5.20) is comparable to that for the constant-signal case (Martin [178]). Heuristically, in either this or the constant- signal case robustness against contamination is achieved by inserting the soft-limiter C, into the feedback loop that incorporates the residual X i - 8j-1si into the updates of the amplitude estimate (see Fig. 26). If it were not for this limiter, then when 8; gets close to the true 19 "too many" large noise values would tend to drive 8, away from 0.

where F, i s the sample cumulative distribution function (cdf)

F , ( x ) & - ( # X ; < X ) , --OO < X < -OO (5.22) 1 n

and where T is some functional mapping the set of cdfs into I?. For example if we define

then the M-estimate 8, is T(F,) since

/-m L ( x - e) dF,( x ) = - 1 " L ( X ; - 8 ) . m n .

1-1

If we assume that L is convex and symmetric about zero and if F, is the common marginal cdf of X , = Ni + 0, then for (5.23) we have T ( 6 ) = 8.

For a wide class of estimates of the form T(F, ) it turns out that, when F is the true data distribution, h( T(Fn) - T ( f ) ) is asymptotically normal with zero mean and variance

A( F , T ) = Jrn [ IC( x , F , T)]' dF( x ) . (5.24) -m

Here IC is the influence curve of the estimate (the reason for the earlier use of this term will become obvious below) defined by

IC( x , F , T ) = Iim T((1 - E) F + cFJ - T( F )

(5.25) f + O E

where F, is the cdf of a random variable taking on only the value x.14 The influence curve for an M-estimate d L i s

IC( x , Fo, T ) = q X - e)

(5.26) - /C!( x - e) dF,( x )

which gives the asymptotic variance V(C, f ) of (5.4). Other useful classes of estimates of the form T( F,) are the

so-called L-estimates, which estimate 0 by linear combina- tions of the order statistics (i.e., the observation sequence put in numerically increasing order), and /?-estimates which are based on the ranks of the data. Explicit representations of these type estimators in the form T(F) and the corre- sponding influence curves can be found in [28]. For any fixed noise density, an asymptotic variance equal to I//( f ) can be achieved within each of these classes. Of particular interest in the class of L-estimates is the so-called a-trim- med mean, which estimates 0 by first removing the [an] largest and [an] smallest samples (0 d a < $), and then computing the sample mean of the remaining sample. The a-trimmed mean can be written as Tu(Fn) where

Tu( F ) = - 1 - 2a

1 f - " F 1 ( x ) d x (5.27)

Fig. 26. Robust recursive signal-amplitude estimator for e contaminated Gaussian errors.

The robustness properties of types of estimators other than the M-estimators and stochastic approximation estima- tors can be studied in a general framework. In particular, for the constant-signal case, many estimators of practical inter- est can be written in the form

an = T(Fn) (5.21)

where F-' is defined by F ' ( x ) = inf { y l F ( y ) 2 x } . For a symmetric noise distribution FN the influence curve of the a-trimmed mean is given by

(5.28)

where

14Note that, in addition to its role in the asymptotic variance formula (5.24), the influence curve characterizes the sensitivity of T(F,) to the incorporation of a datum x into the estimate; i.e., it quantifies the influence of such a datum on the estimate. Hence, the term influence curve.

470 PROCEEDINGS O F THE IEfE. VOL. 73, NO. 3, MARCH 1985

Page 39: Robust Techniques for Signal Processing - A Survey

Fig. 27. Application of robust amplitude estimation to image enhancement. (a) Original image. (b) Noisy image (Gaussian background noise with impulsive outliers). (c) Running mean processing. (d) M-estimate processing

(5.29)

Note the similarity of this influence curve to that of the robust M-estimator for €-contaminated Gaussian noise. In fact, by choosing

a = jdm f,( x ) dx

where f, i s the least favorable density from (4.21) and a i s the limiter breakpoint from (4.22), (4.23), we see that I /C(x, Fe, T,)I < IIC(x, F R , 8 , Tall where

F R , B ( X ) = j-; f,( Y - 0) dY. W

It follows that this a-trimmed mean is also a minimax robust location estimate for €-contaminated Gaussian noise. Unfortunately, there is no known time-varying analog to the a-trimmed mean that carries its minimax property.

A different, intuitive, notion of robustness of an estimator is that small changes in the data set x1;..,xn should not change the estimate much. (This notion is sometimes called resistance.) This property is assured for an estimate T(f,,) if the functional T i s continuous. Thus one way of defining robustness of an estimator of the type T(f,) at a given nominal F, is in terms of the continuity of Tat f,. A related robustness notion arises if we view an estimator 8 ( x ) as a mapping from the marginal distribution of the observations X,; . . ,X , , to the distribution of the estimate 8 ( X ) . Robust-

KASSAM AND P O O R . ROBUST TECHNIQUES F O R S I G N A L P R O C E S S I N G

ness can be formulated in this context by requiring this mapping to be continuous in some way; i.e., a small change in the data distribution should cause only a small change in the distribution of the estimate. It turns out that, for esti- mates of the form T( f,,), these two continuity definitions of robustness are equivalent (within the proper definitions of continuity). The notion of robustness as a continuity prop- erty was introduced in the context of parameter estimation by Hampel [3], although similar ideas were set forth in the context of signal detection in Root’s earlier study of stability in detection [2]. Robustness of this continuity type is usual- ly known as qualitative robustness.

The robust estimators discussed above can be extended to the estimation of signal parameters other than amplitude (see, e.g., Huber [28], Kelly [147], and El-Sawy [146]). Robust estimators have found numerous applications in statistics and signal processing. For example, location M-estimates, a-trimmed means, and modifications thereof have been applied successfully to the problem of image enhancement by Bovik, Huang, and Munson [I801 and by Lee and Kassam [181], [182]. This is a natural application for robust methods since image noise typically consists of a Gaussian-like back- ground with occasional impulsive or “spiky” components. For example the imagS5 of Fig. 27(a) is shown corrupted by a combination of additive Gaussian and impulsive noises in Fig. 27(b). The effects of smoothing this image with a running mean (analogous to the sample mean) and with an M-estimate are shown in Figs. 27(c) and (d), respectively.

‘’The images in Fig. 27 were produced using the facilities of the GRASP Laboratory, Department of Computer and Information Sci- ences, University of Pennsylvania.

471

Page 40: Robust Techniques for Signal Processing - A Survey

The beneficial effects of robustifying the estimate are quite dramatic in this case. Another interesting investigation of the use of robust estimators, in a digital F M receiver, has been considered by Abend et a/. [183].

B. Robust Nonlinear Recursive Filtering and Identification

One of the most commonly used signal processing al- gorithms is the discrete-time Kalman filter, which is based on the linear observation model

Y, = H,X, + V,, n = 0 , 1 , 2 , . . . (5.30)

where, for each n, Y, i s an r X 1 observation vector, H, is the observation matrix, X, is the n x 1 state vector, and V, i s the observation noise; and on the state model

X,+, = F,X, + W,, n = 0,1 , .. . (5.31)

where F, is the one-step state transition matrix and W, is the state noise. Assuming that X,, { W,}:-,, and { V,}:-, are all Gaussian, are independent of one another, and that { W,}?=, and { V,}?=, are sequences of independent vec- tors, the optimum (MMSE) estimators of X, and X,+, given V,; . ., Y, are given recursively by the well-known Kalman filtering algorithm

and

(5.33)

where R, = cov(W,), M,, = F,-,P,-,F~-, + Qn-,, Q, = COV( V,), and Pn = cov(X,,, - X,,). P, is found recursive- ly from a standard formula [I&]. The relations (5.32) and (5.33) are called the measurement update and time update, respectively.

O f course, the linearity of (5.32) and (5.33) follows from the assumptions of Gaussian statistics for the state and the observation noise. If either of these quantities has a distri- bution that deviates from this nominal assumption in that it allows an unexpected number of large observations, then the Kalman estimator may perform poorly. This may be seen from (5.32). In particular, the prediction residual (Y, - H,~nln-l) will contain outliers if either of the Gaussian assumptions is violated towards a heavier tail behavior. Since this residual is fed directly into the estimate, a dis- torted value can cause severe degradation of the estimation performance; and, because of the dynamics, such errors will propagate. However, there is an additional dimension to this problem in that outliers in the state are of interest and should be tracked rather than limited, and so additional considerations arise here. A method for robustification of the Kalman filter against modeling uncertainty has been proposed by Masreliez and Martin in [185].

In view of the ideas described in the preceding subsec- tion, a natural way to try to protect against the detrimental effects noted in the above paragraph is to somehow place a limiter or similar nonlinearity in the feedback loop of the filter (5.32), (5.33). This would limit the effects that any outlier could have on performance, and this, in fact, is the way robustness is achieved in this model if all uncertainty is in the observation noise distribution. However, because of the vector nature of the Kalman filtering problem, some

scaling and transformation of the residual is necessary both in modeling the uncertainty and in processing the residuals for robust estimation. In particular, for ppoint uncertainty, the robust version of (5.32), (5.33) when the prediction errors X,, - XnI,-, are Gaussian but the innovations (Y, - H,X,ln-,) have uncertain distribution (this case corre- sponds to uncertainty in the observation noise) is of the form

'"1, = i n l n - 7 + MnH,TTr*(vn) (5.34)

Xrl+ll" = F,X,l, (5.35)

vn = T,( V, - ~ n i n I n - 7 ) (5 .36)

with

where T, i s a scaling transformation described in [I851 and where 9 ( v , ) is the vector whose j t h component is the j t h component v,,~ of v,, replaced by $ ( v , , ~ ) , fP being the influence curve for robust location estimation in the ppoint model illustrated in Fig. 25(b). The error covariance of this estimator for ppoint uncertainty is always less than its value when the components of the transformed residual vector are iid with the least favorble marginal distribution. Moreover, this worst case covariance is the optimum covari- ance for the case in which the transformed residuals have this least favorable property. However, this estimate does not provide a saddlepoint for the mean square estimation error because of the constraints of the linear model as is discussed in [185].

For the case in which the observation noise is Gaussian but the residual distribution is uncertain due to uncertainty in the state distribution, the Kalman filter is robustified differently than for the case of uncertain observation noise. In particular, without going into detail, the measurement update incorporates Y, linearly into the estimate of in,, but the transformed residual v,, is limited. The linear incorpora- tion of Yn is done because an outlier in Y, indicates an outlier in X,, which should be incorporated into the esti- mate of X,. O n the other hand, an outlier in v, does not necessarily indicate a bad prediction (inI,,-,) and so should not be treated as such.

For further details and simulation results for robust Kalman filtering, the interested reader is referred to [185]. Related results can be found in VandeLinde, Doraiswami, and Yurtseven [I&], Ershov and Lipster [187], Ershov [188], Levin [189], and Tsai and Kurz [I%].

A problem related to Kalman filtering is that of system parameter identification. In this problem we have a set of scalar observations V,;.., Y, and a set of system input vectors X,; . . , X , that are related through the equation

~ = s ( c , X , ) + N;, ; = I ; . . , n (5 37)

where c is a vector parameter and N,; . ., N, is an iid noise sequence with marginal pdf f. The system identification problem is t o estimate c from observation of V,; . -, Y, and X,;. .,X,. Note that (5.37) i s a generalization of the ampli- tude estimation problem of (5.1) in which X;, 8, and s i in (5.1) correspond, respectively, to y , c, and X; in (5.33, and

As in the estimation of signal amplitude, the conven- tional estimators of c in (5.37) are of the form of M-estima- tors, namely,

s( c, X,) = cx;.

472 PROCEEDINGS O F THE IEEE. VOL. 73, NO. 3. MARCH 1935

Page 41: Robust Techniques for Signal Processing - A Survey

I n

with L an error-weighting function. We get least squares estimation with L ( x ) = x 2 , least moduli estimation with L ( x ) = 1x1, and maximum-likelihood estimation with L ( x ) = -log f ( x ) . If the X ; are iid and L is convex, then within mild regularity conditions the estimate S of (5.38) is strongly consistent. Moreover if s(c ,X ; ) is linear (i.e., s(c, X i ) = c r X j ) , then f i ( S - c) is asymptotically Gaussian with zero mean and covariance matrix

P-lv(e, f ) (5.39)

where /3 = E { X,%:} and V ( L , f ) is from (5.4). Thus just as in the location estimation case the minimax robust estimate of c in (5.37) for a given noise uncertainty class .F is the optimum M-estimate (L, = - fL / fR) for the least favorable noise density,

As noted in the preceding subsection, since the M-esti- mator is a batch process, we are often more interested in recursive estimators for on-line applications. In this case (with linear s(c, X , ) ) the appropriate robust recursive identi- fier for a class S i s given by

cn+l = ?n - P-lSnLR( yn - 5( t n , nl( ' R ) ]

(5 4 which, together with f,, also gives a saddlepoint for the minimax robustness problem. For further details on this and related topics the reader is referred to the interesting survey by Poljak and Tsypkin [38].

The problem of system identification is closely related to the problem of estimating the parameters of a linear time series. An excellent survey of methodology for robust esti- mation in this context is found in [37]. To illustrate the type of results that can be obtained, we mention briefly the problem of estimating the parameters of an autoregression. Specifically, suppose we have observations

V, = e + X ; , i = 1,2 , - ,n (5.41)

where 0 is a location parameter and { X ; } is a pth-order autoregression

X , = + , X ; - , + hZX,-, + . . . + + p X , - p + C; (5.42)

where + = (+,,...,(pp)'is a vector of constants, and { c ; } is an iid innovations sequence with marginal pdf f . M-esti- mates of + and 0 can be computed by finding

(?A)

where y e/(l - +, - ... - + p ) and 5 is an estimate of the innovations scale parameter (see [37]), and then setting 8 = +(I - 6, - . . . -6J. These estimates of B and $I are consistent and are asymptotically Gaussian and indepen- dent with

and

ncov(6) - V(L,f )C- ' /u, ' (5.44)

where V ( L , f ) is the asymptotic variance expression for M-estimation of location from (5.4), u,' is the variance of the innovations, and C is the p X p covariance matrix of the { X ; } process when u,' = 1.

Note from (5.43) that LR = - fL/ f , and

gives a saddlepoint for the minimax robust estimation of 6 when the innovations distribution is uncertain. However, this is not necessarily so for robust estimation of + because of the u,' term in (5.44). It is interesting to note from (5.44) that a heavy-tailed innovations distribution may actually be beneficial in estimating +, since cov(6) depends inversely o n u,'. This is not surprising if we note that outliers in the innovations should actually aid in the identification of $I just as the insertion of an impulse into a system allows one to identify its impulse response. Thus we have here a situation in which impulsive phenomena are beneficial to inference.

Extensions of the above ideas to problems of estimation in ARMA models, nonlinear models, models with additive observation noise (in which outliers are again detrimental), and models of unknown order are discussed in [37]. Similar ideas can also be applied to other problems of time series analysis such as forecasting [37] and spectrum estimation [35]. Connections between the problems of robust Kalman filtering and robust time-series regression have been noted in a recent paper by Boncelet and Dickinson [191].

VI. ROBUST DATA QUANTIZATION

Data quantization is a necessary function of systems which digitally process or transmit signals. The optimum design of quantizers based on minimum-distortion criteria has been considered extensively over the past three de- cades and a number of references on this subject can be found in surveys by Morris [I921 and Gersho [I931 and in a special issue of the IEEE TRANSACTIONS ON INFORMATION THEORY (Gray [194]). More recently, designs for quantizers which are optimum for signal detection or estimation pur- poses have also been developed [163], [195]-[197].

Optimum quantizer design is based primarily on statisti- cal definitions of optimality, such as minimum mean dis- tortion [I981 or maximum divergence [I%]. Thus the opti- mum designs usually require an accurate statistical model for the data to be quantized. Since such models are rarely exact, the study of quantizer design for inaccurate models is of practical interest. One approach to this problem is that of adaptive quantization (see, for example, [ I s ] , [2oO]), which is appropriate for very inaccurate models or for situations in which data statistics are changing significantly over moderate time periods. An alternative approach which is primarily of interest when there are relatively small inac- curacies in the statistical model or when a fixed structure is preferred, is a game-theoretic one in which a quantizer with best worst case performance is sought. As demon- strated in the preceding sections, this general approach has also been applied successfully to many other problems of signal processing with inaccurate statistical models, and the

KASSAM AND POOR: ROBUST TECHNIQUES F O R SIGNAL PROCESSING 473

Page 42: Robust Techniques for Signal Processing - A Survey

resulting designs are usually robust. In this section, we consider the problem of minimax design of quantizers for imprecisely modeled data. In particular, we survey several recent results pertaining to the problem of minimax dis- tortion quantization. These results include both asymptotic (as the number of quantization levels becomes infinite) and nonasymptotic treatments of this problem.

A. Robust Quantization for a Small Number of Levels

An M-level quantizer Q can be represented by a set of M output levels ql, q2; .., qM and a set of (M + 1) input breakpoints to, t,; .., tM, satisfying - w = to < t, < . . . < tM- , < tM = + 00, where the quantized value of a real input x i s given by

Q ( x ) = q k , i f x E ( t k _,, t k ] , k = I ;.., M. (6.1)

(see Fig. 28). Thus the design of an M-level quantizer is an

out

Fig. 28. Input/output characteristics of an M-level quan- tizer.

optimization problem on RZM-'. The most common quan- tizer design criterion for a random input X is to choose the quantizer parameters to minimize a mean-distortion quan- tity

E { D [ X t Q ( X ) l l (6.2)

where D[. ; ] i s some appropriate measure of distortion. Usually D [ - ; ] is a difference distortion measure (i.e., it depends only on the difference [X - Q(X)D and the most useful of these are the pth-difference distortion measures given by D[a,b] = la - blP.

The study of minimum-distortion quantization is exem- plified by the classical work of Max [I%], in which design conditions for the tk and qk of minimum-distortion quan- tizers are derived. In particular, for a wide class of dif- ference distortion measures (including the pth-difference ones), M a x shows that the breakpoints should be chosen to satisfy

and he also gives a second set of necessary conditions which, together with (6.3), give a set of nonlinear equations to be solved for the optimum quantizer parameters. For example, if we consider mean square distortion ( D [ a , b ] = la - bI2) and if X has a pdf f , then the necessary conditions are given by

qk = ltk f ( x ) dx , k = l ; . . , M (6.4)

k - 1

that is, qk is the centroid of (fk-l,tk] weighted with f . (An equivalent interpretation is that q k = E { X I X E ( t k - l , tk]}). Many important refinements and generalizations of this theory as well as studies of performance, approximately optimal design, etc., can be found in the literature, and again the reader is referred to [192]-[I941 for further discus- sion.

Of course, designs based on minimizing the quantity of (6.2) will depend on knowledge of the probability distribu- tion of X , as can be seen, for example, in (6.4). Thus, as discussed in the Introduction, when this distribution is not known exactly it is necessary to seek an alternative design strategy to minimizing (6.2). If we assume that the probabil- ity distribution of X lies in some uncertainty class 9, then a useful design objective is the minimization of the alternate quantity

Several recent studies have considered various aspects of the minimization of (6.5) and these are discussed in the following paragraphs. For the purposes of discussion, we will consider only the particular case of mean-square dis- tortion, although all of the cited studies consider more general distortion measures as well. Thus we will be con- sidering the problem

min sup jm 1 x - Q( x ) )'dF( x ) (6.6) Q E ~ M F E B - m

where 1, denotes the class of all M-level quantizers. The first study to consider (6.6) in the context of minimax

robustness is that of Morris and VandeLinde [201], in which B is taken to be the class of all possible probability distri- butions on a fixed interval [ - V, VI. In this case, it is shown in [201] that the minimax quantization problem (6.6) is solved by the M-level uniform quantizer on [ - V, VI, which is given by

t k = - V + Z k V / M , k = l ; . . , M - l

q k = f k - V / M , k = I , . . . , M - I (6.7)

and qm = V - V/M.

The solution to (6.6) was considered next by Bath and VandeLinde in [202], in which the class B is taken to be a unimodal generalized moment constrained (UGMC) set; i.e., B is assumed to consist of the distribution functions which are unimodal (with mode zero) and which satisfy the generalized moment constraint

where the constraint function p is symmetric, continuous, is strictly increasing on (0, co), and satisfies p(0) = 0 and p ( x ) + 00 as x + co. In particular, it is shown in [202] via the Lagrange duality theorem that the minimax quantizer for this problem is given by the solution to

474 PROCEEDINGS OF THE IEEE, V O L 73, NO. 3, MARCH 1985

Page 43: Robust Techniques for Signal Processing - A Survey

where c is from (6.8) and

An O(M) algorithm for solving (6.9) i s also given in [202], and it is demonstrated numerically that the resulting minimax quantizer can perform much better in the worst case over B than both the uniform quantizer and the quantizer which is optimum for Gaussian data.

B. Asymptotic Robust Quantization

Further work on the minimax-distortion quantization problem considers the asymptotic case as the number, M , of quantization levels becomes infinite. Assuming the data are confined to an interval [ - V, VI, i t is convenient to represent (and to implement) a quantizer as an increasing invertible function G:[- V, VI + [- V, VI, followed by a uniform quantizer, which is followed in turn by the inverse of C (see Fig. 29). The function G is termed a compressor and i ts inverse an expander, so that the whole scheme is termed companding. In general, the compressor G can be any invertible function; however, it is usually assumed to have a continuous derivative g. It is also common to as- sume, for simplicity, that the data distribution F has a density f. Under these and further mild regularity condi- tions it can be shown that the mean-squared error associ- ated with the companding scheme is asymptotically of the form D( f,g) . M P 2 , where the functional D( f,g) i s given by

D( f ,g ) = - j v V2 f ( ~ ) [ g ( x ) ] - ~ d x . (6.11) 3 - v

The function g describes the relative density of the quanti- zation intervals within the range of the data.

By way of (6.11), an asymptotically optimum compressor curve G can be chosen by minimizing over g. Straightfor- ward minimization (see, e.g., Gersho [193]) yields that the minimizing compressor is given by

which yields a value of (6.11) of

D( f,g,) = l( f 1 l 3 ( x ) dx) (6.13) 3

12 - v

where go = G. Since it is known that (6.13) gives the

limiting distortion of the minimum mean-square-error quantizer (see, for example, Bucklew and Wise [203]), it follows that the companding structure causes no perfor- mance loss for optimum quantization in the asymptotic case.

Note that, as one would expect, the optimum compressor characteristic of (6.12) depends on an exact knowledge of the probability distribution of the data. Thus if, as above, F i s known only to belong to some class 9 of possible data distributions, then it is reasonable to replace the problem of minimizing (6.11) over g with that of minimizing over g the worst case over B of (6.11). That is, i t is of interest to consider the problem

min ( SUP D( f , g)) (6.14) gc9 F E F

where .F= { f l f = F', f~ B } , and where g i s a set of admissible compressor curve derivatives.

The problem in (6.14) was first considered by Bath and VandeLinde in [204], where the case in which B is a UGMC set as in (6.8) i s treated. It follows from [204] that the minimax compressor for this case is given by

where

g;*( x ) = ( v/3'/2)[ X: + X;p( x ) ] -"2 (6.16)

with A: and X: chosen to solve

and

Here, the constant c and the function p are from (6.8). It is noted in [204] that for the case p ( x ) = x z , the solution to (6.17) involves only finding the root of a simple tran- scendental equation. It is also shown in [204] that the worst case performance over the UGMC set B of the minimax compander is much better than that of the corresponding optimum companders for the Laplace, uniform, and Gaus- sian distributions, However, surprisingly, it is found that "robustified" versions of the classical A-law and p-law companders (see, for example, Cattermole [205]) perform nearly as well in their worst case as does the minimax compander.

A slightly different approach to solving the robust com- panding problem is proposed by Kazakos in [206] and [207]. In particular, [206] and [207] consider uncertainty classes B for which saddlepoint solutions to (6.14) can easily be demonstrated. Note that D( f,g) is linear (and hence con- cave) in f and is convex in g. In [206] the following uncer- tainty class i s considered (here V i s taken to be equal to 1):

B = ( ~ F ( X k + l ) - F ( x k ) = P k , k = 0 , 1 , " ' , N - 1

F ( X ) + F ( - X ) = I , O = X ~ < X , < .. . < x N = 1 } (6.18)

where N, the xk , and the pk are fixed and known with N-1

P k = I . k = O Fig. 29. Configuration for companding quantization

KASSAM AND P O O R . ROBUST TECHNIQUES F O R SIGNAL PROCESSING 475

Page 44: Robust Techniques for Signal Processing - A Survey

This class is an example of the ppoint class discussed in previous sections. Kazakos shows that the member of (6.18) which has the uniform density on each of the intervals ( x k - , , x k ] and its corresponding optimum compressor from (6.12) form a saddlepoint for (6.14) in this case. That is, with

f2*(X)=Pk/(Xk+l - x k ) t X E ( X k , X k + l ] ,

k=O,I ; . . ,N- l

(6.19) and & from (6.12) with f = fp, we have

D( f&) Q D( f?&) Q D( K g ) (6.20)

for all g E 9. Note that the compressor curve q corre- sponding to this & will be piecewise linear. Also note that, of course, the existence of ($*, 8 ) satisfying (6.20) is equiv- alent to the condition

In [207], the saddlepoint properties of (6.14) are con- sidered in a more general setting. in particular, it is noted in [207] that within regularity conditions on 9 and Y, the equality of (6.21) holds for general convex classes 9. Thus in view of (6.13), saddlepoint solutions to (6.14) can be sought by looking for solutions to the alternate problem

where

]( f ) = /" f1I3( x ) dx. - V

(6.22)

(6.23)

That is, under mild conditions, a solution f * to (6.22) and its corresponding optimum compressor characteristic g* from (6.12) form a saddle point for (6.14). As was noted in [208], since ]( f ) is of the form

with C concave, solutions to (5.22) are those "closest" to the uniform density on [ - V, VI. Thus solutions for several uncertainty classes are straightforward. For example, with

9= { f l f = (I - 4 ) 6 + Eh} (6.24)

where 6 and c are fixed and h i s arbitrary, the density solving (6.22) is given by

f ; t ( x ) = max{( l - c ) & ( x ) , m } (6.25)

where m is a constant chosen so that r;t integrates to unity (compare with (2.35)).

Unfortunately, as pointed out in [209], the asymptotic formulation of [206], [207] is weakened by the fact that, for some models (e.g., c-contaminated data), the minimum of the maximum asymptotic error is different from the mini- mum of the asymptotic maximum error due to the discrete nature of the finite-M problem. (This problem does not occur for the UGMC model of Bath and VandeLinde.) This problem is corrected in [209], and the solution for c-con- taminated data are still of the form of (6.25) but with a higher degree of limiting. A typical solution is shown in Fig. 30. Note that, for this case, the levels of the robust quan- tizer are distributed more or less as those of the nominal quantizer near the center of the range but they are spaced

-3 -2 -1 0 1 2 3 Input b l u e , x

Fig. 30. Nominal and robust output levels densities for c-contaminated truncated Gaussian data (c = 0.1).

uniformly (and closer together than for the nominal quan- tizer) near the ends of the range. This latter property guards against a larger than expected number of data points near the ends of the range. Thus the robust quantizer is a mixture of the nominally optimum quantizer and the uni- form quantizer (which is universally minimax). Numerical results illustrating the effectiveness of the robust quantizer are given in [209].

VII. CONCLUSIONS

In this paper we have considered minimax robustness in the context of the signal processing tasks of estimation, detection, and data quantization. We have seen that these robustness formulations take two basic forms: robustness with respect to uncertain second-order statistical properties (e.g., spectral properties) of signals or noise and robustness with respect to uncertainty in the marginal distribution of the noise or signal process. In the first of these cases, the robust methods take the form of the linear procedures discussed in Sections I I and Ill, whereas in the second case, nonlinear procedures are called for. In either case the typical robustification procedure has the effect of lowering the sensitivity of a nominally optimum procedure by tem- pering those characteristics that are accentuated by the nominal model. Thus in a nominally Gaussian noise model with a small fraction of "outliers," enough limiting is intro- duced into the optimum procedure to keep these outliers from destroying the action of the procedure. The fre- quency-domain robustness methods can also be thought of in the same way in which the gain of the appropriate filter is reduced in certain spectral regions to limit the effects of a more than expected amount of energy in those regions (i.e., "spectral outliers").

The relationship of these two types of robustness has been discussed by Franke and Poor in [59] in the context of estimation. in [59] it is noted that, if one knows only spectral or other second-order properties, then linear esti- mation procedures are globally minimax over all estimation schemes. It is only when distributional information is pro- vided that nonlinearity arises in the minimax solutions. Thus the robust filters of Section Ii are globally minimax over all filters (linear or nonlinear) and all random processes (Gaussian or otherwise) with the given spectral properties. Note that the relevant observation or noise processes in Section V have very special spectral characteristics (they are usually white), and only their marginal distribution is al- lowed to vary. Very little progress has been made in dealing

476 PROCEEDINGS O F THE IEEE. VOL. 73. NO. 3, MARCH 1985

Page 45: Robust Techniques for Signal Processing - A Survey

with simultaneous spectral and distributional uncertainty models (an exception is a recent paper by Moustakides and Thomas [154]), although this certainly gives rise to an inter- esting class of problems from a practical viewpoint.

REFERENCES

[l] H. L. Van Trees, Detection, Estimation, and Modulation Theory, I. New York: Wiley, 1968.

[2] W. L. Root, ”Stability in signal detection problems,” in Proc. Symp. in AppliedMathemarics, vol. 16, American Math. SOC. 1964.

[3] F. R. Hampel, “A general qualitative definition of robustness,” Ann. Math. Stat., vol. 42, pp. 1887-1896, 1971.

141 Ya. 2. Tsypkin, “Adaptive optimization algorithms when there i s a priori uncertainty,” Automat Remote Contr., vol. 40, no. 6, pp. 857-868, June 1979.

[SI J . W. Tukey, “A survey of sampling from contaminated distributions,” in Contributions to Probability and Statistics, I. Olkin et a/., Eds. Stanford, CA: Stanford Univ. Press, 1960.

[6] P. J. Huber, “Robust estimation of a location parameter,” Ann. Math. Stat., vol. 35, pp. 73-104, 1964.

[7] -, “A robust version of the probability ratio test,” Ann.

[8] L. A. Zadeh, “General filters for separation of signals and noise,” in Proc. Symp. on Information Networks (Poly. Inst. of Brooklyn, Apr. 1954), pp. 31-49.

(91 W. L. Root, ”Communications through unspecified additive noise,” Inform. Contr., vol. 4, pp. 15-29, Mar. 1961.

[IO] -, “Some notes on jamming,” MIT Lincoln Lab. Tech. Rep. 103, 1956.

[ l l ] M. C. Yovits and J, L. Jackson, “Linear filter optimization with game theory considerations,” in IRE Nat. Conv. Rec., pt. 4,

[12] N. J. Nilsson, “An application of the theory of games to radar reception problems,” in IRE Nat. Conv. Rec., pt. 4, pp. 130-1 40, Mar. 1959.

[13] L.-H. Zetterberg, “Signal detection under noise interference in a game situation,” IRE Trans. Inform. Theory, vol. IT-8, pp. 47-57, Sept. 1%2.

[I41 N. M. Blachman, “Communication as a game,” in IRE Wescon Conv. Rec., pt. 2, pp. 61-66, Aug. 1957.

[IS] R . L. Dobrushin, ”Optimum information transmission through a channel with unknown parameters,” Radio €ng. Electron. Phys., vol. 4, no. 12, pp. 1-8, 1959.

[I61 M. Yu. Cadzhiev, “Determination of the optimum variation mode of the useful signal and noise carrier frequencies in detection problems based on the theory of games,” Auto- mat. Remote Contr., vol. 22, no. 1, pp. 31-39,1961.

[I71 C. E. P. Box, “Non-normality and tests on variances,” Bio- metrika, vol. 4 0 , pp. 318-335, 1953.

[18] P. J. Huber, “Episodes from the early history of robustness,” in Abstracts of Papers: 1982 /FEE Int. Symp. on Information Theory (Les Arcs, France, June 1982), p. 72.

[19] -, “Robust statistics: A review,” Ann. Math. Stat., vol.

[20] -, Robust Statistical Procedures. Philadelphia, PA: SlAM Press, 1977.

[21] F. R. Hampel, ”Robust estimation: A condensed partial survey,” Z. Wahr. verw. Ceb., vol. 27, pp. 87-104, 1973.

[22] P. J. Bickel, “Another look at robustness: A review of reviews and some new developments,” Scand. /. Statist., vol. 3, pp.

[23] R. V. Hogg, “Adaptive robust procedures: A partial review and some suggestions for future applications and theory,” J. Amer. Stat. Assoc., vol. 69, pp. 909-927, 1974.

[24] -, “Statistical robustness: One view of its use in applica-

1979. tions today,” Amer. Statistician, vol. 33, pp. 108-115, Aug.

1251 A. A. Ershov, ”Stable methods of estimating parameters (survey),” Automat. Remote Contr., vol. 39, no. 7, pt. 2, pp.

Math. Stat., VOI. 36, pp. 1753-1758, 1965.

pp. 193-199, 1955.

43, pp. 1041-1067 1972.

145-1 68, 1976.

1152-1181, July 1978.

[47 1

D. F. Andrews, P. J . Bickel, F. R. Hampel, P. J. Huber, W. H. Rogers, and J. W. Tukey, Robust Estimates of Location: Survey and Advances. Princeton, NJ: Princeton Univ. Press, 1972. R. L. Launer and C. N. Wilkinson, Eds., Robustness in Statis- tics. New York: Academic Press, 1979. P. J. Huber, Robust Statistics. New York: Wiley, 1981. S. S. Wolff and J. L. Castwirth, “Robust two-input correlators,” J. Acoust. SOC. Amer., vol. 41, pp. 121 2-1 219, May 1967. R. D. Martin and S. C. Schwartz, “Robust detection of a known signal in nearly Gaussian noise,” / €€E Trans. Inform. Theory, vol. IT-17, pp. 50-56, Jan. 1971. S. A. Kassam and H. V. Poor, “Robust signal processing for communication systems,” I€€€ Commun. Mag., vol. 21, pp. 20-28, Jan. 1983. V. D. VandeLinde, “Robust techniques in communication,” in Robustness in Statistics, R. L. Launer and G. N. Wilkinson, Eds. New York: Academic Press, 1979, pp. 177-200.

.V. M. Krasnenker, “Stable (robust) detection methods for signals against a noise background (survey),” Automat. Re- mote Contr., vol. 41, no. 5, pt. 1, pp. 640-659, May 1980. H. V. Poor, “Robustness in signal detection,” in Communica- tions and Networks: A Survey of Recent Advances, I. F. Blake and H. V. Poor, Eds. New York: Springer-Verlag, 1985 (to appear). 6. Kleiner, R. D. Martin, and D. J. Thomson, “Robust estima- tion of power spectra,” J. Roy. Stat. Soc., Ser. 6, vol. 41, no. 3,

R. D. Martin and D. J. Thomson, “Robust-resistant spectrum estimation,” Proc. I€€€, vol. 70, pp. 1097-1115, Sept. 1982. R. D. Martin, “Robust methods for time series,” in Applied Time Series I / , D. F. Findley, Ed. New York: Academic Press, 1981. B. T. Poljak and Ya. 2. Tsypkin, ”Robust identification,” Automatica. vol. 16, pp. 53-63, 1980. J. 8. Thomas, An lntroduction to Statistical Communication Theory. New York: Wiley, 1968. K. S. Vastola and H. V. Poor, “An analysis of the effects of spectral uncertainty on Wiener filtering,” Automatica, vol.

T. L. Marzetta and S. W. Lang, ”Power spectral density bounds,” /FEE Trans. Inform. Theory, vol. IT-30, pp. 117-122, Jan. 1984. 0. M. Kurkin and I . C. Sidorov, “Signal filtering in the presence of noise with incompletely known probability characteristics,” Radio Eng. Electron. Phys., vol. IO, pp. 54-62, 1980. N. E. Nahi and I. M. Weiss, “Bounding filter: A simple solution to the lack of exact a priori statistics,” Inform.

S. A. Kassam and T. L. Lim, “Robust Wiener filters,” J. Franklin Inst., vol. 304, pp. 171-185, 1977. H. V. Poor, “On robust Wiener filtering,” /FEE Trans. Auto- mat. Contr., vol. AC-25, pp. 531-536, June 1980. S. M. Ali and S. D. Silvey, “A general class of coefficients of divergence of one distribution from another,” J. Roy. Stat. Soc., Ser. B, vol. 28, pp. 131-142, 1966. I. Csiszar, ”Information-type measures of difference of prob- ability distributions and indirect observations,“ Studia Sciem tariurn Mathematicerium Hungarica, vol. 2, pp. 229-318, 1 %7. H. V. Poor, “Minimax linear smoothing for capacities,” Ann. Prob., VOI. IO, pp. 504-507, 1982. H. V. Poor and D. P. Looze, “Minimax state estimation for linear stochastic systems with noise uncertainty,” /FEE Trans. Automat. Contr., vol. AC-26, pp. 932-906, 1981. L. J, Cimini and S. A. Kassam, “Robust and quantized Wiener filters for ppoint spectral classes,” in Proc. 7 9 8 0 Conf. on Inform. Sciences and Systems (Princeton Univ., Princeton, NJ, Mar. l W ) , pp. 314-318. K. S. Vastola and H. V. Poor, ”On generalized band models in robust detection and filtering,” in Proc. 7 9 8 0 Conf. on Inform Sciences and Systems (Princeton Univ., Princeton, NJ, Mar. I W ) , pp. 1-5.

pp. 313-351, 1979.

28, pp. 289-293, 1983.

Contr., VOI. 39, pp. 212-224, NOV. 1978.

K A S S A M A N D P O O R : ROBUST TECHNIQUES FOR SIGNAL PROCESSING 477

Page 46: Robust Techniques for Signal Processing - A Survey

L. J. Cimini and S. A. Kassam, “Optimum piecewise constant Wiener filters,” j . Opt. SOC. Amer., vol. 77, pp. 1162-1171, Oct. 1981. L. Brieman, “A note on minimax filtering,” Ann. Prob., vol. 1, pp. 175-1 79, 1973. S. A. Kassam, “The bounded ppoint classes in robust hy- pothesis testing and filtering,” in Proc. 20th Annual Allerton Conf. on Comm., Control and Computing (Monticello, IL,

K . S. Vastola and H. V. Poor, “Robust Wiener-Kolmogorov theory,” / €E€ Trans. Inform. Theory, vol. IT-30, pp. 316-327, Mar. 1984. E. Wong, Stochastic Processes in Information and Dynamical Systems. New York: Wiley, 1971. J. Franke, “Minimax robust prediction of discrete time series,” 2. Wahr. Verw. Geb., vol. 68, pp. 337-364, 1985. K. S. Vastola, “On robust linear causal estimation of continu- ous-time signals,” in Abstracts of Papers: 7932 I€€€ Int. Symp. on Inform. Theory (Les Arcs, France, June 21-25, 1982), p. 124. J. Franke and H. V. Poor ”Minimax robust filtering and finite-length robust predictors,” in Robust and Nonlinear Time Series Analysis, J. Franke, W. Hardle, and R. D. Martin, Eds. Heidelberg, Germany: Springer-Verlag, 1984. K . Yao, ”An alternative approach to the linear causal least- square filtering theory,” / €€€ Trans. Inform. Theory, vol.

J. Snyders, “Error formulae for optimal linear filtering, predic- tion and interpolation of stationary time series,” Ann. Math.

-, “Error expressions for optimal linear filtering of sta- tionary processes,” /€E€ Trans. Inform. Theory, vol. IT-18, pp.

K. Hoffman, Banach Spaces of Analytic Functions. En- glewood Cliffs, NJ: Prentice-Hall, 1962. Y. Hosoya, ”Robust linear extrapolations of second-order stationary-processes,” Ann. Prob., vol. 6, pp. 574-584, 1978. C. K. Colubev and M. S. Pinsker, “Minimax extrapolation of functions,” Prob. Inform. Transmission, vol. 20, pp. 99-11?. Oct. 1984. Yu. B. Korobochkin, “Minimax linear estimation of a sta- tionary random sequence in the presence of a perturbation with limited variance,” Radio €ng. Electron. Phys., vol. 28, no. 11, pp. 74-78, 1983. H. V. Poor, “The rate-distortion function on classes of sources determined by spectral capacities,” / €€E Trans. Inform. The

M. S. Pinsker, lnforrnation and Information Stability of Ran- dom Variables and Processes. San Francisco, CA: Holden- Day, 1%4. A. Papoulis, “Maximum entropy and spectrum estimation: A review,“ I€€€ Trans. Acoust., Speech, and Signal Process., vol. ASSP-29, pp. 1176-1186, Dec. 1981. C. Moustakides and S. A. Kassam, “Minimax robust equaliza- tion for random signals,” / € € E Trans. Commun., vol. COM-33, 1985 (to be published). -, “Robust Wiener filters for random signals in corre- lated noise,” / €€E Trans. Inform. Theory, vol. IT-29, pp.

-, “Robust Wiener filters for correlated signals and noise,” in Proc. 7 9 8 0 Conf. on Inform. Sciences Syst. (Prince- ton Univ., Princeton, NJ, Mar. 26-28, lW), pp. 308-313. P. J. Huber and V. Strassen, “Minimax tests and the Ney- man-Pearson lemma for capacities,” Ann. Statist., vol. 1, pp.

H. Rieder, “Least favorable pairs for special capacities,” Ann Statist., vol. 5, pp. 909-921, 1977. T. Bednarski, “On solutions of minimax test problems for special capacities,” Z. Wahr. verw. Geb., vol. 58, pp. 397-405, 1981. K. S. Vastola and H. V. Poor, “On the ppoint uncertainty class,” /€E€ Trans. Inform. Theory, vol. IT-30, pp. 374-376, Mar. 1984.

OCt. 6-8, 1982), pp, 526-534.

IT-17, pp. 232-240, 1971.

Statist., VOI. 43, pp. 1935-1943, 1972.

574-582, 1972.

OV, VOI. IT-28, pp. 19-26, 1982.

614-619, July 1983.

251-263, 1973.

M. Taniguchi, “Robust regression and interpolation for time series,” j . Time Ser. Analysis, vol. 2, pp. 53-62, 1981. 5. A. Kassam, “Robust hypothesis testing and robust time series interpolation and regression,” 1. Time Ser. Analysis,

J. A. D’Appolito and C. E. Hutchinson, “A minimax approach to the design of low sensitivity state estimators,” Auto- matica, vol. 8, pp. 599-608, 1972. V. P. Perov, “Optimum linear filtering of random processes when the a priori information is limited,” Radio €ng. Elec- tron. Phys., vol. 21, no. 10, pp. 68-75, 1976. C. T. Chen and S. A. Kassam, “Robust Wiener filtering for multiple inputs with channel distortion,” /€E€ Trans. Inform. Theory, VOI. IT-30, pp. 674-677, July 1%. J. C. Darragh and D. P. Looze, “Noncausal minimax linear state estimation for systems with uncertain second-order statistics,” I€€€ Trans. Automat. Contr., vol. AC-29, pp. 555-557, June 1984. H. Tsaknakis and P. Papantoni-Kazakos, ”Robust linear filter- ing for multivariable stationary time series,” in Proc. 7 9 8 4 Conf. on Inform. Sciences Syst. (Princeton Univ., Princeton, NJ, March 14-16, 1%), pp. 6-10. V. D. VandeLinde, “Robust properties of solutions to linear- quadratic estimation and control problems,” /€€€ Trans. A u tomat. Contr., vol. AC-22, pp. 138-139, Feb. 1977. J. M. Morris, “The Kalman filter: A robust estimator for some classes of linear quadratic problems,” I€€€ Trans. Inform. Theory, vol. IT-22, pp. 526-534, 1976. H. V. Poor and D. P. Looze, “Minimax state estimation for linear stochastic systems with noise uncertainty,” IEEE Trans. Automat. Contr., vol. AC-26, pp. 902-906, 1981. S. Verdu and H. V. Poor, “Minimax linear observers and regulators for stochastic systems with uncertain second-order statistics,” /E€€ Trans. Automat. Contr., vol. AC-29, pp. 499-511, June 1%. C. Martin and M. Mintz, “Robust filtering and prediction for linear systems with uncertain dynamics: A game-theoretic approach,” IEEE Trans. Automat. Contr., voi. AC-28, pp, 888-8%, Sept. 1983. P. P. Gusak, “Upper bound for rms filtering performance criterion in quasilinear models with incomplete information,”

1981. Automat. Remote Contr., vol. 42, no. 4, pp. 466-471, Apr.

T. Kailath, “RKHS approach to detection and estimation problems-Part I : Deterministic signals in Gaussian noise,” /E€€ Trans. Inform. Theory, vol. IT-17, pp. 530-549, Sept. 1971. V. P. Kuznetsov, ”Stable detection when the signal and spectrum of normal noise are inaccurately known,” Tele comm. Radio Eng., vol. 30/31, pp. 58-64, Mar. 1976. S. A. Kassam, T. L. Lim, and L. J. Cimini, “Two-dimensional filters for signal processing under modeling uncertainties,” /E€€ Trans. Geosci. Remote Sensing, vol. GE-18, pp. 331 -336, Oct. 1980. W. C. Knight, R. C. Pridham, and S. M. Kay, “Digital signal processing for sonar,” Proc. /€€E, vol. 69, pp. 1451-1506, Nov. 1981. V. V . Rodionov, “A game theory approach to detection of radar signals in the presence of unknown interference,” Radio Eng. Nectron. Phys., vol. 27, no. 9, pp. 74-80, 1982. C. R. Cahn, “Performance of digital matched filter correlator with unknown interference,” I€€€ Trans. Commun., vol COM-19, pp. 1163-1172, Dec. 1971. C. L. Turin, “Minimax strategies for matched-filter detection,” I€€€ Trans. Commun., vol. COM-23, pp. 1370-1371, Nov. 1975. C.-T. Chen, “Robust and quantized linear filtering for multi- ple input systems,” Ph.D. dissertation, Dept. of Systems Eng., Univ. of Pennsylvania, Philadelphia, 1983. C.-T. Chen and 5. A. Kassam,”Robust multiple-input matched filtering: Frequency and time-domain results,” I€€€ Trans. Inform. Theory (to appear).

VOI. 3, pp. 185-194, 1982.

[99] V. P. Kuznetsov,.”Minimax linear Neyman-Pearson detectors . . .

478 P R O C E E D I N G S O F T H E IEEE. VOL. 7 3 , NO. 3, M A R C H 1985

Page 47: Robust Techniques for Signal Processing - A Survey

for an imprecisely known signal and noise,” Radio €ng. Electron. Phys., vol. 19, no. 12, pp. 73-81, 1974. S. Verdu and H. V . Poor, “Minimax robust discrete-time matched filters,” I€€€ Trans. Commun., vol. COM-31, pp. 208-215, Feb. 1983. V. P. Kuznetsov, “Synthesis of linear detectors when the signal i s inexactly given and the properties of the normal noise are incompletely known,” Radio Eng. Electron. Phys., vol. 19, no. 12, pp. 65-73, 1974. M . V. Burnashev, “On the Minimax detection of an inaccu- rately known signal in a white Gaussian noise background,” Prob. Appl., vol. 24, pp. 107-119, 1979. R. Sh. Aleyner, “Synthesis of stable linear detectors for an inaccurately known signal,” Radio Eng. Electron. Phys., vol. 22, no. 1, pp. 142-145, 1977. -, “On the stable detection of a quasi-deterministic signal the presence of multiplicative interference,” Radio Eng. Nectron. Phys., vol. 24, no. I O , pp. 60-66, 1979. S. Verdu and H. V. Poor, “Signal selection for robust matched filtering,” I€€€ Trans. Commun., vol. COM-31, pp. 667-670, May 1983. K. S. Vastola, J. S. Farnbach, and S. C. Schwartz, “Maximin sonar system design for detection,” in Proc. 1983 I€€€ Int. Conf. on Acoustics, Speech andsignal Processing, Apr. 1983. R. A. Monzingo and T. W. Miller, Introduction to Adaptive Arrays. New York: Wiley, 1980. K. M. Ahmed and R. J. Evans, ”Robust signal and array processing,” Proc. Inst. Elec. Eng., vol. 129, pt. F, no. 4, pp.

C. H. Knapp and G. C. Carter, ”The generalized correlation method for estimation of time delay,” I€€€ Trans. Acoustics, Speech, Signal Process., vol. ASSP-24, pp. 320-327, Aug. 1976. E. K. AI-Hussaini and S. A. Kassam, “Robust Eckart filters for time delay estimation,” /€€E Trans. Acoustic, Speech, Signal Process., vol. ASSP-32, pp. 1052-1063, Oct. 1984. H. V. Poor, ”Robust decision design using a distance crite- rion,” I€€€ Trans. Inform. Theory, vol. IT-26, pp. 575-587, Sept. 1980. -, “Robust matched filters,” I€€€ Trans. Inform. Theory (to be published). C. W. Helstrom, Statistical Theory of Signal Detection, 2nd ed. Oxford, England: Pergamon, 1968. D. Slepian, “On bandwidth,” Proc. IEEE, vol. 64, pp. 292-300, Mar . 1976. S. Verdu and H. V. Poor, “On minimax robustness: A general approach and applications,” I€€€ Trans. Inform. Theory, vol. IT-30, pp. 328-340, Mar. 1984. N. M. Khalfina and L. A. Khalfin, ”On a robust version of the likelihood ratio test,” Theory Prob. Appl., vol. 20, pp. 199-202, 1975. Also in Proc. I€€€, vol. 64, pp. 292-3133, Mar. 1976. G. V. Trunk, ”Non-Rayleigh sea clutter: Properties and detec- tion of targets,” Naval Res. Lab Rep. 7986, Washington, DC, June 1976. J. H. Miller and J. 6. Thomas, “Robust detectors for signals in non-Gaussian noise,” / € € E Trans. Commun., vol. COM-25,

S. A. Kassam, “Robust hypothesis testing for bounded classes of probability densities,” / € E € Trans. Inform. Theory, vol. IT-27, pp. 242-247, Mar. 1981, V. P. Kuznetsov, ”Stable methods of discriminating between two signals,” Radio Eng. Nectron. Phys., vol. 19, no. IO, pp.

-, “Stable detection of a signal that is not precisely known,” Prob. Inform. Transmission, vol. 12, no. 2, pp.

-, ”Stable rules for discrimination of hypotheses,” Prob. Inform. Trans., vol. 18, no. 1, pp. 41-51, 1982. D. Kazakos, “Signal detection under mismatch,” / € € E Trans. Inform. Theory, vol. IT-28, pp. 681-684, July 1982. E. A. Ceraniotis, ”Performance bounds for robust decision problems with uncertain statistics,” in Proc. 2 n d I€€€ Conf.

297-302, Aug. 1982.

pp, 686-691, July 1977.

42-49, 1974.

11 9-1 29, 1976.

I381

1391

I 401

1411

on Decision and Control (San Antonio, TX, Dec. 14-16,

E. A. Ceraniotis and H. V. Poor, “Minimax discrimination for observed Poisson processes with uncertain rate functions,” I€€€ Trans. Inform. Theory, vol. IT-31, 1985 (to appear). P. M. Schultheiss and J. J. Wolcin, “Robust sequential prob- ability ratio detectors,” in EASCON ’75 Conv. Rec., pp.

J. Capon, ”On the asymptotic efficiency of locally optimum detectors,” IRE Trans. Inform. Theory, vol. IT-7, pp. 67-71, Apr. 1961. J. H. Miller and J. 6. Thomas, “Detectors for discrete-time signals in non-Gaussian noise,” / € E € Trans. Inform. Theory, vol. IT-18, pp, 241-250, Mar. 1972. S. A. Kassam and J. 6. Thomas, ”Asymptotically robust detec- tion of a known signal in contaminated non-Gaussian noise,” I€€€ Trans. Inform. Theory, vol. IT-22, pp. 22-26, Jan. 1976. H. V. Poor and J. E. Thomas, “Asymptotically robust quanti- zation for detection,” /€€€ Trans. Inform Theory, vol. IT-24, pp. 222-229, Mar. 1978. A. H. El-Say and V. D. VandeLinde, “Robust detection of known signals,” I€€€ Trans. Inform. Theory, vol. IT-23, pp. 722-727, Nov. 1977. -, “Robust sequential detection of signals in noise,” IEEE Trans. Inform. Theory, vol. IT-25, pp. 346-353, May 1979. C. V. Trunk, ”Small- and large-sample behavior of two detectors against envelope-detected sea clutter,” I€€€ Trans. Inform. Theory, vol. IT-16, pp. 95-99, Jan. 1970. C. V. Trunk and S. F. George, “Detection of targets in non-Gaussian sea clutter,” I€€€ Trans. Aerosp. Nectron. Syst., vol. AES-6, pp. 620-628, Sept. 1970. J. M. Morris, ”On single-sample robust detection of known signals in additive unknown-mean amplitude-bounded ran- dom interference,” /€€€ Trans. Inform. Theory, vol. IT-26, pp. 199-209, Mar. 1980. -, “On single-sample robust detection of known signals in additive unknown-mean amplitude-bounded random in- terference-Part II: The randomized decision rule solution,” / € € E Trans. Inform. Theory, vol. IT-27, pp. 132-1 36, Jan. 1981. -, “Performance of a suboptimal multisample decision rule against known signals with additive unknown-mean amplitude-bounded random interference,” in Proc. 27st I€€€ Conf. on Decision and Control (Orlando, FL, Dec. 8-10,

S. A. Kassam, ”Locally robust array detectors for random signals,” I€€€ Trans. Inform. Theory, vol. IT-24, pp. 309-316, May 1978. M. Kanefsky. “Detection of weak signals with polarity coin- cidence arrays,” /€€€ Trans. Inform. Theory, vol. IT-12, pp. 260-268, Apr. 1%6. R. D. Martin and C. P. McGath, “Robust detection of sto- chastic signals,” / E € € Trans. Inform. Theory, vol. IT-20, pp.

H. V. Poor, M. Mami, and J. 8. Thomas, “On robust detection of discrete-time stochastic signals,” /. Franklin Inst., vol. 309, pp. 29-53, Jan. 1980. H. V. Poor and J. 6. Thomas, ”Locally-optimum detection of discrete-time stochastic signals in non-Gaussian noise,” 1. Acoust. SOC. Amer., vol. 63, pp. 75-80, Jan. 1978. J . W. Modestino and A. Y. Ningo, “Detection of weak signals in narrowband non-Gaussian noise,” I€€€ Trans. Inform The- ory, vol. IT-25, pp. 592-600, Sept. 1979. S. A. Kassam, ”Detection of narrowband signals: Asymptotic robustness and quantization,” in Proc. 1984 Conf. on Inform. Sciences and Systems (Princeton Univ., Princeton, NJ, Mar.

S. A. Kassam, ”Nonparametric detection of narrowband sig- nals,’’ in Proc. 27th Midwest Symp. on Circuits and Systems (Morgentown, WV, June 1984), pp. 518-521. A. H. El-Say, “Detection of signals with unknown phase,” in Proc. 17th Ann. Allerton Conf. on Comm., Control and Computing (Monticello, IL, Oct. 1979). pp. 152-161.

1983). pp. 298-303.

36-A-36-H, 1975.

1982), pp, 437-439.

537-541, July 1974.

1984), pp. 297-301.

KASSAM A N D POOR. ROBUST TECHNIQUES F O R SIGNAL PROCESSING 479

Page 48: Robust Techniques for Signal Processing - A Survey

[I 71 1

P. A. Kelly, “Robust estimation and detection of signals with arbitrary parameters,” in Proc. 27st Annual Allerton Conf. on Commun., Control and Computing (Monticello, IL, Oct. 5-7,

J. G. Shin and 5. A. Kassam, “Robust detector for narrow- band signals in non-Gaussian noise,” J. Acoust. SOC. Amer.,

5. 5. Wolff and F. J . M. Sullivan, “Robust many-input signal detectors,” in Signal Processing: Proc. of NATO Advanced Study Inst. on Signal Processing, J. W. R. Criffiths et a/., Eds. London, England: Academic Press, 1973. V. P. Kuznetsov, “Synthesis of stable, small signal detectors,” Radio Eng. Electron. Phys., vol. 21, no. 2, pp. 26-34, 1976. H. V. Poor, “Signal detection in the presence of weakly dependent noise-Part II: Robust detection,” I€€€ Trans. Inform. Theory, vol. IT-28, pp. 744-752, 1982. 5. L. Portnoy, “Robust estimation in dependent situations,” Ann. Statist., vol. 5, pp. 22-43, 1977. -, “Further remarks on robust estimation in dependent situations,” Ann. Statist., vol. 7, pp. 244-251, 1979. G. V. Moustakides and J. B. Thomas, “Min-max detection of weak signals in +-mixing noise,” / € E € Trans. Inform. Theory, vol. IT-30, pp. 529-537, May 1984. S. A. Kassam and J. B. Thomas, ”Dead-zone limiter: An application of conditional tests in nonparametric detection,” J. Acoust. SOC. Amer., vol. 60, pp. 857-862, 1976. R. D. Martin, “Robust estimation of signal parameters with dependent data,” in Proc. 27st /FEE Conf. on Decision and Control (Orlando, FL, Dec. 8-10, 1982), pp. 433-436. 5. A. Kassam, C. Moustakides, and J . C. Shin, ”Robust detection of known signals in asymmetric noise,” /FEE Trans. Inform. Theory, vol. IT-28, pp. 84-91, Jan. 1982. J . R. Collins, “Robust estimation of a location parameter in the presence of asymmetry,” Ann. Statist., vol. 4, pp. 68-85, 1976. C. V. Moustakides, Robust Detection of Signals: A Large Deviations Approach, IRISA (CNRS) Publ. no. 243, Univ. de Rennes, France, Nov. 1984. R. D. Martin and 5. C. Schwartz, “On mixture quasi-mixture, and nearly normal random processes,” Ann. Math. Stat., vol. 43, pp. 948-967, 1972. P. A. Kelly and W. L. Root, “Stable linear detectors with optimality constraints,” in Proc. 22nd Ann. Allerton Conf. on Comm., Control, and Computing (Monticello, IL, Oct.

-, “Stability in detection of signals in noise,” in Proc. 23rd /FEE Conf. on Decision and Control (Las Vegas, NV, Dec. 12-14,1984), pp. 1436-1443. 5. A. Kassam, “Optimum data quantization for signal detec- tion,” in Communications and Networks: A Survey of Recent Advances, I. F. Blake and H. V. Poor, Eds. New York: Springer-Verlag, 1985. L. Kurz, “Nonparametric detectors based on partition tests,” in Nonparametric Methods in Communications, P. Papantoni-Kazakos and D. Kazakos, Eds. New York: Marcel-Dekker, 1977. R. F. Dwyer, “Robust sequential detection of weak signals in undefined noise using acoustical arrays,” J. Acoust. SOC. Amer., vol. 67, pp. 833-841, 1980. Y. C. Ching and L. Kurz, “Nonparametric detectors based on m-interval partitioning,” /E€€ Trans. Inform. Theory, vol. IT-18, pp. 251-257, Mar. 1972. P. Kersten and L. Kurz, “Robustized vector Robbins-Monro algorithm with applications to m-interval detection,” Inform.

-, “Bivariate minterval classifiers with application to edge detection,” Inform. Contr., vol. 34, pp. 152-168, 1977. R. F. Dwyer, ”Robust sequential detection of narrowband acoustic signals in noise,” in Proc. /FEE Int. Conf. on Acous- tics, Speech and Signal Processing, pp. 140-143, Apr. 1979. -, “A technique for improving detection and estimation of signals contaminated by under ice noise,” J. Acoust. SOC. Amer., vol. 74, pp. 124-130, July 1983. J. H. Miller and J. B. Thomas, “The detection of signals in impulsive noise modeled as a mixture process,” / €€E Trans. Commun., vol. COM-24, pp. 559-563, May 1976.

1983), pp. 602-609.

VOI. 74, pp. 527-533, Aug. 1983.

3-5, 1984), pp. 318-327.

Sci., VOI. 11, pp. 121 -1 40, 1076.

J . W. Modestino, ”Adaptive detection of signals in impulsive noise environments,” /FEE Trans. Commun., vol. COM-25,

L. B. Milstein, D. D. Schilling, and J. K . Wolf, ”Robust detection using extreme-value theory,” /FEE Trans. Inform. Theory, vol. 15, pp. 370-375, 1%9. 5. A. Kassam and J. B. Thomas, Eds., Nonparametric Detec- tion: Theory and Applications. Stroudsburg, PA: Dowden, Hutchinson and Ross, 1980. 5. A. Kassam, “A bibliography on nonparametric detection,” /E€€ Trans. Inform. Theory, vol. IT-26, pp. 595-602, Sept. 1980. R. D. Martin and C. J. Masreliez, “Robust estimation via stochastic approximation,” / €€E Trans. Inform. Theory, vol. IT-21, pp. 263-271, May 1975. A. B. Tsybakov, “Robust estimation of a function,” Prob. Inform. Transmission, vol. 18, no. 3, pp. 190-200, 1983. R. D. Martin, ”Robust estimation of signal amplitude,” /E€€ Trans. Inform. Theory, vol. IT-18, pp. 5%-64%, 1972. E. L. Price and V. D. VandeLinde, “Robust estimation using the Robbins-Munro stochastic approximation algorithm,” /E€€ Trans. Inform. Theory, vol. IT-25, pp. 698-704, 1979. A. C. Bovik, T. 5. Huang, and D. C. Munson, Jr., “A generali- zation of median filtering using linear combinations of order statistics,” I€€€ Trans. Acoust., Speech, and Signal Process., vol. ASSP-31, pp. 1342-1350, Dec. 1982. Y. H. Lee and S. A. Kassam, ”Nonlinear edge preserving filtering techniques for image enhancement,” in Proc. 27th Midwest Symp. Circuits Syst. (Morgentown, WV, June I%), pp. 554-557. -, “Applications of nonlinear adaptive filters for image enhancement,” in Proc. 7th Int. Conf. Pattern Recogn. (Montreal, Que., Canada, Aug. l W ) , pp. 930-932. K. Abend et a/., “Advanced digital receiver techniques,” Tech. Rep. RADC-TR-68-169, Rome Air Development Cen., Rome, NY, Sept. 1978. B. D. 0. Anderson and J. B. Moore, Optimal Filtering. En- glewood Cliffs, NJ: Prentice-Hall, 1979. C. J . Masreliez and R. D. Martin, “Robust Bayesian estimation for the linear model and robustifying the Kalman filter,” /FEE Trans. Automat. Contr., vol. AC-22, pp. 361-371, 1977. V. D. VandeLinde, R. Doraiswami, and H. 0. Yurtseven, “Robust filtering for linear systems,” in Proc. 77th /FEE Conf. on Decision and Control (New Orleans, LA, Dec. 13-15,

A. A. Ershov and R. Sh. Lipster, “A robust Kalman filter in discrete time,” Automat. Remote Contr., vol. 39, no. 3, 1978. A. A. Ershov, “Robust filtering algorithms,” Automat. Remote Contr., vol. 39, no. 7, pt. 1, pp. 992-996, July 1978. I. K. Levin, “Accuracy analysis of a robust filter of a certain type by the method of convex hull,” Automat. Remote Contr., no. 5, pp. 660-669,1980. C. Tsai and L. Kurz, “Adaptive robustizing approach to

1983. Kalman filtering,” Automatica, vol. 18, no. 3, pp. 279-288,

C. G. Boncelet, Jr., and B. W. Dickinson, “An approach to robust Kalman filtering,” in Proc. 2Znd / € € E Conf. Decision Control (San Antonio, TX, Dec. 14-16,1983), pp: 304-305. J. M. Morris, “Quantization and source encoding with a fidelity criterion: A survey,” Naval Res. Lab. Rep. 8082, Washington, DC, Mar. 25,1977. A. Gersho, “Principles of quantization,” /FEE Trans. Circuits

R. M. Gray, Ed. I€€€ Trans. Inform. Theory (Special Issue on Quantization), vol. IT-28, Mar. 1982. B. Aazhang and H. V. Poor, ”On optimum and nearly opti- mum data quantization for signal detection,” / €€E Trans. Commun., vol. COM-32, pp. 745-751, July 1984. H. V. Poor, “A companding approximation for the statistical divergence of quantized data,” in Proc. 2 n d I€€€ Conf. Decision Control (San Antonio, TX, Dec. 14-16, 1983). pp.

-, “The effects of data quantization on filtering of stationary Gaussian processes,” in Proc. 23rd IF€€ Conf. Decision Control (Las Vegas, NV, Dec. 12-14, 1 W), pp.

pp. 1022-1027, Sept. 1977.

1972), pp. 652-656.

Syst., VOI. CAS-25, pp. 427-436, July 1978.

697-702.

1430-1 435.

PROCEEDINGS O F THE IEEE. VOL. 73, NO. 3. MARCH 1985

Page 49: Robust Techniques for Signal Processing - A Survey

[I981 J , Max, ”Quantizing for minimum distortion,” IRE Trans. Inform. Theory, vol. IT-6, pp. 7-12, Mar. 1960.

[I991 D. J. Goodman and A. Gersho, “Theory of ara adaptive quantizer,” /FEE Trans. Commun., vol. COM-22, pp.

[200] A. Gersho and D. J. Goodman, ”A training mode adaptive quantizer,” /FEE Trans. Inform. Theory, vol. IT-20, pp.

[201] J. M. Morris and V. D. VandeLinde, ”Robust quantization of discrete-time signals with independent samples,” /FEE Trans. Commun. Techno/., vol. COM-22, pp. 1897-1901, Dec. 1974.

[202) W. C. Bath and V. D. VandeLinde, “Robust memoryless quantization for minimum signal distortion,” /€€E Trans. Inform. Theory, vol. IT-28, pp. 296-306, Mar. 1982.

[203] J , A . Bucklew and G. L. Wise, “Multidimensional asymptotic quantization theory with r-th power distortion measures,” /FEE Trans. tnform. Theory, vol. IT-28, pp. 239-247, Mar.

1037-1045, Aug. 1974.

746-749, NOV. 1974.

1982. [204] W. G. Bath and V. D. VandeLinde, “Robust quantizers de-

signed using the companding approximation,” in Proc. 78th /€E€ Conf. on Decision and Control (Ft. Lauderdale, FL, Dec.

[205] P. W. Cattermole, Principles of Pulse Code Modulation. London, England: Ilffe, 1969.

[206] D. Kazakos, ”On the design of robust quantizers,” in 7987 / €€E Telecomm. Conf., Conf. Rec. (New Orleans, LA, Dec.

[207] -, “New results on robust quantization,” /€E€ Trans.

[208] H. V. Poor, “Some results on robust data quantization,” in Proc. 27st / € € E Conf. Decision Control (Orlando, FL, Dec.

[209] -, “Robust quantization of <-contaminated data,” /E€€

1979), pp. 483-487.

1981), pp. F4.5.1-F4.5.4.

Commun., VOI. COM-31, pp. 965-974, Aug. 1983.

8-10, 1982), pp. 440-445.

Trans. Commun., vol. COM-33, pp. 218-222, Mar. 1985.

K A S S A M A N D POOR: ROBUST TECHNIQUES F O R SIGNAL PROCESSING 481

Page 50: Robust Techniques for Signal Processing - A Survey