The intent of detection theory is to provide rational (instead of arbitrary) techniques for determining which of several conceptions--models--of data generation and measurement is most "consistent" with a given set of data. In digital communication, the received signal must be processed to determine whether it represents a binary "0" or "1"; in radar or sonar, the presence or absence of a target must be determined from measurements of propagating fields; in seismic problems, the presence of oil deposits must be inferred from measurements of sound propagation in the earth. Using detection theory, we will derive signal processing algorithms which will give good answers to questions such as these when the information-bearing signals are corrupted by superfluous signals (noise).

The detection theory's foundation rests on statistical
hypothesis testing (Cramér, 1946, Chapter
35; Lehman, 1986; Poor, 1988, Chapter 2; van
Trees, 1968, pp 19-52). Given a probabilistic model (an event
space *not* to impose rigid
constraints on the algorithms that solve the hypothesis
testing problem. Over a variety of reasonable criteria, one
central solution to evaluating which model describes
observations--the likelihood ratio test--will persistently
emerge; this result will form the basis of
*all* detection algorithms.

Detection problems become more elaborate and complicated when models become vague. Models are characterized by probability distributions, and these distributions suffice in the likelihood ratio test. Vagueness does not refer to this stochastic framework; rather, it refers to uncertainties in the probability distribution itself. The distribution may depend on unknown parameters, like noise power level. The distribution most certainly depends on signal structure; suppose that is partially or completely unknown? The most difficult (and interesting) problems emerge when uncertainties arise in the probability distributions themselves. For example, suppose the only model information we have is through data; how would an optimal detector be derived then?

Along the way we will discover that a general geometric picture of detection emerges: Ease of a detection problem depends on how "far apart" the models are from each other. This geometric framework turns out to be elaborate, but underlies modern detection theory and forms links to information theory.