Open Access
April, 1965 Large-Sample Estimation of an Unknown Discrete Waveform Which is Randomly Repeating in Gaussian Noise
Melvin Hinich
Ann. Math. Statist. 36(2): 489-508 (April, 1965). DOI: 10.1214/aoms/1177700159


Suppose we have an input $X(t)$ made up of an unknown waveform $\theta(t)$ of known length, which is repeated randomly, and is imbedded in Gaussian noise with a known covariance function. The rate of recurrence of the waveform is a known small constant. In addition, the signal-to-noise ratio of the input $X(t)$ is quite low. We wish to estimate the waveform $\theta(t)$ and its autocorrelation $\psi(\tau) = \int \theta(t + \tau)\theta(t) dt$. Restricting ourselves to discrete-time observations on $X(t)$, we shall derive an optimal estimator of the discrete version of $\psi(\tau)$. This estimator is a weighted average of the sample autocorrelation and the square of a linear estimator of the time average (the zero-frequency or DC value) of the waveform. For the estimation of $\theta$, the problem is more complicated. The optimality concept (asymptotic efficiency) used in this work is based upon large-sample theory and the Cramer-Rao Inequality (Chernoff [2] and Cramer [4]). The problem stated above was motivated by a problem of electronic surveillance of an enemy communication system based upon pulse position modulation, PPM. To illustrate this system suppose station $A$ is sending a message, which we wish to intercept and decode, to station $B$ by PPM over a certain FM bandwidth. $A$ continues to repeat a fixed pulse-type waveform $\theta(t) = \sum^n_{i = 1} \theta_iH\lbrack t - (i - 1)T\rbrack$ where $H(t) = 1\quad\text{if} \quad 0 \leqq t < T,\\ = 0\quad\text{otherwise}$. The vector $\theta' = (\theta_1, \cdots, \theta_n)$, the parameter $n$, and $T$--the pulse width--are known to both $A$ and $B$. Notice that $nT$ is the time duration (length) of $\theta(t)$. Since $\theta, n,$ and $T$ are known to the receiver $B$, one may ask what the coding scheme is for the information that $A$ is sending. The answer is that the length of time between successive recurrences of the waveform, is the variable which contains the information. In many applications the average length of time between successive occurrences of $\theta(t)$ is around $10^2$ times $nT$. While the intervals between repetitions are fundamental in the transmission of information, someone who does not yet know the waveform may regard the recurrences to be purely random in time. This modulation technique has the effect of spreading the power in the FM bandwidth over a wider swath of the frequency scale. This makes surveillance more difficult because it requires that we somehow determine the actual bandwidth being used. Moreover, the spreading of the power makes jamming of the channel difficult. To further complicate matters, $A$ transmits the pulses with low power so that $B$ picks up an input $X(t)$ with low signal-to-noise ratio. Since $B$ knows $\theta(t)$, he uses matched filtering to detect the times of occurrence of the $\theta's$. If the noise is assumed to be additive Gaussian noise with known covariance, then matched filtering is optimal in a decision theoretic sense (Wainstein and Zubakov [8]). But suppose that we are listening in on this channel without knowing $\theta$ and we wish to find out what $A$ is saying. First we must determine the frequency band of the channel which $A$ is using. Then we must detect the times of occurrence of the $\theta's$, although we do not know $\theta$. However, let us assume that we have already determined $n$ and $T$, although it will turn out that $n$ is not a vital parameter in the estimator developed in this paper. Jakowatz, Shuey, and White [6] present a special discrete-time (sampled data) system, called the Adaptive Filter, which estimates an unknown waveform which is repeating in additive noise. The system uses a complicated stochastic iterative procedure. The Filter obtains a crude estimate of the waveform from the initial input and uses the discrete cross-correlation between this estimate and the input to detect the times of occurrence of the waveform. When it decides that a waveform is present in the input, it refines the estimate by averaging it with the section of input where the waveform is thought to be present. Provided the autocorrelation of the waveform--$\psi(\tau)$--has $\psi(0)$, its maximum, a good deal larger than the relative maxima of $\psi$, and provided the noise is well behaved, then this iterative procedure results in an asymptotically stable estimate for the waveform. This stable estimate is then used as the matching element in matched filter detection of the waveform. Thus we could call the Adaptive Filter an adaptive matched filter. A partial analysis of this system is given by Hinich [5]. The estimate of the discrete autocorrelation of the waveform is helpful to the analysis of systems which are based upon discrete-time cross-correlation such as the Adaptive Filter, since the autocorrelation is a basic parameter in the distributions of the random variables (correlations) which arise in the operation of these systems. Suppose we can obtain an expression for $\theta$ in terms of $\psi, \theta = f(\psi)$. Once we have obtained an asymptotically efficient estimator of $\psi$, call it $\hat\psi$, then $\hat\theta = f(\hat\psi)$ would be an asymptotically efficient estimator of $\theta$. Unfortunately, there is a multiplicity of $\theta's$ which have $\psi$ as their autocorrelation. In Section 6 we will present a method for obtaining the correct $\theta$ (the one which appears in the input) from $\psi$ by using the observations on $X(t)$. Unfortunately this method does not seem to be sufficiently practical. To conclude, let us outline the rest of this paper. In Section 2 we give a formal statement and description of the problem posed above. In Section 3 we state the Cramer-Rao theorem and derive the information matrices relevant to the estimation of $\theta$ and $\psi$, as well as their inverses. In Section 4 we present the optimal estimator of $\psi$. We also show that while the normalized sample correlation is an unbiased estimator of $\psi$, it is not efficient. In Section 5 we discuss three examples. These are the general case white noise, the case of white noise when the discrete waveform has only two components $\theta_1, \theta_2$, and the case of Gaussian Markov noise $(EN (t + \tau)N(t) = \rho^{|\tau|}, 0 < \rho < 1)$. In Section 6 we discuss the problem of estimating $\theta$ after the autocorrelation $\psi$ has been estimated.


Download Citation

Melvin Hinich. "Large-Sample Estimation of an Unknown Discrete Waveform Which is Randomly Repeating in Gaussian Noise." Ann. Math. Statist. 36 (2) 489 - 508, April, 1965.


Published: April, 1965
First available in Project Euclid: 27 April 2007

zbMATH: 0132.39104
MathSciNet: MR174400
Digital Object Identifier: 10.1214/aoms/1177700159

Rights: Copyright © 1965 Institute of Mathematical Statistics

Vol.36 • No. 2 • April, 1965
Back to Top