20
Multichannel Time of Delay Estimation

Time of delay (ToD) estimation is a very common problem in many applications that employ wavefield propagation such as radar, remote sensing, and geophysics (Section 18.5) where the information of interest is mapped onto ToDs. Features of the wavefields call for measurement redundancy from a multichannel system where there are images signals that are collected as s1(t), s2(t), …, sM(t), and this leads to massive volumes of data for estimating the ToD metric. The uniform linear array in Chapter 19 represents a special case of a multichannel system where the source(s) is narrowband and the only parameter(s) that can be estimated is the DoA(s). Here the narrowband constraint is removed while still preserving the multidimensionality of the signals in order to have the needed redundancy for estimating the ToD variation vs. measuring channels (e.g., sensors) with known geometrical positions.

The multichannel concept is illustrated in Figure 20.1 from seismic exploration (see Section 18.5 or [82] ): the wavefield generated at the surface propagates in the subsurface and is backscattered at depth discontinuities, and measured with sensors at the surface. The wavefield is collected by a regularly spaced 2D array of sensors to give the measured wavefields s(x, y, t) in space (x, y) and time t in the figure. Note that in Figure 20.1 the 3D volume s(x, y, t) of 1.3 km images 10 km is sliced along planes to illustrate by gray‐scale images the complex behavior of multiple backscattered wavefronts, and the corresponding ToD discontinuities of wavefronts due to subsurface faults. Multichannel ToD estimation refers to the estimation of the ensemble of ToDs pertaining to the same wavefront represented by a 2D surface τ(x, y) that models the th delayed wavefront characterized by a spatial continuity.1 The measured wavefield can be modeled as a superposition of echoes (for any position x, y)

images

where the amplitudes α(x, y) and ToDs τ(x, y) are wavefront and sensor dependent, and g(t) is the excitation waveform.

3D Graph illustrating the multichannel measurements for subsurface imaging, depicting wavefront and ground surface (arrows).

Figure 20.1 Multichannel measurements for subsurface imaging.

The ToD from multiple measurements is by chaining multiple distinct sensors. A radar system can be interpreted as a multichannel model when estimating the ToD variation vs. the scanning time when the target(s) move(s) from one time‐scan to another, or the radar antenna moves for a still target (e.g., satellite remote sensing of the earth). Regardless of the origin of the multiplicity of measurements, the most general model for multichannel ToD for the i th measurement (Figure 20.2):

images

is the superposition of L echoes each with waveform g(t), amplitude α,i, and ToD τ,i as sketched below in gray‐intensity plot with superimposed the trajectory of the ToD {τ,i} vs. the sensor’s position (black‐line). ToD estimation refers to the estimation of the delays images for a specific measurement, and the amplitudes images could be nuisance parameters. However, when considering a collection of measurements s1(t), s2(t), …, sM(t), multichannel ToD estimation refers to the estimation of the ToD profile for the th wavefront (i.e., a collection of ToD vs. sensor):

images
Multichannel ToD model displaying 3 horizontal and 1 vertical fluctuating waves .

Figure 20.2 Multichannel ToD model.

Parametric wavefront: Estimation of ToDs of the wavefront(s) as a collection of the corresponding delays. The wavefront can be parameterized by a set of wavefront‐dependent geometrical and kinematical shape parameters θ (e.g., sensors position, sources position, and propagation velocity) as illustrated in Section 18.5, so that the ToD dependency is

images

To exemplify, one can use a linear model for the ToDs images so that the wavefront can be estimated by estimating the two parameters images instead. The estimation of the shape parameters from a set of redundant (images) measurements has the benefit of tailoring ToD methods to estimate the sequence of ToDs vs. a measurement index as in Figure 20.2. Applications are broad and various such as wavefield based remote sensing, radar, and ToD tracking using Bayesian estimation methods (Section 13.1). To illustrate, when two flying airplanes cross, one needs to keep the association between the radar’s ToDs and the true target positions as part of the array processing (Chapter 19). Only the basics are illustrated in this chapter, leaving the full discussion to application‐related specialized publications.

Active vs. passive sensing systems: The waveforms g1(t), g2(t), …, gL(t) can be different to one another and dependent on the sources, which are not necessarily known. A radar/sonar system (Section 5.6) or seismic exploration (Section 18.5) are considered as active systems as the waveforms are generated by a controlled source and all waveforms are identical (images), possibly designed to maximize the effective bandwidth (Section 9.4) and thus the resolution for multiple echoes. In contrast to active systems, in passive systems the waveforms are different from one another, modeled as stochastic processes characterized in terms of PSD. An example of a passive system is the acoustic signal in an concert hall, which is the superposition of signals from many musical instruments. The scenario is complex, and Table 20.1 offers a (limited and essential) taxonomy of the methods and summarizes those discussed in this chapter.

Table 20.1 Taxonomy of ToD methods.

# sensors (M) # ToD (L) g(t) ToD estimation method
1 1 known MLE (Section 9.4)
2 1 not‐known Difference of ToD (DToD) (Section 20.3)
1 images images MLE for multiple echoes (Section 20.2)
images (wavefront) 1 known MLE/Shape parameters estimation (Section 20.5)
images (wavefront) 1 non‐known Shape parameters estimation (Section 20.6)
images (wavefront) images known Multi‐target ToD tracking [92]

20.1 Model Definition for ToD

The ToD problem for multiple echoes is continuous‐time estimation that must be modified to accommodate discrete‐time data. The time‐sampled signals are modeled as in Section 5.6:

images

where images is the column that orders the th delayed waveform sampled in {t1, t2, …, tN}, and noise images. Dependence of the ToDs is non‐linear and it is useful to exploit the quadratic dependency on the amplitudes αi in the ML method (Section 7.2.2):

images

The MLE follows from the optimization

images

where images. In the case of a white Gaussian noise process (images), the optimization for ML ToD estimation reduces to the maximization of the metric

images

where the projection matrix images imbeds into the matrix images all the products images that account for the mutual interference of the different waveforms. A special case is when the amplitudes are changing, but the delays (τ1, …, τL) are not:

images

This model reduces to the MLE in Section 7.2.3 and gives

images

The influence of noise and amplitude fluctuations is mitigated by the averaging of the outer products images.

Recall (Section 9.4) that for images and images, the MLE depends on the maximization of

that is the cross‐correlation between the samples of {s[n]} and the continuous‐time waveform g(t) arbitrarily shifted (by τ) and regularly sampled by the same sampling interval Δt of images. However, if the waveform is only available as sampled g[k] at sampling interval Δt, the cross‐correlation (20.1) can be evaluated over a coarse sample‐grid (Section 10.3), or it requires that the discrete‐time waveform is arbitrarily shifted by a non‐integer and arbitrary delay τ, and this can be obtained by interpolation and resampling (Appendix B).

20.2 High Resolution Method for ToD (L=1)

Let the model be

with known waveform g(t). The MLE of the ToDs

images

for images is a multidimensional search over L‐dimensions (see Section 9.4 for images) that might be affected by many local maxima, and the ToD estimate could be affected by large errors when waveforms are severely overlapped. High‐resolution methods adapted from the frequency estimation in Chapter 16 have been proposed in the literature and are based on some simple similarities reviewed below.

20.2.1 ToD in the Fourier Transformed Domain

The model after the Fourier transform of the data (20.2) is

Since the waveform is known, its Fourier transform G(ω) is known and the observation can be judiciously deterministically deconvolved (i.e., S(ω) is divided by G(ω) except when images) so that after deconvolution:

images

This model resembles the frequency estimation in additive Gaussian noise, where the ToD τ induces a linear phase‐variation vs. frequency, rather than a linear phase‐variation vs. time as in the frequency estimation model (Chapter 16).

Based on this conceptual similarity, the discrete‐time model can be re‐designed using the Discrete Fourier transform (Section 5.2.2) over N samples images to reduce the model (20.2) to (20.3):

where the th ToD τ makes the phase change linearly over the sample index k

images

behaving similarly to a frequency that depends on the sampling interval Δt. If considering multiple independent observations with randomly varying amplitudes images as in the model in Section 20.1, the ToD estimation from the estimate of images coincides with frequency estimation, and the high‐resolution methods in Chapter 16 can be used with excellent results.

In many practical applications, there is only one observation (images) and the model (20.4) can be represented in compact notation (N is even):

images

where

images

and

images

contains the DFT of the waveform images. The estimation of multiple ToDs follows from LS optimization of the cost function:

images

The LS method coincides with MLE when the noise samples W[k] are Gaussian and mutually uncorrelated. Since the DFT of Gaussian white process w(nΔt) leaves the samples W[k] to be uncorrelated due to the unitary transformation property of DFT, the minimization of J(α, ω) coincides with the MLE.

One numerical method to optimize J(α, ω) is by the weighted Fourier transform RELAXation (W‐RELAX) iterative method that resembles the iterative steps of EM (Section 11.6), where at every iteration the estimated waveforms are stripped from the observations for estimation of one single parameter [95]. In detail, let the set images be known as estimated at previous steps, after stripping these waveforms

images

one can use a local metric for the th ToD and amplitude

images

As usual, this is quadratic wrt α and it can be easily solved for the frequency ω (i.e., the th ToD)

images

and the amplitude follows once images has been estimated as

images

Iterations are carried out until some degree of convergence is reached. Improved and faster convergence is achieved by adopting IQML (Section 16.3.1) in place of the iterations of W‐RELAX.

The methods above can be considered extensions of statistical methods discussed so far, but some practical remarks are necessary:

  • Any frequency domain method needs to pay attention to the shape of G(ω), or the equivalent DFT samples G[k]. In definitions, the range of G should be limited to those values that are meaningfully larger than zero to avoid instabilities in convergence. However, restricting the range for ω reduces ToD accuracy considerably as it is dominated by higher frequencies in the computation of the effective bandwidth (Section 9.4). Similar reasoning holds when performing deterministic deconvolution if using other high‐resolution methods, as the deconvolution 1/G(ω) can be stabilized by conditioning that is adding a constant term to avoid singularities such as images, where η is a small fraction of max{|G(ω)|}.
  • Conditioning and accuracy are closely related. In some cases in iterative estimates it is better to avoid temporary ToD estimates getting too close one another: one can set a ToD threshold, say Δω, that depends on the waveforms. In this case, the estimates from the iterative methods above that are too close, say images and images with images and images, are replaced by images and images to improve convergence and accuracy.

20.2.2 CRB and Resolution

As usual, the CRB is a powerful analytic tool to establish the accuracy bound. In the case of ToD, the performance depends on the waveforms and their spectral content as detailed in Section 9.4 for images ToDs. When considering multiple delays, the CRB depends on the FIM entries, which for white Gaussian noise with power spectral density N0/2 can be evaluated in closed form as:

images

Recall that the choice of the waveform is a powerful degree of freedom in system design to control ToD accuracy when handling multiple and interfering echoes. Choice of waveform and range can satisfy images and images so that two (or more) ToDs are decoupled and the correlation‐based ML estimator can be applied iteratively on each echo. In all the other contexts when images and/or images, the CRB for multiple echoes is larger than the CRB for one isolated echo due to the interaction of the waveforms, and it can be made dependent on their mutual interaction as illustrated in the simple example in Section 9.4 with images.

With multiple echoes, resolution is the capability to distinguish all the echoes by their different ToDs. Let a measurement be of images echoes in τ1 and τ2 (images) with the same waveform; it is obvious that when the two ToDs are very close, if not coincident, the two waveforms coincide and are misinterpreted as one echo, possibly with a small waveform distortion that might be interpreted as noise. Most of the tools to evaluate the performance in this situation are by numerical experiments using common sense and experience. To exemplify in Figure 20.3, if the variance from the CRB is images and images the two echoes cannot be resolved as two separated echoes and estimates interfere severely with one another. On the contrary, when noise is small enough to make images, one can evaluate the resolution probability for the two ToD estimates images and images. A pragmatic way to do this is by choosing two disjoint time intervals images and images around τ1 and τ2; the resolution probability Pres follows from the frequency analysis that images and images are into the two intervals:

images

and the corresponding MSE in a Montecarlo analysis is evaluated for every simulation run conditioned to the resolution condition. Usually images and images have the same width images around each ToD: images and imagesimages.

Graphs of resolution probability displaying 2 overlapping bell-shape (top) and triangle (bottom) curves, with 3 rightward arrows with markers labeled estimated ToD.

Figure 20.3 Resolution probability.

20.3 Difference of ToD (DToD) Estimation

In ToD when images, the absolute delay is estimated wrt a reference time and this is granted by the knowledge of the waveform that is possible for an active system. images traces s1(t), s2(t) collected by two sensors allow one to estimate the difference of ToDs even when the waveform g(t) is not know, as one trace acts as reference waveform wrt the other. The degenerate array with two sensors is an exemplary case: the DToD for the same source signal depends on the angular position that can be estimated in this way (e.g., the human sense of hearing uses the ear to collect and transduce sound waves into electrical signals that allow the brain to perceive and spatially localize sounds). Let a pair of signals be

images

measured over a symmetric interval images where the waveform g(t) can be conveniently considered a realization of a stochastic process characterized by the autocorrelation images (but the sample waveform is not known!) and its power spectral density images. Additive impairments w1(t) and w2(t) are Gaussian and uncorrelated with the waveform. The obvious choice is to resume the cross‐correlation estimator (20.1) by using one signal as reference for the other, this is evaluated first in terms of performance. Later in the section, the elegant and rigorous MLE for DToD will be derived.

20.3.1 Correlation Method for DToD

In DToD, what matters is the relative ToD of one signal wrt the other as the two measurements are equivalent, and there is no preferred one. The cross‐correlation method is

where the two signals are shifted against one another until the similarities of the two signals are proved by the cross‐correlation metric. For discrete‐time sequences, the method can be adapted either to evaluate the correlation for the two discrete‐time signals, or to interpolate the signals while correlating (Appendix B). Usually the (computationally expensive) interpolation is carried out as second step after having localized the delay and a further ToD estimation refinement becomes necessary for better accuracy.

Performance Analysis

The cross‐correlation among two noisy signals (20.5) is expected to have a degradation wrt the conventional correlation method for ToD, at least for the superposition of noise and the unknown waveform. Noise is uncorrelated over the two sensors, images for every α, but it has the same statistical properties. The MSE performance is obtained by considering the cross‐correlation for all the terms that concur in s1(t) and s2(t):

images

and using a sensitivity analysis around the estimated value images of DToD that peaks the cross‐correlation, from the derivative:

images

Recall that images is the sample estimate of the stochastic cross‐correlation for a stationary process:

images

To simplify the notation, let the DToD be images; then the derivative of the correlation can be expanded into terms2

images

where the Taylor series has been applied only for the deterministic term on true DToD (images): images. For performance analysis of DToD, it is better to neglect any fluctuations of the second order derivative of the correlation of the waveform, which are replaced by its mean value images, and thus

images

It is now trivial to verify that the DToD is unbiased, while its variance (see properties in Appendix A, with images)

depends on the power spectral densities of stochastic waveform Sgg(f) and noise Sww(f). Once again, the effective bandwidth for stochastic processes for signal and noise

images

plays a key role in the DToD method. When considering bandlimited processes sampled at sampling interval Δt and bandwidth smaller than 1/2Δt (Nyquist condition), the effective bandwidth can be normalized to the sampling frequency to link the effective bandwidth for the sampled signal (say images or images for signal and noise)

images

Some practical examples can help to gain insight.

DToD for Bandlimited Waveforms

Let the noise be spectrally limited to Bw in Hz, with power spectral density images for images, and the spectral power density of the waveform upper limited to Bg (to simplify, images) with signal to noise ratio

images

The variance of DToD from (20.6) is

This decreases with the observation length T and ρ, when the term in brackets is negligible as for large ρ, or when the noise is wideband (images) and the ratio images cannot be neglected anymore. For small signal to noise ratio it decreases as 1/ρ2.

Another relevant example is when all processes have the same bandwidth (images) with the same power spectral density S(f) up to a scale factor: images, when

images

From the Schwartz inequality images, the variance is lower bounded by

This value is attained only for images const.

The case of time sequences with white Gaussian noise can be derived from the results above. In this case the overall number of samples is images with noise bandwidth images and power images. The effective bandwidth of the noise is images and for the signal is images so that the variance of DToD follows from (20.7):

images

When considering sampled signals, it is more convenient to scale the variance to the sample interval so as to have the performance measured in sample units:

images

This depends on the number of samples (N) and the signal to noise ratio ρ; again the term 1/ρ2 disappears for large ρ. In the case that both waveform and noise are white (images), images and thus

More details on DToD in terms of bias (namely when images) and variance are in [96].

20.3.2 Generalized Correlation Method

The cross‐correlation method for DToD is the most intuitive extension of ToD, but there is room for improvement if rigorously deriving the DToD estimator from ML theory. Namely, the ML estimator was proposed by G.C. Carter [97] who gave order to the estimators and methods discussed within the literature of the seventies. We give the essentials here while still preserving the formal derivations.

The basic assumption is that the samples of the Fourier transform of data si(t) (for images) evaluated at frequency binning images

images

are zero‐mean Gaussian random variables. Provided that T is large enough to neglect any truncation effect due to the DToD τ (see Section 20.4), the correlation depends on the cross‐spectrum

images

and it follows that

images

is uncorrelated vs. frequency bins. The rv at the k th frequency bin is

images

where

images

The set of rvs for all the N frequency bins images are Gaussian and independent, and the log‐likelihood is (apart from constant terms)

images

Still considering a large interval T, the ML metric for images becomes

images

where the dependency on the DToD τ is in the phase of S(f) as

images

The inverse of the covariance

images

depends on the coherence of the two rvs defined as:

images

The ML metric images(τ) can be expanded (recall the property: images) to isolate only the delay‐dependent term:

The MLE is based on is the search for the delayed cross‐correlation of the two signals (term images) after being filtered by Ψ(f). In practice, the DToD is by the correlation method provided that the two signals are preliminarily filtered by two filters H1(f) and H2(f) such that images as sketched in Figure 20.4.

Diagram of generalized correlation method displaying boxes labeled from H1(f) and H2(f) to ∫s̃1(t) s̃2(t = ζ)dt connected by arrows from s1(t) and s2(t) to s̃1(t) and s̃2(t).

Figure 20.4 Generalized correlation method.

Filter Design

Inspection of the coherence term based on the problem at hand

images

shows that it depends on the signal to noise ratio for every frequency bin

images

and it can be approximated as

images

it downweights the terms when noise dominate the signal. The filter Ψ(f) is

images

and a convenient way to decouple Ψ(f) into two filters is

where the first term decorrelates the signal for the waveform, and the second term accounts for the signal to noise ratio. If ρ(f) is independent of frequency, the second term is irrelevant for the MLE of DToD.

CRB for DToD

CRB can be derived from the log‐likelihood term that depends on the delay (20.10) and it follows from the definitions in Section 8.1:

images

or compactly

To exemplify, consider the case where signal and noise have the same bandwidth (e.g., this might be the case that lacks of any assumption on what is signal and what is noise, and measurements are prefiltered within the same bandwidth to avoid “artifacts”) images with images (i.e., this does not imply that imagesconst), then

images

coincides with the variance of correlation estimator (20.8) derived from a sensitivity analysis, except that any knowledge of the power spectral density of the waveform images could use now the prefiltering (20.11) to attain an improved performance for the DToD correlator. If sampling the signals at sampling interval images (i.e., maximal sampling rate compatible with the processes here), the sequences have images samples in total and the CRB coincides with (20.9).

Even if the generalized correlation method is clearly the MLE for DToD, there are some aspects to consider when planning its use. The filtering (20.11) needs to have a good control of the model, and this is not always the case as the difference ToD is typically adopted in passive sensing systems where there is poor control of the waveform from the source, and the use of one measurement as reference for the others is a convenient and simple trick, namely when measurements are very large. In addition, the filtering introduces edge effects due to the filters’ transitions h(t) that limits the valid range T of the measurements. When filters are too selective, the filter response h(t) can be too long for the interval T and the simple correlation method (20.5) is preferred.

20.4 Numerical Performance Analysis of DToD

Performance of DToD depends on the observation interval T as this establishes the interval where the two waveforms maximally correlate. However, in the case that the true DToD τ is significantly different compared to T (Figure 20.5), the overlapping region within the area showing the same delayed waveform reduces by images and thus the variance (20.8) scales accordingly as the non‐overlapped region acts as self‐noise:

images
Graphs of truncation effects in DToD displaying 2 fluctuating curves (combination of solid and dashed) with arrows depicting T (solid and dashed) and τ (dashed).

Figure 20.5 Truncation effects in DToD.

The same holds for any variance computation. This loss motivates the choice of “slant windowing” whenever one has the freedom to choose the two portions of observations s1(t) and s2(t) to correlate.

Measurements are usually gathered as sampled and all methods above need to be implemented by taking the sampling into account. Correlation for sampled signals needs to include the interpolation and resampling as part of the DToD estimation step as for the ToD method. However, the filtering might introduce artifacts that in turn degrade the overall performance except for the case when the DToD is close to zero. This is illustrated in the example below by generating a waveform from a random process with PSD within images and white noise (images). The signal is generated by filtering a white stochastic process with a Butterworth filter of order 8, and the second signal has been delayed using the DFT for fractional delay (see Appendix B). The fractional delay is estimated by using the parabolic regression discussed in Section 9.4. Below is the Matlab code for the Montecarlo test to evaluate the performance wrt the CRB.

Figure 20.6 shows the performance in terms of RMSE (measured in samples) compared to the CRB (dashed lines) for varying choices of the delay images and images samples. The performance attains the CRB for large SNR and images, while there is a small bias that justifies the floor for large SNR due to the parabolic regression on correlation evaluated for discrete‐time signals.

Image described by caption.

Figure 20.6 RMSE vs. SNR for DToD (solid lines) and CRB (dashed line).

20.5 Wavefront Estimation: Non‐Parametric Method (L=1)

When the waveform is not known and the measurements are highly redundant (images), the wavefront can be estimated by iterated DToD among all measurements [98] even if, because of small waveform distortion vs. distance (see Figure 20.8), it would be preferable to use the DToD of neighboring signals. The set of delayed waveforms over a wavefront

images

depends on two geometrical parameters x, y according to the planar coordinates of the sensors (see Figure 20.7). For every pair of sensors out of M, say i and j at geometrical positions pi and pj, the correlation estimator on s(t; pi) and s(t; pj) yields the estimate of the DToD Δτi,j that is dependent on the DToD of the true delays

images

apart from an error αi,j that depends on the experimental settings (waveform, noise, any possible distortion, and even delays) as proved in (20.12).

Graph of wavefront estimation from multiple DToD displaying 4 vertical fluctuating waves traversing along 4 downward arrows.

Figure 20.7 Wavefront estimation from multiple DToD.

All the pairs that are mutually correlated can be used to estimate the DToDs up to a total of images estimates (if using all‐to‐all pairs) to build a linear system

where the matrix images has values images along every row to account for the pair of sensors used in DToD, and all the other terms have obvious meanings. Assuming that the DToD estimates are

images

then the LS estimate of the ensemble of ToDs is

images

Notice that the wavefront images is estimated up to an arbitrary translation as DToDs are not sensitive to a constant, and this reflects to a null eigenvalue of ATA that calls for the LS estimate using the pseudoinverse (Section 2.7). Any iterative estimation method for the solution of the LS estimate (e.g., iterative methods in Section 2.9) are not sensitive to null eigenvalues as that part of the solution in images corresponding to the eigenvector of the constant value of τ is not updated.

20.5.1 Wavefront Estimation in Remote Sensing and Geophysics

In remote sensing and geophysical signal processing, the estimation of the wavefront of delays (in seismic exploration these are called horizons since they define the ToD originated by the boundary between different media, as illustrated at the beginning of this Chapter) is very common and of significant importance. Even if these remote sensing systems are usually active with known waveform, the propagation over the dispersive medium and the superposition of many backscattered echoes severely distort the waveform, and make it almost useless for ToD estimation. However, in DToD estimation the waveform is similarly distorted for two neighboring sensors. This is exemplified in Figure 20.8 where the waveforms are not exactly the same shifted copy, but there is some distortion both in wavefront and in waveform (see the shaded area that visually compares the waveforms at the center and far ends of the array).

Schematic displaying waves radiating from a commons source, depicting curved wavefront.

Figure 20.8 Curved and distorted wavefront (upper part), and measured data (lower part).

One common way to handle these issues is by preliminarily selecting those pair of sensors where the waveforms are not excessively distorted between one another, and this reduces the number of equations associated with the selected pairs of the overdetermined problem (20.13) to be smaller than images (all‐to‐all configuration). This has a practical implication as the number of scans can be images as in the seismic image at the beginning in Figure 20.1, and the use of M2/2 would be intolerably complex. Selecting the K neighbors (say images) can reduce the complexity to manageable levels KM/2. If M is not excessively large, one could still use all the images pairs and account separately for the waveform distortion by increasing the DToD variance (20.12). As a rule of thumb, the distortion increases with the sensors’ distance images, and this degrades the variance proportionally

images

for an empirical choice of the scaling term η and exponent γ.

It is quite common to have outliers in DToD and this calls for some preliminary data classification (Chapter 23). In addition to the robust estimation methods (Section 7.9), another alternative is to solve for (20.13) by using an L1 norm. These methods are not covered here as they are based on the specific application, but usually they are not in closed form even if the optimization is convex, and are typically solved by iterative methods.

20.5.2 Narrowband Waveforms and 2D Phase Unwrapping

In narrowband signals, the waveform is the product of a slowly varying envelope a(t) with a complex sinusoid:

images

where ωo is the angular frequency. The delays are so small compared to the envelope variations that the signals at the sensors (i, j) are

images

and the correlator‐based DToD estimator (neglecting the noise only for the sake of reasoning)

images

yields the estimate of the phase difference

that is directly related to the DToD. Wavefront estimation is the same as above with some adaptations [99]: phase difference (and phase in general3) is defined modulo‐ 2π from the arg[.] of the correlation of signals. Since images phase is wrapped within images (herein indicated as [.]2π) so that the true measured phase difference differs from the true one by multiple of 2π, unless images. 2D phase unwrapping refers to the estimation of the continuous phase‐function

images

(or the wavefront images) from the phase‐differences (20.14) evaluated modulo‐ 2π (also called wrapped phase). The LS unwrapping is similar to problem (20.13) by imposing for any pair of measurements the equality

in the LS sense. The system becomes

(20.16)images

that yields the LS solution

(20.17)images

Since phase is measured modulo‐ 2π, the choice of the sensor’s pairs is not extended to all the sensors, but rather to the neighborhood of each sensor as the phase‐variation among neighbor sensors is expected to be mostly smaller than π.

The concept of phase aliasing is illustrated by the example in Figure 20.9 where phase ϑi increases along the sensors’ position and the wrapped phase [ϑi]2π has 2π jumps when close to images. If considering the angular difference of the phase between two points coarsely sampled 1:2 (gray points with double‐spacing images), the modulo‐ 2π value coincides with the true value except where phase difference images (gray lines) while the wrapped phase difference images (i.e., the wrapped phase difference can be considered as the smallest angular difference between two consecutive vectors). Since the phase difference estimated from the wrapped phase differs from the true one by (a multiple of) 2π, there is no way to retrieve these images jumps that impair the overall estimated phase‐function by the unwrapping algorithm. This is called phase aliasing condition. If sampling is dense enough (gray and black points in Figure 20.9), the phase‐differences are likely to be coincident with the true (unwrapped) phase without any artifacts.

2 Graphs displaying wave curves with markers connected by vertical dashed lines, illustrating phase-function and wrapped phase in (−π, π) from modulo-2π.

Figure 20.9 Phase‐function and wrapped phase in images from modulo‐ 2π.

20.5.3 2D Phase Unwrapping in Regular Grid Spacing

In addition to wavefront estimation, 2D phase unwrapping has several notable applications in magnetic resonance imaging, optical interferometry, and satellite synthetic aperture radar (SAR) imaging. All these applications collect measurements over a regular grid spacing and the phase‐differences (20.15) are computed between neighboring points along two orthogonal directions. Since the transformation A is the derivative along the two dimensions, the phase differences can be partitioned along these two directions as

images

so that the LS estimate becomes

which can be solved efficiently by exploiting the structure of the linear operators and their sparsity—see Appendix C.

Inspection of the term images shows that it contains the second order derivatives (i.e., discrete Laplace operator) of the unwrapped phase field images, while the terms on the right of (20.18) are the first‐order derivatives of the phase‐differences. This is the basis for several detailed lines of thought on the impact of phase‐difference artifacts such as the 2π discontinuities on the final solutions. The reader is referred to the copious literature on this fascinating topic that deals with the phase field as a continuous field that is sampled, and phase unwrapping artifacts are the results of coarse sampling of these phase fields.

The example in Figure 20.10 shows the consequences of phase aliasing from the wrapped phase image [ϑ(x, y)]2π using (20.18) when the phase image is down‐sampled 1:n (i.e., take one sample out of n in both directions) to simulate phase aliasing when evaluating phase‐differences. The estimated phase images is (visually) not affected by 1:2 decimation even if some residual is still visible on residual images, while it is more severely affected when decimation is larger, as for 1:5, most of the artifacts are in those areas where there is a sudden change of the phase‐function (e.g., in the face, where luminance is abruptly changing from black to white). Residuals for 1:2 or 1:5 show a pattern that resembles the electric field with dipoles, there is a close analytical relationship, but the explanation is beyond the scope of the introductory discussion.

6 Images of a dog illustrating 2D phase unwrapping.

Figure 20.10 Example of 2D phase unwrapping.

One might argue that the solution to these artifacts could be to oversample the wavefields, but this is not always possible as the sensors’ density is given by the application. The interpolation of phase images cannot provide the solution as unfortunately phase information is related to the dynamics of the phase and not (easily) to the spectrum of the signals. Since phase aliasing introduces missing 2π jumps that are smeared by the LS method as diffused error (see 1:5 decimation around the face with abrupt black‐white discontinuities), these can be interpreted as outliers that are surely not‐Gaussian and the L2‐norm can be replaced with L1 metrics, possibly iterated. There are methods that solve for the LS phase estimate by cosine‐transforms, but this method is questionable for small images due to the poor control of artifacts at the boundary as consequence of the periodicity of the cosine‐transformation.

20.6 Parametric ToD Estimation and Wideband Beamforming

The ToDs images depend on a set of shape parameters that fully characterize each waveform. A common and intuitive, but not unique, example is the wavefield originated by point sources at distances h from a linear array of sensors located at images (see Figure 20.11). The 2D ray‐geometrical model of propagation in a homogeneous medium with propagating velocity v (in m/s) gives the hyperbolic ToD vs. array space x (here assumed as continuous just to ease the notation) in the two equivalent formulations:

images

where the two shape parameters

images

are the geometrical position of the source, or alternatively when v is not known, the ToD at the vertex of the hyperbola images: images. Note that in general, the wavefront is curved and, for large distances (far‐field approximation for the wavefront) images, it is linear with a slope

images

that depends on the slowness 1/v of the propagating medium. The example in Figure 20.11 shows the wavefields from 3 sources for wideband g(t) (left) and narrowband waveforms, and wideband s(x, t) with white noise or the same after filtering the noisy wavefield within the bandwidth of the signal. All slopes of the wavefronts are below the value defined by the propagation velocity images 1500 m/s that mimics the acoustic waves in water as for a sonar sensing system, and hyperbolic ToDs images are indicated by the dashed lines superimposed on the wavefield images in various settings.

Graphs displaying delayed waveforms (dashed lines) from three sources impinging on a uniform linear array, with labels (left–right) array, g(t) narrowband, s(x,t) + white noise, and s(x,t) + filtered noise.

Figure 20.11 Example of delayed waveforms (dashed lines) from three sources impinging on a uniform linear array.

Furthermore, there are applications such as in subsurface remote sensing where backscattered wavefields are used for depth imaging and the propagation velocity v is unknown and this becomes an additional parameter to be estimated—or even worse (but unfortunately very common in many geophysical seismic exploration applications), the propagating medium is not homogeneous, the wavefield is not spherical, and the wavefront deviates from the hyperbolic shapes of the Figure 20.11. In these contexts, the number of parameters of a parametric ToD model become very complex and this is not covered herein. However, the approach is still based on estimation of the deviations from a nominal hyperbolic ToD shape based on some initial assumptions of the propagation velocity.

In complex contexts, there are many wavefronts impinging on an array of sensors and the signal at position images

is characterized by the superposition of these delayed waveforms. The waveforms are not known but redundantly measured by the array to guarantee the possibility of estimating one waveform and mitigating (or attenuating) the others. Wideband beamforming refers to the method to delay and combine the multiple signals (20.19) to estimate the waveform of interest (say images for the images ‐th wavefront), and possibly mitigate the others considered as interference.

20.6.1 Delay and Sum Beamforming

Delay and sum beamforming extends the beamforming methods in Chapter 19 to wideband waveforms. Making reference to the model (20.19), the wideband beamforming method delays each signal by a certain shape parameters images and sums all the N delayed signals from all sensors:

When the shape parameters match the images th wavefield images, this value flattens s(t, x) in (20.20) and the sum enhances it; the resulting signal becomes

where the three terms are the signal of interest, interference from the other wavefronts (images) that are reduced by the summing, and the noise that is reduced by 1/N. When images, the signal is somewhat attenuated by the sum.

Wideband beamforming is shown in the example in Figure 20.12 for a point source at images and images (or images as from the vertex of the hyperbolic wavefield) from the linear array, and propagation velocity is images as for ultrasound systems with Ricker waveform (Section 9.4). The delay and sum beamforming images is for varying source point images and the clearest waveform is when images as expected, with small artifacts when images that decrease with N. The interference mitigation capability of delay and sum beamforming is shown for the same example when considering the presence on another interfering wavefield at images and images, and noise images is affected by the impairment, but surely much less than the original observation s(t, x).

4 Graphs with labels (left–right) s(t,x), sout(t), s(t,x) w/noise and interference, and sout(t), illustrating delay and sum beamforming for wideband signals.

Figure 20.12 Delay and sum beamforming for wideband signals.

The corresponding Matlab code for delay and sum beamforming described here follows, together with the generation of the signal (a gray‐scale image is used in place of wiggle plot in the figure, which enhances any waveform distortion after the sum).

Remark 1: The signal images after the delay and sum beamforming is maximized for those choices of the shape parameters images that flattens the wavefields images before summing up. The energy of images within a predefined window can be used to estimate the presence or not of one source for a certain combination of shape parameters. The search is over the bidimensional pairs images for the value(s) that peak(s) the energy metric, and time resolution is limited by the time window adopted in the energy metric.

Remark 2: After delaying, the sum in (20.21) can take into account the different degree of noise and interference occurring in the flattened wavefield. This is obtained by weighting the terms before summing, and the weighting can be designed according to the degree of noise and interference that is experienced in the sum. The optimal weighting is the same as for BLUE (Section 6.5) and depends on the inverse of the power of the noise and interference images for the specific shape parameters images:

images

with normalization images

20.6.2 Wideband Beamforming After Fourier Taransform

Implementation of delay and sum beamforming for discrete‐time signals needs to interpolate the signals when the delay is fractional. This is carried out by appropriate filtering as part of the delay for discrete‐time signals as detailed in Appendix B.

Alternatively, one can use the delay after the Fourier transform of the wavefields vs. time, which has several benefits. The model (20.19) after FT is

images

and, for every frequency, it is the superposition of spatial sinusoids each with appropriate spatial varying amplitudes α(x). A wideband beamforming becomes:

images

and avoids any explicit interpolation step.

Another simplification of wideband modeling is for far‐field when the wavefront is planar, or around a certain position images.

images

When the amplitude variations across space is negligible (images) the model is

images

where

images

These are the sum of spatial sinusoids that can be handled using the theory of narrowband array processing as illustrated in Chapter 19.

Appendix A: Properties of the Sample Correlations

Preliminarily, we can notice that for any WSS process x(t) the mean values can be evaluated as

images

where the last equality is due to the limited range of the autocorrelation within images. If the interval T is much larger than the decorrelation of the process:

images

These approximations ease the computations of the terms:

images
images

that reduce to those in the main text.

Appendix B: How to Delay a Discrete‐Time Signal?

Fractional delay (not sample‐spaced) of a discrete‐time signal is very frequent in statistical signal processing. Let x1, x2, …, xN be a sequence to be delayed by τ samples, with τ non‐integer (otherwise, this would be simply a time‐shift of the samples) and sample index indicated by subscript. The solution is to interpolate the sequence with an appropriate continuous‐time interpolating function h(t) that estimates at best the continuous‐time signal:

images

assuming that the sampling interval is unitary. The interpolated signal images can now be arbitrarily delayed

images

and the delayed signal is uniformly resampled to have the delayed sequence:

images

The last equality shows that the delayed sequence is the convolution of the original sequence and the interpolating function delayed by τ and resampled.

Accuracy of fractional delay depends on the choice of the interpolating function h(t) that must honor the original samples (i.e., images) as discussed in Section 5.5. The ideal interpolating function

images

is not commonly used as the tails decay too slowly and the sequence needs to be very long to avoid edge effects. One of the most common ways to interpolate is by using the spline (see the book by de Boor [100] for an excellent overview).

A very pragmatic way to delay a sequence is by the use of the discrete Fourier transform (DFT) under the assumption of periodicity of the sequence. Namely, from the sequence x1, x2, …, xN, the DFT is computed and the result in turn is delayed by a linear phase:

images

In spite of its simplicity, there are pitfalls to be considered to avoid “surprises” after fractional delays. Namely, the use of the DFT still uses an interpolation, which can be easily seen from the definition of the inverse DFT of Y(k):

images

which is periodic over N samples. Namely, when using the DFT all sequences are periodic, even if not originally. If the delay τ is small and the original sequence tapered to zero, these wrap‐around effects are negligible. Alternatively, the original sequence can be padded by zeros up to N′ (with images) to make the implicit periodic sequence long enough to reduce these artifacts (but still not zero after non‐fractional delays). Needless to say, the DFT must be computed with basis N′.

Appendix C: Wavefront Estimation for 2D Arrays

Let images denote the matrix of (unknown) delay function (bold font indicates matrix), images and images the matrixes of gradients, and images the shift matrix (i.e., images). The ToD of the wavefront can be related to the DToD along the two directions for the neighboring points (gradients of the wavefront):

(20.22)images

or equivalently

where images and images are operators for the evaluation of the gradients from the wavefront images. The equation (20.23) represents an overdetermined set of images linear equations and MN unknowns that can be rearranged using the properties of the Kronecker products (Section 1.1.1)

(20.24)images

here images and images. The LS wavefront estimator is

(20.25)images

or equivalently

(20.26)images

which is the discrete version of the partial differential equation with boundary conditions [99], where images (or images) is the discrete version of the second order derivative along x (or y). The matrixes are large but sparse, and this can be taken into account when employing iterative methods for the solution (Section 2.9) that converge even if A is ill conditioned (i.e., the update rule is independent of any constant additive value).

The Matlab code below exemplifies how to build the numerical method to solve for the wavefront from the set of DToD, or for 2D phase unwrapping when the phase‐differences images are evaluated modulo‐ 2π. Since all matrixes Dx and Dy are sparse, its is (highly) recommended to build these operators using the sparse toolbox (or any similar toolbox).

Notes

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.16.51.3