The Wiener filter as originally proposed by Norbert Wiener is a signal processingfilter which uses knowledge of the statistical properties of both the signal and the noise to reconstruct an optimal estimate of the signal from a noisy one-dimensional time-ordered data stream. The generalized Wiener filter generalizes the same idea beyond the domain of one-dimensional time-ordered signal processing, with two-dimensional image processing being the most common application.[1]
Description
Consider a data vector which is the sum of independent signal and noise vectors with zero mean and covariances and . The generalized Wiener Filter is the linear operator which minimizes the expected residual between the estimated signal and the true signal, . The that minimizes this is , resulting in the Wiener estimator . In the case of Gaussian distributed signal and noise, this estimator is also the maximum a posteriori estimator.
The generalized Wiener filter approaches 1 for signal-dominated parts of the data, and S/N for noise-dominated parts.
An often-seen variant expresses the filter in terms of inverse covariances. This is mathematically equivalent, but avoids excessive loss of numerical precision in the presence of high-variance modes. In this formulation, the generalized Wiener filter becomes using the identity .
An example
The cosmic microwave background (CMB) is a homogeneous and isotropic random field, and its covariance is therefore diagonal in a spherical harmonics basis. Any given observation of the CMB will be noisy, with the noise typically having different statistical properties than the CMB. It could for example be uncorrelated in pixel space. The generalized Wiener filter exploits this difference in behavior to isolate as much as possible of the signal from the noise.
The Wiener-filtered estimate of the signal (the CMB in this case) requires the inversion of the usually huge matrix . If S and N were diagonal in the same basis this would be trivial, but often, as here, that isn't the case. The solution must in these cases be found by solving the equivalent equation , for example via conjugate gradients iteration. In this case all the multiplications can be performed in the appropriate basis for each matrix, avoiding the need to store or invert more than their diagonal. The result can be seen in the figure.[citation needed]