In digital signal processing, linear prediction is often called linear predictive coding (LPC) and can thus be viewed as a subset of filter theory. In system analysis, a subfield of mathematics, linear prediction can be viewed as a part of mathematical modelling or optimization.
The prediction modelEdit
The most common representation is
where is the predicted signal value, the previous observed values, with , and the predictor coefficients. The error generated by this estimate is
where is the true signal value.
These equations are valid for all types of (one-dimensional) linear prediction. The differences are found in the way the predictor coefficients are chosen.
For multi-dimensional signals the error metric is often defined as
Estimating the parametersEdit
The most common choice in optimization of parameters is the root mean square criterion which is also called the autocorrelation criterion. In this method we minimize the expected value of the squared error , which yields the equation
for 1 ≤ j ≤ p, where R is the autocorrelation of signal xn, defined as
where the autocorrelation matrix is a symmetric, Toeplitz matrix with elements , the vector is the autocorrelation vector , and , the parameter vector.
Another, more general, approach is to minimize the sum of squares of the errors defined in the form
where the optimisation problem searching over all must now be constrained with .
On the other hand, if the mean square prediction error is constrained to be unity and the prediction error equation is included on top of the normal equations, the augmented set of equations is obtained as
where the index ranges from 0 to , and is a matrix.
Specification of the parameters of the linear predictor is a wide topic and a large number of other approaches have been proposed. In fact, the autocorrelation method is the most common and it is used, for example, for speech coding in the GSM standard.
Solution of the matrix equation is computationally a relatively expensive process. The Gaussian elimination for matrix inversion is probably the oldest solution but this approach does not efficiently use the symmetry of . A faster algorithm is the Levinson recursion proposed by Norman Levinson in 1947, which recursively calculates the solution. In particular, the autocorrelation equations above may be more efficiently solved by the Durbin algorithm.
In 1986, Philippe Delsarte and Y.V. Genin proposed an improvement to this algorithm called the split Levinson recursion, which requires about half the number of multiplications and divisions. It uses a special symmetrical property of parameter vectors on subsequent recursion levels. That is, calculations for the optimal predictor containing terms make use of similar calculations for the optimal predictor containing terms.
For equally-spaced values, a polynomial interpolation is a linear combination of the known values. If the discrete time signal is estimated to obey a polynomial of degree then the predictor coefficients are given by the corresponding row of the triangle of binomial transform coefficients. This estimate might be suitable for a slowly varying signal with low noise. The predictions for the first few values of are
- Ramirez, M. A. (2008). "A Levinson Algorithm Based on an Isometric Transformation of Durbin's" (PDF). IEEE Signal Processing Letters. 15: 99–102. doi:10.1109/LSP.2007.910319.
- Delsarte, P. and Genin, Y. V. (1986), The split Levinson algorithm, IEEE Transactions on Acoustics, Speech, and Signal Processing, v. ASSP-34(3), pp. 470–478
- Hayes, M. H. (1996). Statistical Digital Signal Processing and Modeling. New York: J. Wiley & Sons. ISBN 978-0471594314.
- Levinson, N. (1947). "The Wiener RMS (root mean square) error criterion in filter design and prediction". Journal of Mathematics and Physics. 25 (4): 261–278.
- Makhoul, J. (1975). "Linear prediction: A tutorial review". Proceedings of the IEEE. 63 (5): 561–580. doi:10.1109/PROC.1975.9792.
- Yule, G. U. (1927). "On a Method of Investigating Periodicities in Disturbed Series, with Special Reference to Wolfer's Sunspot Numbers". Phil. Trans. Roy. Soc. A. 226: 267–298. doi:10.1098/rsta.1927.0007. JSTOR 91170.