Jump to content

Minimax approximation algorithm

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Gareth Jones (talk | contribs) at 18:29, 29 August 2012 (wlink author, correct typo). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

A minimax approximation algorithm is a method which aims to find an approximation such that the maximum error is minimized. Suppose we seek to approximate the function f(x) by a function p(x) on the interval [a,b]. Then a minimax approximation algorithm will aim to minimize[1]

Polynomial expansions such as the Taylor series expansion are often convenient for theoretical work but less useful for practical applications. For practical work it is often desirable to minimize the maximum absolute or relative error of a polynomial fit for any given number of terms in an effort to reduce computational expense of repeated evaluation.

One popular approach is the Remez algorithm.

References

  1. ^ Powell, M. J. D. (1981). "7: The theory of minimax approximation". Approximation Theory and Methods. Cambridge University Press. ISBN 0521295149.