Beruflich Dokumente
Kultur Dokumente
Approximation Theory
Sometimes, we even further require that, for each i, the rst ri derivatives
of agree with those of f at xi .
1
2 CHAPTER 7. APPROXIMATION THEORY
Theorem 7.0.1 Let f (x) be a piecewise continuous function over the interval
[a, b]. Then for any > 0, there exist an integer n and numbers a0 , . . . , an such
b n
that a {f (x) ai xi }2 dx < .
i=0
converges to f (x) as n .
n
(x xj )
j=1
E(x) := f (x) p(x) = f (n) () (7.9)
n!
where min{x1 , . . . , xn , x} < < max{x1 , . . . , xn , x}.
n1
(pf): Suppose p(x) = ak xk where the coecients ak are to be deter-
k=0
n1
mined. Then p(xi ) = ak xki = f (xi ), i = 1, . . . , n, can be written in the
k=0
form
1, x1 , . . . , xn1
1 a0 f (x1 )
= . (7.10)
1, xn , . . . , xn1
n an1 f (xn )
The matrix, known as the van Dermonde matrix, has determinant (xi
i>j
xj ). Since all xi s are distinct, we can uniquely solve (7.10) for the unknowns
4 CHAPTER 7. APPROXIMATION THEORY
n
(x xi )
F (x) = f (x) p(x) (f (x0 ) p(x0 )) i=1 .
n
(x0 xi )
i=1
n!
f (n) () (f (x0 ) p(x0 )) = 0.
n
(x0 xi )
i=1
n
(x0 xi )
Thus E(x0 ) = f (x0 ) p(x0 ) = i=1
n! f (n) (). Since x0 is arbitrary, the
theorem is proved.
Denition 7.1.1 The polynomial p(x) dened by (7.7) is called the Lagrange
interpolation polynomial.
1
f (x) = (7.12)
1 + 25x2
7.2. NEWTONS INTERPOLATION FORMULA 5
in the interval [1, 1] at equally spaced points. Runge (1901) discovered that as
the degree n of the interpolating polynomial pn (x) tends toward innity, pn (x)
diverges in the intervals .726 . . . |x| < 1 while pn (x) works pretty well in the
central portion of the interval.
Let Pi0 i1 ...ik (x) represent the k-th degree polynomial for which
for j = 0, . . . , k.
holds.
(pf): Denote the right-hand side of (7.14) by R(x). Observe that R(x) is a
polynomial of degree k. By denition, it is easy to see that R(xij ) = f (xij )
for all j = 0, . . . , k. That is, R(x) interpolates the same set of data as does the
polynomial Pi0 i1 ...ik (x). By Theorem 7.1.1 the assertion is proved.
6 CHAPTER 7. APPROXIMATION THEORY
The dierence Pi0 i1 ...ik (x) Pi0 i1 ...ik1 (x) is a k-th degree polynomial which
vanishes at xij for j = 0, . . . , k 1. Thus we may write
Pi0 i1 ...ik (x) = Pi0 i1 ...ik1 (x) + fi0 ...ik (x xi0 )(x xi1 ) . . . (x xik1 ). (7.15)
The leading coecients fi0 ...ik can be determined recursively from the formula
(7.14), i.e.,
where fi1 ...ik and fi0 ...ik1 are the leading coecients of the polynomials Pi1 ...ik (x)
and Pi0 ...ik1 (x), respectively.
Remark. Note that the formula (7.16) starts from fi0 = f (xi0 ).
Remark. The polynomial Pi0 ...ik (x) is uniquely determined by the set of sup-
port data {(xij , fij )}. The polynomial is invariant to any permutation of the
indices i0 , . . . , ik . Therefore, the divided dierences (7.16) are invariant to per-
mutation of the indices.
It follows that the k-th degree polynomial that interpolates the set of support
data {(xi , fi )|i = 0, . . . , k} is given by
(0) (r )
Given {xi }, i = 1, . . . k and values ai , . . . , ai i where ri are nonnegative inte-
gers. We want to construct a polynomial P (x) such that
(j)
P (j) (xi ) = ai (7.21)
k
Remark. The degree of P (x) is at most (ri + 1) 1.
i=1
(j)
Theorem 7.3.1 Given the nodes {xi }, i = 1, . . . , k and values {ai }, j = 0, . . . , ri ,
there exists a unique polynomial satisfying (7.21).
k
(j)
Then P (x) is of degree (ri + 1) 1. Now P (j) (x1 ) = a1 for j = 0, . . . , r1
i=1
(j)
(0) (0) (j) (j) (j) a1
implies =a1 c1 , . . . , a1 = c1 j!. So q1 (x) is determined with c1 = j! .
Now we rewrite (7.23) as
P (x) q1 (x)
R(x) := = q2 (x) + (x x2 )r2 +1 q3 (x) + . . .
(x x1 )r1 +1
Note that R(j) (x2 ) are known for j = 0, . . . , r2 since P (j) (x2 ) are known. Thus
(j)
all c2 , hence q2 (x), may be determined. This procedure can be continued to
determine all qi (x). Suppose Q(x) = P1 (x) P2 (x) where P1 (x) and P2 (x) are
k
two polynomials of the theorem. Then Q(x) is of degree (ri + 1) 1,
i=1
and has zeros at xi with multiplicity ri + 1. Counting multiplicities, Q(x) has
k
(ri + 1) zeros. This is possible only if Q(x) 0.
i=1
8 CHAPTER 7. APPROXIMATION THEORY
(2) Suppose ri = 1 for all i = 1, . . . , k. That is, suppose values of f (xi ) and
f (xi ) are to be interpolated. Then the resultant (2k 1)-degree polynomial
is called the Hermite interpolating polynomial. Recall that the (k 1)-degree
polynomial
k
x xj
i (x) = (7.24)
j=1
xi xj
j=i
hi (xj ) = ij ;
gi (xj ) = 0; (7.28)
hi (xj ) = [1 2(x xi )i (xi )]2i (x)i (x) 2i (xi )i (x)|x=xj = 0;
2
k
P (x) = f (xi )hi (x) + f (xi )gi (x)). (7.29)
i=1
Thus far for a given function f of an interval [a, b], the interpolation has been
to construct a polynomial over the entire interval [a, b]. There are at least two
disadvantages for the global approximation:
7.4. SPLINE INTERPOLATION 9
1. For better accuracy, we need to supply more support data. But then
the degree the resultant polynomial gets higher and such a polynomial is
dicult to work with.
2. Suppose f is not smooth enough. Then the error estimate of an high
degree polynomial is dicult to establish. In fact, it is not clear whether
or not that the accuracy will increase with increasing number of support
data.
Denition 7.4.1 Let the interval [a, b] be partitioned into a = x1 < x2 < . . . <
xn = b. A function p(x) is said to be a cubic plite of f on the partition if
f (xi ) i Mi
ci = (7.32)
i 6
f (xi+1 ) i Mi+1
di = . (7.33)
i 6
Thus
2
3i1 Mi f (xi ) f (xi1 ) i1 (Mi Mi1 )
+
6i1 i1 6
3i2 Mi f (xi+1 ) f (xi ) i (Mi+1 Mi )
= + , (7.36)
6i i 6
or equivalently,
i1 i + i1 i
Mi1 + Mi + Mi+1
6 3 6
f (xi+1 ) f (xi ) f (xi ) f (xi1 )
= . (7.37)
i i1
7.5. TRIGONOMETRIC INTERPOLATION 11
a0
M
(x) = + (ah cos hx + bh sin hx) (7.42)
2
h=1
a0
M1
aM
(x) = + (ah cos hx + bh sin hx) + cos M x, (7.43)
2 2
h=1
12 CHAPTER 7. APPROXIMATION THEORY
For simplicity, we shall consider equally spaced notes. Without loss of gen-
erality, we shall assume
2k
xk = , k = 0, . . . , N 1.
N
Observe that
ei(hxk ) = ei(2hk/N ) = ei2(N h)k/N = ei(N h)xk .
Thus, we may write
eihxk + ei(N h)xk
cos hxk = ;
2
eihxk ei(N h)xk
sin hxk = . (7.44)
2i
a0
M
eihxk + ei(N h)xk eihxk ei(N h)xk
(xk ) = + (ah + bh )
2 2 2i
h=1
2k
(pf): It only remains to show (7.50). Recall that k = eixk = ei N . Thus
k = jk and kj = kj . Observe that jh
j N
= 1. That is, jh is a root of the
N1
polynomial N 1 = ( 1) k . Thus it must be either jh = 1 which is
k=0
N 1 N1
N 1
the case j = h, or k
jh = kjh = kj kh = 0. We may therefore
k=0 k=0 k=0
summarize that
N 1
0, if j= 0
kj kh = (7.51)
N, if j = h.
k=0
Introducing (h) := (1, 1h , . . . , N
h
1 )
T
C N , we may rewrite (7.51) in terms
of the complex inner product
0, if j = h
(j) , (h) = . (7.52)
N, if j=h
That is, the vectors (0) , . . . (N 1) form an orthogonal basis of C N . Denote
f := (f0 , . . . , fN 1 )T .
N 1
2
bh = fk sin hxk . (7.55)
N
k=0
Theorem 7.5.2 Let p(x) be the phase polynomial that interpolates a given set
of support data (xk , fk ), k = 0, . . . , N 1 with xk = 2k/N . Then the s-segment
ps (x) of p(x) minimizes the sum
N 1
S(q) = |fk q(xk )|2 (7.57)
k=0
s
(pf): Introducing the vectors ps := (ps (x0 ), . . . ps (xN 1 ))T = j (j) and
j=0
s
q := (q(x0 ), . . . , q(xN 1 ))T = j (j) in C N , we write S(q) = f q, f q.
j=0
1
f ps , w(j) = j j = 0,
N
and hence
s
f ps , ps q = f ps , (j j ) (j) = 0.
j=1
Remark. (7.5.2) states the important property that the truncated trigono-
metric interpolation polynomial p(x) produces the least-squares trigonometric
approximation ps (x) of all data.
7.6. FAST FOURIER TRANSFORM 15
Suppose that we have a series of sines and cosines which represent a given
function of [L, L], say,
a0 nx nx
f (x) = + an cos + bn sin . (7.58)
2 n=1
L L
and
L
nx mx
sin cos dx = 0, (7.61)
L L L
one can show that
L
1 nx
an = f (x) cos dx, n = 0, 1, . . . (7.62)
L L L
L
1 nx
bn = f (x) sin dx, n = 1, 2, . . . (7.63)
L L L
Denition 7.6.1 The series (7.58), with coecients dened by (7.62) and
(7.63), is called the fourier series of f (x) on [L, L].
Given a function f (x), its Fourier series does not necessarily converge to
f (x) at every x. In fact,
Theorem 7.6.1 Suppose f (x) is piecewise continuous on [L, L]. Then (1) If
f (x0 + h) f (x0 )
x0 (L, L) and f (x+
0 ) and f (x0 ) both exist (where f (x ) := lim ),
h0 h
+
f (x0 )+f (x0 )
then the Fourier series converges to 2 . (2) At L or L, if f (L+ )
f (L+ )+f (L )
and f (L ) exist, then the Fourier series converges to 2 .
16 CHAPTER 7. APPROXIMATION THEORY
Remark. Suppose that f (x) is dened on [0, L]. We may extend f (x) to become
an even function on [L, L] (simply by dening f (x) := f (x) for x [L, 0]).
In this way, the Fourier coecients for the extended function become
L
2 nx
an = f (x) cos dx (7.64)
L 0 L
bn 0.
Similarly, we may extend f (x) to become an odd function on [L, L], in which
case we obtain a pure sine series
nx
f (x) = bn sin (7.66)
n=1
L
with
L
2 nx
bn = f (x) sin dx. (7.67)
L 0 L
Example.
4 cos 3x cos 5x
f (x) = (cos x + + + . . .).
2 32 52
2. If g(x) = x, x , then
3. If
x( x), for 0x
h(x) =
x( + x), for x 0
then
8 sin 3x sin 5x
h(x) = (sin x + + + . . .).
33 53
7.6. FAST FOURIER TRANSFORM 17
N 1
2
an g(yk ) cos(nyk )
N
k=0
N 1
2 2k 2kn
= (1)n f( ) cos := a
n . (7.69)
N N N
k=0
a0
M
+ an cos ny + bn sin ny
2 n=1
0
M
a
+ n cos n(x ) + bn sin n(x ).
a
2 n=1
Remark. The basis of the Fourier analysis method for smoothing data is as
follows: If we think of given numerical data as consisting of the true values of
a function with random errors superposed, the true functions being relatively
smooth and the superposed errors quite unsmooth, then the examples above sug-
gest a way of partially separating functions from error. Since the true function
is smooth, its Fourier coecients will decrease quickly. But the unsmoothness
of the error suggests that its Fourier coecients may decrease very slowly, if at
all. The combined series will consist almost entirely of error, therefore, beyond
a certain place. If we simply truncate the series at the right place, then we are
discarding mostly error (although there will be error contribution in the terms
retained).
jx kx
Lemma 7.6.1 The functions ei L and ei L are orthogonal in the following
sense:
L
i jx i kx 0 if k = j
e L e L = . (7.70)
L 2L, if k = j
f (x) = fn einx (7.75)
n=
7.6. FAST FOURIER TRANSFORM 19
where 2
1
fn = f (x)einx dx. (7.76)
2 0
Consider the function g(x) of the form
g(x) = dj eijx (7.77)
j=
N 1
= dj ( eijxk einxk ). (7.79)
j= k=0
The central idea behind the fast Fourier transform (FFT) is that when N is
the product of integers, the numbers dj (or j ) prove to be closely interdepen-
dent. This interdependence can be exploited to substantially reduce the amount
of computation required to generated these numbers. We demonstrate the idea
as follows:
N 1 N 1
1 i 2jn 1
j = j1 +t1 j2 = f (xn )e N = f (xn ) nj
N n=0 N n=0
20 CHAPTER 7. APPROXIMATION THEORY
N 1
1
= f (xn ) j1 n2 +j1 t2 n1 +t1 j2 n2
N n=0
t2 1 t
1 1
1
= ( f (xn2 +t2 n1 ) j1 t2 n1 ) j1 n2 +t1 j2 n2 . (7.81)
N n =0 n =0
2 1
t2 1
1
j = F2 (j1 , j2 ) := F1 (j1 , n2 ) j1 n2 +t1 j2 n2 . (7.83)
N n =0
2
n2
0 1 2
0 0 0 0
j1
1 0 2 2
j2
0 1 2
0 0 0 0
j1
1 2 3i 0 2 3i
7.7. UNIFORM APPROXIMATION 21
Suppose N = t1 t2 t3 . Let
j = j1 + t1 j2 + t1 t2 j3
n = n3 + t3 n2 + t3 t2 n1 .
Then in the nine power terms in nj , three will contain the product t1 t2 t3 and
hence may be neglected. The remaining six power terms may be grouped into
a three step algorithm:
1 1
t
F1 (j1 , n2 , n3 ) := f (xn ) j1 t3 t2 n1 (7.84)
n1 =0
2 1
t
F2 (j1 , j2 , n3 ) := F (j1 , n2 , n3 ) (j1 +t1 j2 )t3 n2 (7.85)
n2 =0
t3 1
1
j = F3 (j1 , j2 , j3 ) := F2 (j1 , j2 , n3 ) (j1 +t1 j2 +t1 t2 j3 )n3 . (7.86)
N n =0
3
3
We note that if N = 10 , then only 3% of the original terms are needed.
(pf): Let
M := {x [a, b]||f (x) g(x)| = }. (7.89)
where e(x) := f (x) g(x). The inequality in (7.90) is possible if and only if the
sign of p(x) is the same as that of e(x). That is, we must have (f (x)g(x))p(x) >
0 for all x M . By (7.88), it follows that the polynomial p must change signs
at least n times in [a, b]. That is, p must have at least n zeros. This contradict
with the assumption that p is not identically zero.
Remark. The above theorem asserts only that g is a best approximation when-
ever there are at least n + 1 points satisfying (7.87) and (7.88). In general, there
can be more points where the maximal deviation is achieved.
Remark. The only property of Pn1 we have used to establish Theorem 7.7.1
is a weaker form of the Fundamental Theorem of Algebra, i.e., any polynomial
of degree n 1 has at most n 1 distinct zeros in [a, b]. This property is in fact
shared by a larger class of functions.
We now state without proof the famous result that Theorem 7.7.1 is not
only sucient but is also necessary for a polynomial g to a best approximation.
The following theorem is also known as the Alternation Theorem:
Denition 7.7.2 The set of points {x1 , . . . , xn+1 } in Theorem 7.7.2 is referred
to as an alternant for f and g.
Corollary 7.7.1 For any f C[a, b], there is a unique best approximation.
Theorem 7.7.1 also provides a basis for designing a method for the compu-
tation of best approximations of continuous functions. The idea, known as the
exchange method of Remez, is as follows:
4. Replace {xk } by the new extreme points and repeat the sequence of steps
given above beginning with step (2).
The objective here is to have the set {xk } converge to a true alternant and
hence the polynomial converge to a best approximation. It can be proved that
the process does converge for any choice of starting values in step (1) for which
the value of computed in step (2) in not zero. With additional assumptions on
the dierentibility of f , it can also be shown that convergence is quadratic. Also
the assumption that e(x) possesses exactly n + 1 extrem points in step (3) is
not essential. For more details, refer to G. Meinarduss book Approximation of
Functions: Theory and Numerical Methods, Springer-Verlag, New York, 1967.