Adomian decomposition method
The Adomian decomposition method (ADM) is a semi-analytical method for solving ordinary and partial nonlinear differential equations. The method was developed from the 1970s to the 1990s by George Adomian, chair of the Center for Applied Mathematics at the University of Georgia.[1] It is further extensible to stochastic systems by using the Ito integral.[2] The aim of this method is towards a unified theory for the solution of partial differential equations (PDE); an aim which has been superseded by the more general theory of the homotopy analysis method.[3] The crucial aspect of the method is employment of the "Adomian polynomials" which allow for solution convergence of the nonlinear portion of the equation, without simply linearizing the system. These polynomials mathematically generalize to a Maclaurin series about an arbitrary external parameter; which gives the solution method more flexibly than direct Taylor series expansion.[4]
Ordinary Differential Equations
Adomian method is well suited to solve Cauchy problems, an important class of problems which include initial conditions problems.
Application to a first order nonlinear system
An example of initial condition problem for an Ordinary Differential Equation is the following:
To solve the problem, the highest degree differential operator (written here as L) is put on the left side, in the following way:
with and . Now the solution is assumed to be an infinite series of contributions:
Replacing in the previous expression, we obtain:
Now we identify with some explicit expression on the right, and , with some expression on the right containing terms of lower order than . For instance:
In this way, any contribution can be explicitly calculated at any order. If we settle for the four first terms, the approximant is the following:
Application to Blasius equation
A second example, with more complex boundary conditions is the Blasius Equation for a flow in a boundary layer:
With the following conditions at the boundaries:
Linear and non-linear operators are now called and , respectively. Then, the expression becomes:
and the solution may be expressed, in this case, in the following simple way:
where:
If:
and:
Adomian’s polynomials to linearize the non-linear term can be obtained systematically by using the following rule:
, where:
Boundary conditions must be applied, in general, at the end of each approximation. In this case, the integration constants must be grouped into three final independent constants. However, in our example, the three constants appear grouped from the beginning in the form shown in the formal solution above. After applying the two first boundary conditions we obtain the so-called Blasius series:
To obtain we have to apply boundary conditions at , which may be done by writing the series as a Padé approximant:
where L=M. The limit at of this expression is .
If we choose , M linear equations for the b coefficients are obtained:
Then, we obtain the a coefficients by means of the following sequence:
In our example:
Which when becomes:
With the limit:
Which is approximately equal to 1 (the boundary condition (3))with an accuracy of 4/1000.
Partial Differential Equations
Application to a rectangular system with nonlinearity
One of the most frequent problems in physical sciences is to obtain the solution of a (linear or nonlinear) partial differential equation which satisfies a set of functional values on a rectangular boundary. For instance, let us consider the following problem:
with the following boundary conditions defined on a rectangle:
and
and
This kind of partial differential equation appears frequently coupled with others in science and engineering. For instance, in the incompressible fluid flow problem, the Navier–Stokes equations must be solved in parallel with a Poisson equation for the pressure.
Let us use the following notation for the problem (1):
where are double derivate operators and is a non-linear operator.
The formal solution of (2) is:
Expanding now u as a set of contributions to the solution we have:
By substitution in (3) and making a one-to-one correspondence between the contributions on the left side and the terms on the right side we obtain the following iterative scheme:
where the couple is the solution of the following system of equations:
is the n-order approximant to the solution and has been consistently expanded in Adomian polynomials:
,
where and in the example (1).
Here are products (or sum of products) of &\nu; components of u whose subscripts sum up to n, divided by the factorial of the number of repeated subscripts. It is only a thumb-rule to order systematically the decomposition to be sure that all the combinations appearing are utilized sooner or later.
The is equal to the sum of a generalized Taylor series about (Adomian, 1994).
For the example (1) the Adomian polynomials are:
Other possible choices are also possible for the expression of .
Series solutions
Cherruault (1989) established that the series terms obtained by Adomian's method approach zero as if m is the order of the highest linear differential operator and that .[5]
With this method the solution can be found by systematically integrating along any of the two directions: in direction X we would use expression (3); in the alternative Y direction we would use the following expression:
where: is obtained from the boundary conditions at and :
If we call x-partial solution and y-partial solution to the two respective solutions so obtained, one of the most interesting consequences of the method is that the x-partial solution uses only the two boundary conditions (1-a) and the y-partial solution uses only the conditions (1-b).
Thus, one of the two sets of boundary functions or is redundant, and this implies that a partial differential equation with boundary conditions on a rectangle cannot have arbitrary boundary conditions on the borders, since the conditions at , must be consistent with those imposed at and .
An example to clarify this point is the solution of the Poisson problem with the following boundary conditions:
By using Adomian's method and a symbolic processor (as Mathematica or Maple) it is easy to obtain the third order approximant to the solution. This approximant has an error lower than $5 \times 10^{-16}$ in any point, as it can be proved by substitution in the initial problem and by displaying the absolute value of the residual obtained as a function of (x, y) (García-Olivares 2003-a).
The solution at and is given by specific functions that in this case are:
and respectively.
If a (double) integration is now performed in the Y direction using these two boundary functions the same solution will be obtained, which satisfy and and cannot satisfy any other condition on these borders.
Some people react with incredulity to these results. Is it not a well established fact that any elliptic equation has one and only one solution for any functional conditions in the four sides of a rectangle provided there is no discontinuity on the edges?
The cause of the misconception is that scientists and engineers normally think in a boundary condition in terms of weak convergence in a Hilbert space (the distance to the boundary function is small enough to practical purposes). In contrast, Cauchy problems impose a point-to-point convergence to a given boundary function and to all its derivatives (and this is a quite strong condition!).
For the first ones, a function satisfies a boundary condition when the area (or another functional distance) between it and the true function imposed in the boundary is so small as desired; for the second ones, however, the function must tend to the true function imposed in any and every point of the interval.
The commented Poisson problem does not have a solution for any functional boundary conditions ; however, given it is always possible to find boundary functions so close to as desired (in the weak convergence meaning) for which the problem has solution. This property makes it possible to solve Poisson's and many other problems with arbitrary boundary conditions but never for analytic functions exactly specified on the boundaries.
The reader can convince himself (herself) of the high sensitivity of PDE solutions to small changes in the boundary conditions by solving this problem integrating along the X-direction, with boundary functions slightly different even tough visually not distinguishable. For instance, the solution with the boundary conditions:
at x=0 and x=0.5, and the solution with the boundary conditions:
at x=0 and x=0.5, produce lateral functions with different sign convexity even tough both functions are visually not distinguishable.
Solutions of elliptic problems and other partial differential equations are highly sensitive to small changes in the boundary function imposed when only two sides are used. And this sensitivity is not easily compatible with models that are supposed to represent real systems, which are described by means of measurements containing experimental errors and are normally expressed as initial-boundary value problems in a Hilbert space.
Improvements to the decomposition method
At least three methods have been reported [6] [7] [8] to obtain the boundary functions that are compatible with any lateral set of conditions imposed. This makes it possible to find the analytical solution of any PDE boundary problem on a closed rectangle with the required accuracy, so allowing to solve a wide range of problems that the standard Adomian's method was not able to address.
The first one perturbs the two boundary functions imposed at and (condition 1-a) with a N-order polynomial in y: in such a way that: , where the norm of the two perturbation functions are smaller than the accuracy needed at the boundaries. These depend on a set of polynomial coefficients . Then, the Adomian method is applied and functions are obtained at the four boundaries which depend on the set of . Finally, a boundary function is defined as the sum of these four functions, and the distance between and the real boundary functions ((1-a) and (1-b)) is minimized. The problem has been reduced, in this way, to the global minimization of the function which has a global minimum for some combination of the parameters . This minimum may be found by means of a genetic algorithm or by using some other optimization method, as the one proposed by Cherruault (1999).[9]
A second method to obtain analytic approximants of initial-boundary problems is to combine Adomian decomposition with spectral methods.[6]
Finally, the third method proposed by García-Olivares is based on imposing analytic solutions at the four boundaries, but modifying the original differential operator in such a way that it is different to the original one only in a narrow region close to the boundaries, and it forces the solution to satisfy exactly analytic conditions at the four boundaries.[8]
References
- ^ Adomian, G. (1994). Solving Frontier problems of Physics: The decomposition method. Kluwer Academic Publishers.
- ^ Adomian, G. (1986). Nonlinear Stochastic Operator Equations. Kluwer Academic Publishers. ISBN 0-12-044375-9. [1]
- ^ Liao, S.J. (2012), Homotopy Analysis Method in Nonlinear Differential Equation, Berlin & Beijing: Springer & Higher Education Press, ISBN 978-3642251313 [2]
- ^ Wazwaz, Abdul-Majid (2009). Partial Differential Equations and Solitary Waves Theory. Higher Education Press. p. 15. ISBN 90-5809-369-7.
- ^ Cherruault, Y. (1989), "Convergence of Adomian's Method", Kybernetes, 18, Bingley, U.K.: Emerald: 31–38
- ^ a b García-Olivares, A. (2002), "Analytical approximants of time-dependent partial differential equations with tau methods", Mathematics and Computers in Simulation, 61, Amsterdam, Netherlands: Elsevier: 35–45 [3]
- ^ García-Olivares, A. (2003), "Analytic solution of partial differential equations with Adomian's decomposition", Kybernetes, 32, Bingley, U.K.: Emerald: 354–368 [4]
- ^ a b García-Olivares, A. (2003), "Analytical solution of nonlinear partial differential equations of physics", Kybernetes, 32, Bingley, U.K.: Emerald: 548–560 [DOI: 10.1108/03684920310463939] [5]
- ^ Cherruault, Y. (1999). Optimization, Méthodes locales et globales. Presses Universitaires de France. ISBN 2-13-049910-4.