Steffensen's method

In numerical analysis, Steffensen's method is a root-finding technique named after Johan Frederik Steffensen which is similar to Newton's method. Steffensen's method also achieves quadratic convergence, but without using derivatives as Newton's method does.

Simple description

The simplest form of the formula for Steffensen's method occurs when it is used to find the zeros, or roots, of a function ; that is: to find the value that satisfies  . Near the solution , the function is supposed to approximately satisfy this condition makes adequate as a correction-function for for finding its own solution, although it is not required to work efficiently. For some functions, Steffensen's method can work even if this condition is not met, but in such a case, the starting value must be very close to the actual solution , and convergence to the solution may be slow.

Given an adequate starting value , a sequence of values can be generated using the formula below. When it works, each value in the sequence is much closer to the solution than the prior value. The value from the current step generates the value for the next step, via this formula:[1]

for n = 0, 1, 2, 3, ... , where the slope function is a composite of the original function given by the following formula:

or equivalently

where .

The function is the average value for the slope of the function between the last sequence point and the auxiliary point , with the step . It is also called the first-order divided difference of between those two points.

It is only for the purpose of finding for this auxiliary point that the value of the function must be an adequate correction to get closer to its own solution, and for that reason fulfill the requirement that . For all other parts of the calculation, Steffensen's method only requires the function to be continuous and to actually have a nearby solution.[1] Several modest modifications of the step in the slope calculation exist to accommodate functions that do not quite meet the requirement.

Advantages and drawbacks

The main advantage of Steffensen's method is that it has quadratic convergence[1] like Newton's method that is, both methods find roots to an equation just as ‘quickly’. In this case quickly means that for both methods, the number of correct digits in the answer doubles with each step. But the formula for Newton's method requires evaluation of the function's derivative as well as the function , while Steffensen's method only requires itself. This is important when the derivative is not easily or efficiently available.

The price for the quick convergence is the double function evaluation: Both and must be calculated, which might be time-consuming if is a complicated function. For comparison, the secant method needs only one function evaluation per step. The secant method increases the number of correct digits by "only" a factor of roughly 1.6 per step, but one can do twice as many steps of the secant method within a given time. Since the secant method can carry out twice as many steps in the same time as Steffensen's method,[lower-alpha 1] when both algorithms succeed, the secant method converges faster than Steffensen's method in actual practice: The secant method achieves a factor of about (1.6)2  2.6 times as many digits for every two steps (two function evaluations), compared to Steffensen's factor of 2 for every one step (two function evaluations).

Similar to most other iterative root-finding algorithms, the crucial weakness in Steffensen's method is the choice of the starting value  . If the value of is not ‘close enough’ to the actual solution  , the method may fail and the sequence of values may either flip-flop between two extremes, or diverge to infinity (possibly both!).

Derivation using Aitken's delta-squared process

The version of Steffensen's method implemented in the MATLAB code shown below can be found using the Aitken's delta-squared process for accelerating convergence of a sequence. To compare the following formulae to the formulae in the section above, notice that  . This method assumes starting with a linearly convergent sequence and increases the rate of convergence of that sequence. If the signs of agree and is ‘sufficiently close’ to the desired limit of the sequence , we can assume the following:

then

so

and hence

 .


Solving for the desired limit of the sequence gives:



which results in the more rapidly convergent sequence:

Implementation in Matlab

Here is the source for an implementation of Steffensen's Method in MATLAB.

function Steffensen(f,p0,tol)
% This function takes as inputs: a fixed point iteration function, f, 
% and initial guess to the fixed point, p0, and a tolerance, tol.
% The fixed point iteration function is assumed to be input as an
% inline function. 
% This function will calculate and return the fixed point, p, 
% that makes the expression f(x) = p true to within the desired 
% tolerance, tol.

format compact % This shortens the output.
format long    % This prints more decimal places.

for i=1:1000   % get ready to do a large, but finite, number of iterations.
               % This is so that if the method fails to converge, we won't
               % be stuck in an infinite loop.
    p1=f(p0);  % calculate the next two guesses for the fixed point.
    p2=f(p1);
    p=p0-(p1-p0)^2/(p2-2*p1+p0) % use Aitken's delta squared method to
                                % find a better approximation to p0.
    if abs(p-p0)<tol  % test to see if we are within tolerance.
        break         % if we are, stop the iterations, we have our answer.
    end
    p0=p;              % update p0 for the next iteration.
end
if abs(p-p0)>tol       % If we fail to meet the tolerance, we output a
                       % message of failure.
    'failed to converge in 1000 iterations.'
end

Implementation in Python

Here is the source for an implementation of Steffensen's Method in Python.

def g(f, x: float):
    """First-order divided difference function.

    Arguments:
        f(callable): Function input to g
        x(float): Point at which to evaluate g
    """
    return lambda x: f(x + f(x)) / f(x) - 1

def steff(f, x: float):
    """Steffenson algorithm for finding roots.

    This recursive generator yields the x_n+1 value first then, when the generator iterates,
    it yields x_n + 2 from the next level of recursion.

    Arguments:
        f(callable): Function whose root we are searching for
        x(float): Starting value upon first call, each level n that the function recurses x is x_n
    """

    if g(f, x)(x) != 0:
        yield x - f(x) / g(f, x)(x)  # First give x_n + 1
        yield from steff(f, x - f(x) / g(f, x)(x))  # Then give new iterator

Generalization

Steffensen's method can also be used to find an input for a different kind of function that produces output the same as its input: for the special value  . Solutions like are called fixed points. Many such functions can be used to find their own solutions by repeatedly recycling the result back as input, but the rate of convergence can be slow, or the function can fail to converge at all, depending on the individual function. Steffensen's method accelerates this convergence, to make it quadratic.

This method for finding fixed points of a real-valued function has been generalised for functions on a Banach space  . The generalised method assumes that a family of bounded linear operators associated with and can be found to satisfy the condition[2]

In the simple form given in the section above, the function simply takes in and produces real numbers. There, the function is a divided difference. In the generalized form here, the operator is the analogue of a divided difference for use in the Banach space. The operator is equivalent to a matrix whose entries are all functions of vector arguments and .

Steffensen's method is then very similar to the Newton's method, except that it uses the divided difference instead of the derivative  . It is thus defined by

for  , and where is the identity operator.

If the operator satisfies

for some constant  , then the method converges quadratically to a fixed point of if the initial approximation is ‘sufficiently close’ to the desired solution  , that satisfies  .

Notes

  1. Because requires the prior calculation of , the two evaluations must be done sequentially – the algorithm per se cannot be made faster by running the function evaluations in parallel. This is yet another disadvantage of Steffensen's method.
gollark: ?remind 666d https://discord.com/channels/346530916832903169/445375649511768074/707611563162337370
gollark: Oh, I see.
gollark: I don't know if the rules allow it but meh.
gollark: Move stuff so it's BABA IS MELT AND YOU at the bottom, and you should be left with BABA IS, and push the IS to beside the HOT then put the BABA beside that.
gollark: Or probably "WATER IS YOU".

References

  1. Dahlquist, Germund; Björck, Åke (1974). Numerical Methods. Translated by Anderson, Ned. Englewood Cliffs, NJ: Prentice Hall. pp. 230–231.
  2. Johnson, L.W.; Scholz, D.R. (June 1968). "On Steffensen's method". SIAM Journal on Numerical Analysis. 5 (2): 296–302. doi:10.1137/0705026. JSTOR 2949443.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.