Sidi's generalized secant method

Last updated

Sidi's generalized secant method is a root-finding algorithm, that is, a numerical method for solving equations of the form . The method was published by Avram Sidi. [1]

Contents

The method is a generalization of the secant method. Like the secant method, it is an iterative method which requires one evaluation of in each iteration and no derivatives of . The method can converge much faster though, with an order which approaches 2 provided that satisfies the regularity conditions described below.

Algorithm

We call the root of , that is, . Sidi's method is an iterative method which generates a sequence of approximations of . Starting with k + 1 initial approximations , the approximation is calculated in the first iteration, the approximation is calculated in the second iteration, etc. Each iteration takes as input the last k + 1 approximations and the value of at those approximations. Hence the nth iteration takes as input the approximations and the values .

The number k must be 1 or larger: k = 1, 2, 3, .... It remains fixed during the execution of the algorithm. In order to obtain the starting approximations one could carry out a few initializing iterations with a lower value of k.

The approximation is calculated as follows in the nth iteration. A polynomial of interpolation of degree k is fitted to the k + 1 points . With this polynomial, the next approximation of is calculated as

 

 

 

 

(1)

with the derivative of at . Having calculated one calculates and the algorithm can continue with the (n + 1)th iteration. Clearly, this method requires the function to be evaluated only once per iteration; it requires no derivatives of .

The iterative cycle is stopped if an appropriate stop-criterion is met. Typically the criterion is that the last calculated approximation is close enough to the sought-after root .

To execute the algorithm effectively, Sidi's method calculates the interpolating polynomial in its Newton form.

Convergence

Sidi showed that if the function is (k + 1)-times continuously differentiable in an open interval containing (that is, ), is a simple root of (that is, ) and the initial approximations are chosen close enough to , then the sequence converges to , meaning that the following limit holds: .

Sidi furthermore showed that

and that the sequence converges to of order , i.e.

The order of convergence is the only positive root of the polynomial

We have e.g. ≈ 1.6180, ≈ 1.8393 and ≈ 1.9276. The order approaches 2 from below if k becomes large: [2] [3]

Sidi's method reduces to the secant method if we take k = 1. In this case the polynomial is the linear approximation of around which is used in the nth iteration of the secant method.

We can expect that the larger we choose k, the better is an approximation of around . Also, the better is an approximation of around . If we replace with in ( 1 ) we obtain that the next approximation in each iteration is calculated as

 

 

 

 

(2)

This is the Newton–Raphson method. It starts off with a single approximation so we can take k = 0 in ( 2 ). It does not require an interpolating polynomial but instead one has to evaluate the derivative in each iteration. Depending on the nature of this may not be possible or practical.

Once the interpolating polynomial has been calculated, one can also calculate the next approximation as a solution of instead of using ( 1 ). For k = 1 these two methods are identical: it is the secant method. For k = 2 this method is known as Muller's method. [3] For k = 3 this approach involves finding the roots of a cubic function, which is unattractively complicated. This problem becomes worse for even larger values of k. An additional complication is that the equation will in general have multiple solutions and a prescription has to be given which of these solutions is the next approximation . Muller does this for the case k = 2 but no such prescriptions appear to exist for k > 2.

Related Research Articles

In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots of a real-valued function. The most basic version starts with a single-variable function f defined for a real variable x, the function's derivative f ′, and an initial guess x0 for a root of f. If the function satisfies sufficient assumptions and the initial guess is close, then

Quantum harmonic oscillator Important, well-understood quantum mechanical model

The quantum harmonic oscillator is the quantum-mechanical analog of the classical harmonic oscillator. Because an arbitrary smooth potential can usually be approximated as a harmonic potential at the vicinity of a stable equilibrium point, it is one of the most important model systems in quantum mechanics. Furthermore, it is one of the few quantum-mechanical systems for which an exact, analytical solution is known.

Julia set Fractal sets in complex dynamics of mathematics

In the context of complex dynamics, a topic of mathematics, the Julia set and the Fatou set are two complementary sets defined from a function. Informally, the Fatou set of the function consists of values with the property that all nearby values behave similarly under repeated iteration of the function, and the Julia set consists of values such that an arbitrarily small perturbation can cause drastic changes in the sequence of iterated function values. Thus the behavior of the function on the Fatou set is "regular", while on the Julia set its behavior is "chaotic".

In mathematics, the Hermite polynomials are a classical orthogonal polynomial sequence.

Beta distribution Probability distribution

In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] parameterized by two positive shape parameters, denoted by α and β, that appear as exponents of the random variable and control the shape of the distribution. The generalization to multiple variables is called a Dirichlet distribution.

In quantum mechanics, the particle in a one-dimensional lattice is a problem that occurs in the model of a periodic crystal lattice. The potential is caused by ions in the periodic structure of the crystal creating an electromagnetic field so electrons are subject to a regular potential inside the lattice. It is a generalization of the free electron model, which assumes zero potential inside the lattice.

In physics, the S-matrix or scattering matrix relates the initial state and the final state of a physical system undergoing a scattering process. It is used in quantum mechanics, scattering theory and quantum field theory (QFT).

The adiabatic theorem is a concept in quantum mechanics. Its original form, due to Max Born and Vladimir Fock (1928), was stated as follows:

LSZ reduction formula

In quantum field theory, the LSZ reduction formula is a method to calculate S-matrix elements from the time-ordered correlation functions of a quantum field theory. It is a step of the path that starts from the Lagrangian of some quantum field theory and leads to prediction of measurable quantities. It is named after the three German physicists Harry Lehmann, Kurt Symanzik and Wolfhart Zimmermann.

Muller's method is a root-finding algorithm, a numerical method for solving equations of the form f(x) = 0. It was first presented by David E. Muller in 1956.

In physics, Fujikawa's method is a way of deriving the chiral anomaly in quantum field theory. It uses the correspondence between functional determinants and the partition function, effectively making use of the Atiyah–Singer index theorem.

Harmonic balance is a method originally introduced by Michel Nakhla. It is used to calculate the steady-state response of nonlinear differential equations, and is mostly applied to nonlinear electrical circuits . It is a frequency domain method for calculating the steady state, as opposed to the various time-domain steady state methods. The name "harmonic balance" is descriptive of the method, which starts with Kirchhoff's Current Law written in the frequency domain and a chosen number of harmonics. A sinusoidal signal applied to a nonlinear component in a system will generate harmonics of the fundamental frequency. Effectively the method assumes the solution can be represented by a linear combination of sinusoids, then balances current and voltage sinusoids to satisfy Kirchhoff's law. The method is commonly used to simulate circuits which include nonlinear elements, and is most applicable to systems with feedback in which limit cycles occur.

The finite potential well is a concept from quantum mechanics. It is an extension of the infinite potential well, in which a particle is confined to a "box", but one which has finite potential "walls". Unlike the infinite potential well, there is a probability associated with the particle being found outside the box. The quantum mechanical interpretation is unlike the classical interpretation, where if the total energy of the particle is less than the potential energy barrier of the walls it cannot be found outside the box. In the quantum interpretation, there is a non-zero probability of the particle being outside the box even when the energy of the particle is less than the potential energy barrier of the walls.

In mathematics, in the area of complex analysis, Nachbin's theorem is commonly used to establish a bound on the growth rates for an analytic function. This article provides a brief review of growth rates, including the idea of a function of exponential type. Classification of growth rates based on type help provide a finer tool than big O or Landau notation, since a number of theorems about the analytic structure of the bounded function and its integral transforms can be stated. In particular, Nachbin's theorem may be used to give the domain of convergence of the generalized Borel transform, given below.

Subgradient methods are iterative methods for solving convex minimization problems. Originally developed by Naum Z. Shor and others in the 1960s and 1970s, subgradient methods are convergent when applied even to a non-differentiable objective function. When the objective function is differentiable, sub-gradient methods for unconstrained problems use the same search direction as the method of steepest descent.

This is a glossary for the terminology often encountered in undergraduate quantum mechanics courses.

Chandrasekhars <i>H</i>-function

In atmospheric radiation, Chandrasekhar's H-function appears as the solutions of problems involving scattering, introduced by the Indian American astrophysicist Subrahmanyan Chandrasekhar. The Chandrasekhar's H-function defined in the interval , satisfies the following nonlinear integral equation

In mathematics, the Fuchs relation is a relation between the starting exponents of formal series solutions of certain linear differential equations, so called Fuchsian equations. It is named after Lazarus Immanuel Fuchs.

The Fuchsian theory of linear differential equations, which is named after Lazarus Immanuel Fuchs, provides a characterization of various types of singularities and the relations among them.

Tau functions are an important ingredient in the modern theory of integrable systems, and have numerous applications in a variety of other domains. They were originally introduced by Ryogo Hirota in his direct method approach to soliton equations, based on expressing them in an equivalent bilinear form. The term Tau function, or -function, was first used systematically by Mikio Sato and his students in the specific context of the Kadomtsev–Petviashvili equation, and related integrable hierarchies. It is a central ingredient in the theory of solitons. Tau functions also appear as matrix model partition functions in the spectral theory of Random Matrices, and may also serve as generating functions, in the sense of combinatorics and enumerative geometry, especially in relation to moduli spaces of Riemann surfaces, and enumeration of branched coverings, or so-called Hurwitz numbers.

References

  1. Sidi, Avram, "Generalization Of The Secant Method For Nonlinear Equations", Applied Mathematics E-notes 8 (2008), 115–123, http://www.math.nthu.edu.tw/~amen/2008/070227-1.pdf
  2. Traub, J.F., "Iterative Methods for the Solution of Equations", Prentice Hall, Englewood Cliffs, N.J. (1964)
  3. 1 2 Muller, David E., "A Method for Solving Algebraic Equations Using an Automatic Computer", Mathematical Tables and Other Aids to Computation 10 (1956), 208–215