Sidi's generalized secant method is a root-finding algorithm, that is, a numerical method for solving equations of the form . The method was published by Avram Sidi. [1]
The method is a generalization of the secant method. Like the secant method, it is an iterative method which requires one evaluation of in each iteration and no derivatives of . The method can converge much faster though, with an order which approaches 2 provided that satisfies the regularity conditions described below.
We call the root of , that is, . Sidi's method is an iterative method which generates a sequence of approximations of . Starting with k + 1 initial approximations , the approximation is calculated in the first iteration, the approximation is calculated in the second iteration, etc. Each iteration takes as input the last k + 1 approximations and the value of at those approximations. Hence the nth iteration takes as input the approximations and the values .
The number k must be 1 or larger: k = 1, 2, 3, .... It remains fixed during the execution of the algorithm. In order to obtain the starting approximations one could carry out a few initializing iterations with a lower value of k.
The approximation is calculated as follows in the nth iteration. A polynomial of interpolation of degree k is fitted to the k + 1 points . With this polynomial, the next approximation of is calculated as
(1) |
with the derivative of at . Having calculated one calculates and the algorithm can continue with the (n + 1)th iteration. Clearly, this method requires the function to be evaluated only once per iteration; it requires no derivatives of .
The iterative cycle is stopped if an appropriate stopping criterion is met. Typically the criterion is that the last calculated approximation is close enough to the sought-after root .
To execute the algorithm effectively, Sidi's method calculates the interpolating polynomial in its Newton form.
Sidi showed that if the function is (k + 1)-times continuously differentiable in an open interval containing (that is, ), is a simple root of (that is, ) and the initial approximations are chosen close enough to , then the sequence converges to , meaning that the following limit holds: .
Sidi furthermore showed that
and that the sequence converges to of order , i.e.
The order of convergence is the only positive root of the polynomial
We have e.g. ≈ 1.6180, ≈ 1.8393 and ≈ 1.9276. The order approaches 2 from below if k becomes large: [2] [3]
Sidi's method reduces to the secant method if we take k = 1. In this case the polynomial is the linear approximation of around which is used in the nth iteration of the secant method.
We can expect that the larger we choose k, the better is an approximation of around . Also, the better is an approximation of around . If we replace with in ( 1 ) we obtain that the next approximation in each iteration is calculated as
(2) |
This is the Newton–Raphson method. It starts off with a single approximation so we can take k = 0 in ( 2 ). It does not require an interpolating polynomial but instead one has to evaluate the derivative in each iteration. Depending on the nature of this may not be possible or practical.
Once the interpolating polynomial has been calculated, one can also calculate the next approximation as a solution of instead of using ( 1 ). For k = 1 these two methods are identical: it is the secant method. For k = 2 this method is known as Muller's method. [3] For k = 3 this approach involves finding the roots of a cubic function, which is unattractively complicated. This problem becomes worse for even larger values of k. An additional complication is that the equation will in general have multiple solutions and a prescription has to be given which of these solutions is the next approximation . Muller does this for the case k = 2 but no such prescriptions appear to exist for k > 2.
The quantum harmonic oscillator is the quantum-mechanical analog of the classical harmonic oscillator. Because an arbitrary smooth potential can usually be approximated as a harmonic potential at the vicinity of a stable equilibrium point, it is one of the most important model systems in quantum mechanics. Furthermore, it is one of the few quantum-mechanical systems for which an exact, analytical solution is known.
In the context of complex dynamics, a branch of mathematics, the Julia set and the Fatou set are two complementary sets defined from a function. Informally, the Fatou set of the function consists of values with the property that all nearby values behave similarly under repeated iteration of the function, and the Julia set consists of values such that an arbitrarily small perturbation can cause drastic changes in the sequence of iterated function values. Thus the behavior of the function on the Fatou set is "regular", while on the Julia set its behavior is "chaotic".
In mathematics, the Hermite polynomials are a classical orthogonal polynomial sequence.
In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or in terms of two positive parameters, denoted by alpha (α) and beta (β), that appear as exponents of the variable and its complement to 1, respectively, and control the shape of the distribution.
In probability theory and statistics, the gamma distribution is a versatile two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:
In quantum mechanics, the particle in a one-dimensional lattice is a problem that occurs in the model of a periodic crystal lattice. The potential is caused by ions in the periodic structure of the crystal creating an electromagnetic field so electrons are subject to a regular potential inside the lattice. It is a generalization of the free electron model, which assumes zero potential inside the lattice.
In physics, the S-matrix or scattering matrix relates the initial state and the final state of a physical system undergoing a scattering process. It is used in quantum mechanics, scattering theory and quantum field theory (QFT).
The Gram–Charlier A series, and the Edgeworth series are series that approximate a probability distribution in terms of its cumulants. The series are the same; but, the arrangement of terms differ. The key idea of these expansions is to write the characteristic function of the distribution whose probability density function f is to be approximated in terms of the characteristic function of a distribution with known and suitable properties, and to recover f through the inverse Fourier transform.
In quantum field theory, the Lehmann–Symanzik–Zimmermann (LSZ) reduction formula is a method to calculate S-matrix elements from the time-ordered correlation functions of a quantum field theory. It is a step of the path that starts from the Lagrangian of some quantum field theory and leads to prediction of measurable quantities. It is named after the three German physicists Harry Lehmann, Kurt Symanzik and Wolfhart Zimmermann.
Muller's method is a root-finding algorithm, a numerical method for solving equations of the form f(x) = 0. It was first presented by David E. Muller in 1956.
Harmonic balance is a method used to calculate the steady-state response of nonlinear differential equations, and is mostly applied to nonlinear electrical circuits. It is a frequency domain method for calculating the steady state, as opposed to the various time-domain steady-state methods. The name "harmonic balance" is descriptive of the method, which starts with Kirchhoff's Current Law written in the frequency domain and a chosen number of harmonics. A sinusoidal signal applied to a nonlinear component in a system will generate harmonics of the fundamental frequency. Effectively the method assumes a linear combination of sinusoids can represent the solution, then balances current and voltage sinusoids to satisfy Kirchhoff's law. The method is commonly used to simulate circuits which include nonlinear elements, and is most applicable to systems with feedback in which limit cycles occur.
In mathematics, in the area of complex analysis, Nachbin's theorem is commonly used to establish a bound on the growth rates for an analytic function. This article provides a brief review of growth rates, including the idea of a function of exponential type. Classification of growth rates based on type help provide a finer tool than big O or Landau notation, since a number of theorems about the analytic structure of the bounded function and its integral transforms can be stated. In particular, Nachbin's theorem may be used to give the domain of convergence of the generalized Borel transform, given below.
The Wiener–Hopf method is a mathematical technique widely used in applied mathematics. It was initially developed by Norbert Wiener and Eberhard Hopf as a method to solve systems of integral equations, but has found wider use in solving two-dimensional partial differential equations with mixed boundary conditions on the same boundary. In general, the method works by exploiting the complex-analytical properties of transformed functions. Typically, the standard Fourier transform is used, but examples exist using other transforms, such as the Mellin transform.
In mathematics, the spectral theory of ordinary differential equations is the part of spectral theory concerned with the determination of the spectrum and eigenfunction expansion associated with a linear ordinary differential equation. In his dissertation, Hermann Weyl generalized the classical Sturm–Liouville theory on a finite closed interval to second order differential operators with singularities at the endpoints of the interval, possibly semi-infinite or infinite. Unlike the classical case, the spectrum may no longer consist of just a countable set of eigenvalues, but may also contain a continuous part. In this case the eigenfunction expansion involves an integral over the continuous part with respect to a spectral measure, given by the Titchmarsh–Kodaira formula. The theory was put in its final simplified form for singular differential equations of even degree by Kodaira and others, using von Neumann's spectral theorem. It has had important applications in quantum mechanics, operator theory and harmonic analysis on semisimple Lie groups.
This is a glossary for the terminology often encountered in undergraduate quantum mechanics courses.
Input-to-state stability (ISS) is a stability notion widely used to study stability of nonlinear control systems with external inputs. Roughly speaking, a control system is ISS if it is globally asymptotically stable in the absence of external inputs and if its trajectories are bounded by a function of the size of the input for all sufficiently large times. The importance of ISS is due to the fact that the concept has bridged the gap between input–output and state-space methods, widely used within the control systems community.
In atmospheric radiation, Chandrasekhar's H-function appears as the solutions of problems involving scattering, introduced by the Indian American astrophysicist Subrahmanyan Chandrasekhar. The Chandrasekhar's H-function defined in the interval , satisfies the following nonlinear integral equation
In mathematics, the Fuchs relation is a relation between the starting exponents of formal series solutions of certain linear differential equations, so called Fuchsian equations. It is named after Lazarus Immanuel Fuchs.
The Fuchsian theory of linear differential equations, which is named after Lazarus Immanuel Fuchs, provides a characterization of various types of singularities and the relations among them.
Tau functions are an important ingredient in the modern mathematical theory of integrable systems, and have numerous applications in a variety of other domains. They were originally introduced by Ryogo Hirota in his direct method approach to soliton equations, based on expressing them in an equivalent bilinear form.