Newton's method

But, in the absence of any intuition about where the zero might lie, a "guess and check" method might narrow the possibilities to a reasonably small interval by appealing to the intermediate value theorem.)

[2] The method first appeared roughly in Isaac Newton's work in De analysi per aequationes numero terminorum infinitas (written in 1669, published in 1711 by William Jones) and in De metodis fluxionum et serierum infinitarum (written in 1671, translated and published as Method of Fluxions in 1736 by John Colson).

He applied the method only to polynomials, starting with an initial root estimate and extracting a sequence of error corrections.

Newton applied this method to both numerical and algebraic problems, producing Taylor series in the latter case.

[3] The essence of Viète's own method can be found in the work of the Persian mathematician Sharaf al-Din al-Tusi.

[5] The Japanese mathematician Seki Kōwa used a form of Newton's method in the 1670s to solve single-variable equations, though the connection with calculus was missing.

In the same publication, Simpson also gives the generalization to systems of two equations and notes that Newton's method can be used for solving optimization problems by setting the gradient to zero.

For example, in some cases, if the first derivative is not well behaved in the neighborhood of a particular root, then it is possible that Newton's method will fail to converge no matter where the initialization is set.

On the other hand, if the multiplicity m of the root is not known, it is possible to estimate m after carrying out one or two iterations, and then use that value to increase the rate of convergence.

Applying Newton's method to find the root of g(x) recovers quadratic convergence in many cases although it generally involves the second derivative of f(x).

According to Taylor's theorem, any function f(x) which has a continuous second derivative can be represented by an expansion about a point that is close to a root of f(x).

That is, Taking the absolute value of both sides gives Equation (6) shows that the order of convergence is at least quadratic if the following conditions are satisfied: where M is given by

From geometrical principles, it can be seen that the Newton iteration xi starting at the left endpoint is monotonically increasing and convergent, necessarily to ζ.

[12] Joseph Fourier introduced a modification of Newton's method starting at the right endpoint: This sequence is monotonically decreasing and convergent.

If f is twice continuously differentiable, it can be proved using Taylor's theorem that showing that this difference in locations converges quadratically to zero.

[12] All of the above can be extended to systems of equations in multiple variables, although in that context the relevant concepts of monotonicity and concavity are more subtle to formulate.

[13] In the case of single equations in a single variable, the above monotonic convergence of Newton's method can also be generalized to replace concavity by positivity or negativity conditions on an arbitrary higher-order derivative of f. However, in this generalization, Newton's iteration is modified so as to be based on Taylor polynomials rather than the tangent line.

By performing this iteration, it is possible to evaluate a square root to any desired accuracy by only using the basic arithmetic operations.

The following three tables show examples of the result of this computation for finding the square root of 612, with the iteration initialized at the values of 1, 10, and −20.

Each row in a "xn" column is obtained by applying the preceding formula to the entry above it, for instance The correct digits are underlined.

This kind of subtle dependence on initialization is not uncommon; it is frequently studied in the complex plane in the form of the Newton fractal.

One may also use Newton's method to solve systems of k equations, which amounts to finding the (simultaneous) zeroes of k continuously differentiable functions

[29] In the 1950s, John Nash developed a version of the Newton's method to apply to the problem of constructing isometric embeddings of general Riemannian manifolds in Euclidean space.

The loss of derivatives problem, present in this context, made the standard Newton iteration inapplicable, since it could not be continued indefinitely (much less converge).

In the 1960s, Jürgen Moser showed that Nash's methods were flexible enough to apply to problems beyond isometric embedding, particularly in celestial mechanics.

Since then, a number of mathematicians, including Mikhael Gromov and Richard Hamilton, have found generalized abstract versions of the Nash–Moser theory.

Because of the more stable behavior of addition and multiplication in the p-adic numbers compared to the real numbers (specifically, the unit ball in the p-adics is a ring), convergence in Hensel's lemma can be guaranteed under much simpler hypotheses than in the classical Newton's method on the real line.

Also, this may detect cases where Newton's method converges theoretically but diverges numerically because of an insufficient floating-point precision (this is typically the case for polynomials of large degree, where a very small change of the variable may change dramatically the value of the function; see Wilkinson's polynomial).

[citation needed] The following is an example of a possible implementation of Newton's method in the Python (version 3.x) programming language for finding a root of a function f which has derivative f_prime.

We will check during the computation whether the denominator (yprime) becomes too small (smaller than epsilon), which would be the case if f′(xn) ≈ 0, since otherwise a large amount of error could be introduced.

An illustration of Newton's method.
Illustration of Newton's method
x n +1 is a better approximation than x n for the root x of the function f (blue curve)
Illustration of Newton's method
Iteration typically improves the approximation
The tangent lines of x 3 − 2 x + 2 at 0 and 1 intersect the x -axis at 1 and 0 respectively, illustrating why Newton's method oscillates between these values for some starting points.
Basins of attraction for x 5 − 1 = 0 ; darker means more iterations to converge.
Graph of a strictly concave quadratic function with unique maximum.
Optimization computes maxima and minima.