Shape Operator

Counting operations instead of degree in polynomial approximation

27 Mar 2026

A famous problem in approximation theory is to approximate the function \left|x\right| as accurately as possible on the interval -1 \le x \le 1, measured by maximum absolute error, using either polynomials or rational functions of fixed degree.

Perhaps it seems silly to approximate a function as simple as this, but it serves as a useful test problem for probing how well non-smooth functions can be approximated by polynomial or rational functions.

I first read about this problem in Trefethen’s book Approximation Theory and Approximation Practice, which points out the significant gap in convergence behavior between polynomials and rationalsSee also Trefethen, Nakatsukasa, and Weideman, “Exponential node clustering at singularities for rational approximation, quadrature, and PDEs,” 2020 for some more recent discussion of rational approximation problems like this.:

Theorem 25.1. Approximation of |x| on [−1, 1]. The errors in best polynomial and rational approximation of |x| on [−1, 1] satisfy as n → ∞
E_{n0}(|x|) \sim \beta/n, \quad \beta = 0.2801\ldots
and
E_{nn}(|x|) \sim 8e^{-\pi\sqrt{n}}

Polynomials can converge only linearly in their degree here, which we will call the “Bernstein bound,” and rationals can converge root-exponentially in their degree, which we will call the “Newman bound.” (The notation E refers to the minimax error — the smallest achievable maximum absolute error over the interval.)

The previous two posts explored repeated composition of low-degree polynomials, which can generate polynomials of exponentially high degree in the number of operations. What happens when we apply this to the problem above?

In particular, consider Newton’s iteration for the inverse square root,

y_{n+1} = y_n\left(\frac{3}{2}-\frac{x}{2}y_n^2\right),

which can be derived by applying Newton’s method to the function f(y)=1/y^2-x. This iteration evaluates a polynomial of degree \left(3^n-1\right)/2 in x using 3n multiplications and n additions. To simplify a bit, we’ll consider a multiply followed by an addition as a single operation and say each iteration requires 3 “operations.” The absolute value follows from the inverse square root through |x|=x^2\cdot(1/\sqrt{x^2}).

Here is a plot showing the absolute error in this approximation, starting from y_0 = 1, for 10, 20, 30, 40, and 50 iterations:

Graph of absolute error when using Newton's inverse square root iteration to approximate |x| for 10, 20, 30, 40, and 50 iterations.
x
abs. err.

The steady march of these curves on a log-log scale shows that as the number of iterations increases, the maximum error occurs at exponentially decreasing values of x and decreases exponentially with iterations.

The following table gives the maximum absolute error read from each of these curves.

iterations operations max. abs. error
10 30 4.7 × 10-3
20 60 8.1 × 10-5
30 90 1.4 × 10-6
40 120 2.5 × 10-8
50 150 4.3 × 10-10

Fitting this data to an exponential gives a very good fit with about 0.18 decimal digits of accuracy added per iteration (0.059 digits per operation).

This is exponential convergence in operation count — a qualitatively different rate than either the Bernstein or Newman bounds.

The rate constant, however, is somewhat low for practical purposes. The crossover points when comparing this Newton iteration against minimax polynomials and rationals of fixed degree on a per-operation basis are, respectively, 23 operations (max. abs. err. ≈ 1.2 × 10-2) and 217 operations (max. abs. err. ≈ 5 × 10-14). Newton iteration is a clear win over minimax polynomials of fixed degree, but only a win over minimax rationals of fixed degree if you need relatively high precision.

But this isn’t yet a fair comparison. The Newton iteration is not specifically optimized for this problem (i.e. for this approximation interval and error metric) at all, whereas the minimax polynomials and rationals have many optimized coefficients. What might be possible with optimized coefficients at each iteration?

The Bernstein bound still sets a limit here. After n iterations (3n operations), a Newton-like method with optimized coefficients would still evaluate a polynomial of degree \left(3^n-1\right)/2. Using this degree in the Bernstein bound implies exponential convergence with a maximum rate constant of log10(3) ≈ 0.48 digits per iteration (0.16 digits per operation). Perhaps surprisingly, the unoptimized Newton iteration is already within a factor of about 2.7 of this hypothetical optimum rate. If this rate were achievable, it would move the crossover points with minimax polynomials and rationals of fixed degree down to 8 operations (max. abs. err. ≈ 4 × 10-2) and 71 operations (max. abs. err. ≈ 3 × 10-12) respectively.

Further improvement may be possible by considering iterated rational maps, which can also evaluate rationals of exponentially high degree in the number of operations. Combining that observation with the Newman bound suggests that super-exponential convergence in operation count could be possible.

After some early experiments, I have not yet observed such super-exponential convergence, and I have doubts that it will in fact be possible. Consider, for example, Heron’s method for approximating square roots,

y_{n+1} = \frac{1}{2}\left(y_n+\frac{x}{y_n}\right),

which can be derived by applying Newton’s method to the function f(y)=y^2-x. This iteration evaluates a rational function of degree 2^n-1 (in both numerator and denominator) after n iterations, using 3 operations per iteration (counting division as one operation). Using this iteration to approximate the absolute value through |x|=\sqrt{x^2}, starting with y_0 = 1, the maximum absolute error always appears to occur at x=0. This makes the error particularly easy to analyze: it is 2^{-n} after n iterations. Exponential convergence, with rate constant 0.30 decimal digits per iteration (0.10 digits per operation). This is faster than the polynomial inverse square root iteration above, but not super-exponential.

How might this be improved with optimized coefficients per iteration? The Newman bound on degree, suggesting that super-exponential convergence in operation count may be possible, seems unlikely to be realizable to me, but I don’t yet know a lower bound than that.

I have suggested optimizing coefficients in polynomial and rational iterations. This is likely to be a challenging optimization problem because the maximum absolute error is a highly nonlinear function of these coefficients. However, I see some reason for optimism. Polynomial and rational iterations of the type I am describing are very similar to the layers of neural networks, and in fact, there has already been substantial research on polynomial neural networks. When compared to the neural networks that are being trained successfully for machine learning and artificial intelligence, the iterations I am describing are very small networks.

The central observation of this post is simple: the classical convergence rates for polynomial and rational approximation are stated in terms of degree, but degree is not the same as computational cost. When measured by operation count, iterated polynomial maps can achieve exponential convergence for problems where fixed-degree polynomials converge only linearly and fixed-degree rationals converge only root-exponentially. Whether the full potential of this approach can be realized is an open question.

At this stage, I have many more questions than answers, and I will close with a list of them:

  1. We have seen that Newton’s iteration for the inverse square root achieves full exponential convergence in operation count for the problem of approximating \left|x\right| on the interval -1 \le x \le 1 by a polynomial. There is a factor of 2.7 gap between the rate constant realized by this technique and the maximum possible rate constant set by applying the Bernstein bound to the degree. Is it possible to achieve the full Bernstein bound? This would be surprising, in a way, since these iteratively generated polynomials have exponentially fewer coefficients than a polynomial in standard form of the same degree. But if the limit is lower, what is it?
  2. Is super-exponential convergence in operation count possible for iterated rationals in this problem? If not, what is the maximum possible exponential convergence rate?
  3. How generically can iterated polynomials and rationals approximate non-smooth functions with exponential (or super-exponential) convergence in operation count? The iterations we have examined are specific to approximating the square root and inverse square root, but similar iterations with optimized coefficients should be applicable more generally in principle. What performance can they achieve, say for the problem of approximating |x|^{\alpha} on -1 \le x \le 1 for other positive \alpha?
  4. Minimax polynomial and rational approximations are widely used in standard library implementations of mathematical functions, but almost always over intervals where the functions they approximate are smooth. Could iterated polynomials or rationals with optimized coefficients offer practical advantages in these settings?