If a polynomial function is trapped in a box, how much can it wiggle? This question is answered by Markov’s inequality, which states that for a degree-
polynomial
that maps
into
, it holds that
(1) ![]()
How tight is this inequality? Do polynomials we know and love come close to saturating it, or is this bound very loose for them? A first polynomial which is natural to investigate is the power
. This function maps
into
, and the maximum value of its derivative is
![]()
To saturate the inequality, we need something wigglier. The heavyweight champions for polynomial wiggliness are the Chebyshev polynomials
, which are motivated and described at length in this previous post. For our purposes, what’s important is that Chebyshev polynomials really know how to wiggle. Just look at how much more rapidly the degree-7 Chebyshev polynomial (blue solid line) moves around than the degree-7 power
(orange dashed line).
In addition to seeming a lot more wiggly to the eye, the degree-
Chebyshev polynomials have much larger derivatives, saturating Markov’s inequality:
![]()
These two examples illustrate a good rule of thumb. The derivative of a polynomial which is “simple” or “power-like” will be of size about
, whereas the derivative of a “clever” or “Chebyshev-like” polynomial will be much higher at about
.
The inequality (1) was proven by Andrey Markov. It is much less well-known than Andrey Markov’s famous inequality in probability theory. A generalization of Markov’s inequality (1) was proven by Andrey’s brother Vladimir Markov, who proved a bound on the
th derivative of any polynomial
mapping
into
:
(2) ![]()
For the rest of this post, we will focus on the basic Markov inequality (1). This inequality is easily extended to polynomials trapped in a box
of general sidelengths. Indeed, any polynomial
mapping
can be transmuted to a polynomial
mapping
to
by an affine change of variables:
![]()
![]()
Markov’s inequality (general domain/codomain). Let
be a degree-
polynomial that maps
to
. Then
(3)
What’s Markov’s inequality good for? A lot, actually. In this post, we’ll see one application of this inequality: proving polynomial inapproximability. There are lots of times in computational math where its valuable to approximate some function like
,
, or
by a polynomial. There are lots of techniques for producing polynomial approximations and understanding the rate of convergence of the best-possible polynomial approximation. But sometimes we just want a quick estimate of the form, say, “a polynomial needs to be of degree at least
to approximate that function up to additive error
“. That is, we seek lower bounds on the polynomial needed to approximate a given function to a specified level of accuracy.
The general Markov’s inequality (3) provides a direct way of doing this. Our treatment follows Chapter 5 of Faster Algorithms via Approximation Theory by Sushant Sachdeva and Nisheesh K Vishnoi. The argument will consist of two steps. First, we trap the function in a box. Then, we show it wiggles a lot (i.e., there is a point at which the derivative is large). Therefore, by Markov’s inequality, we conclude that the degree of the polynomial must be sufficiently large.
Let’s start with the function
and ask the question:
What polynomial degree
do we need to approximate this function to error 0.1 on the interval
?
We are interested in the case where the interval is large, so we assume
. To address this question, suppose that
is a polynomial that satisfies
![]()
- Trap it in a box. Since
for positive
, it therefore must hold that
. Therefore, the polynomial
is trapped in the box
. - Show it wiggles. The function
decreases quite rapidly. At zero, it is
and at two, it is
. Therefore, since
is within
of
for all
, it must hold that
is at least
and
is at most
. Therefore, by the intermediate value theorem, there is some
between
and
for which ![Rendered by QuickLaTeX.com \[p'(t^*) = \frac{p(2) - p(0)}{2-0} \ge \frac{0.9 - 0.24}{2}=0.33.\]](https://www.ethanepperly.com/wp-content/ql-cache/quicklatex.com-128d3240b1f7024150800ae2aea13002_l3.png)
We are ready for the conclusion. Apply the bullet point above and Markov’s inequality (3) to obtain
![]()
![]()
We’ve illustrated with just a single example, but this same technique can also be used to give (sharp) inapproximability results for other functions we know and love like
and
. For a bit of fun, see if you can get results for approximating a power
on
by a polynomial
of degree
. You may be surprised by what you find!
I find the Markov inequality technique for proving polynomial inapproximability to be pretty cool. As we saw in a previous post, we usually understand the difficulty of approximating a function in terms of the rate of convergence of the best (polynomial) approximation, which is tied to fine properties of the function and its smoothness. The Markov inequality approach answers a different question and uses entirely different information about the function. Rather than asking about the asymptotic rate of convergence, the Markov inequality approach tells you at what degree does a polynomial start approximating the function at all. And rather than using information about smoothness, the Markov approach shows that polynomial inapproximability is hard for any function that changes a lot over a small interval. As an exercise, you can show that the same argument for inapproximability of
also shows a polynomial of degree
is necessary for approximating the ramp function
![Rendered by QuickLaTeX.com \[f_{\rm ramp}(t) = \begin{cases} 1-t/2, & 0 \le t \le 2, \\ 0, & 2 < t \le b.\end{cases}.\]](https://www.ethanepperly.com/wp-content/ql-cache/quicklatex.com-26e453856802dfaca0eef8e4b383a649_l3.png)
Reference. This blog post is my take on an argument presented in Chapter 5 of Faster Algorithms via Approximation Theory by Sushant Sachdeva and Nisheesh K Vishnoi. It’s a very nice monograph, and I highly recommend you check it out!