Intuition behind proving that $f$ is constant

You could see there as being two parts to that answer.

  1. How to relate the $h$ and $2h$ offsets to the usual $(f(x+h)-f(x))/h$ quotient used in derivatives
  2. Applying a Rolle's theorem type argument to show that a zero one-sided derivative implies $f$ is constant

For 1, a more general statement is true. If $f$ is continuous and $\lim_{h\to 0^+}\frac{f(x+2h)-f(x+h)}{h}=s$ then $\lim_{h\to 0^+}\frac{f(x+h)-f(x)}{h}=s.$ Consider a small $\epsilon$ such that $\frac{f(x+2h)-f(x+h)}{h}$ is close to $s$ for $0<h<\epsilon.$ In geometric terms, the slope between the points $(x+h,f(x+h))$ and $(x+2h,f(x+2h))$ is approximately $s.$ In these terms, the trick is then to notice that the slopes between $(x+h,f(x+h))$ and $(x+h/2,f(x+h/2)),$ and then to $(x+h/4,f(x+h/4)),$ etc, are all approximately $s$ for positive $k$ and all $h<\epsilon.$ And these approach $(x,f(x))$ as $k\to\infty.$ The overall slope $(f(x+h)-f(x))/h$ is therefore approximately $s.$

When writing this out as a proof, it makes life easier to assume $s=0$ by applying a linear transformation. (Similar to how you prove the mean value theorem given Rolle's theorem.)

For 2, I think you just have to have the idea of applying something like Rolle's theorem, as in the mean value theorem. Define $g(x)=f(x)+\lambda x$ for an arbitrary $\lambda\neq 0,$ and apply the argument from Rolle's theorem: the derivative at a local extremum is constrained. You need to check that the argument makes sense for one-sided derivatives, but that's fine: it doesn't tell you the derivative is zero, but it tells you the right derivative must be non-positive at a maximum. So $g$ can't have any local maximum for $\lambda> 0.$ This can be turned into a proof as in the suggested answer.


Here's an alternate approach to 2 you might enjoy. Fix any $a,$ any $\epsilon>0,$ and set $b=\inf\{x>a\mid f(x)-f(a)>\epsilon(x-a)\}$ if this set is non-empty. Then $f(b)-f(a)=\epsilon(b-a)$ and there is a sequence $h\to 0^+$ with $(f(b+h)-f(b))/h>\epsilon,$ contradicting the result that the right derivative is zero. A similar argument shows $|f(x)-f(a)|\leq \epsilon(x-a)$ for all $\epsilon$ and hence $f$ is constant.

Another approach is to use the following lemma: if $g:[a,b]\to\mathbb R$ is continuous and for each $x\in[a,b)$ there exists $x'\in(x,b]$ such that $f(x')\geq f(x),$ then $f(b)\geq f(a).$ Proof of lemma: consider $\sup\{x\in[a,b]\mid f(x)\geq f(a)\}.$ The lemma can be applied to $g(x)=f(x)+\lambda x$ to show $f(b)\geq f(a)-\lambda (b-a)$ for all $\lambda>0,$ hence $f(b)\geq f(a),$ and similarly $f(b)\leq f(a).$ The trick of using $\lambda$ is then "giving yourself an $\epsilon$ of room" - it's easier to show a function with positive derivative is strictly non-decreasing than to show a function with zero derivative is non-decreasing.


We want to show this by contradiction. So we will have to assume $a < b$ which are two points at which there is a difference of values. All we know is that $f$ is continuous and $[a,b]$ is compact : this has to be suitably used. Of course, this gives the existence of a maximum and minimum value of $f$ attained on $[a,b]$.

What we do is this : since we are assuming $f(a) > f(b)$, what we ideally want to do, is to show that the maximum, which is not attained at $b$ definitely, is not attained anywhere.

What if we were to experiment with $f$?

If $c$ were such a maximum point, then nearby $c$, we have , by interpretation of the limit being zero, that given $d$ just above $c$, the values of $f(d)$ and $f(\frac{c+d}{2})$ are close. A rather naive idea would then be as follows : through some iteration, that is to say, $f(\frac{c+d}{2})$ and $f(\frac{\frac{c+d}{2} + c}{2})$ are close, and now take another arithmetic mean, and a sum of "close differences" should be close etc., we want to try to get to the conclusion in some manner. There are two roadblocks:

  • Firstly, the iteration and limit process does not help us conclude that $f(d)$ is close to $f(c)$, because an infinite sum of very small quantities may not even be finite, let alone small enough or something.

  • Second and more importantly, even if $f(d)$ was close, it could be either side of $f(c)$ : we would prefer if we could somehow get $f(d) > f(c)$, because that is the contradiction we seek. This cannot happen here.

That, is the need for a "contrived" function. What this contrived function must do, is to make the given limit strictly positive for all $x$ in $[a,b]$. What this small change does, is magic : it suitably modifies $f$ in the interval $[a,b]$, so that $g(d)$ ends up not being just close to $g(\frac {c+d}2)$, but also larger. Now,you can throw both points out of the window : there are no sum of small quantities to be taken, and inequalities are preserved in the limit!

In other words, the condition given to us, is not manipulable, because the fact that the limit is zero at every point, and not any other positive quantity, hinders us from contradicting the existence of maxima/minima on a compact interval like $[a,b]$. So what we have done with the creation of $g$, is created a similar situation, but the crucial point I have emphasised : the limit being positive, which has helped, in that we have been able to contradict the situation.

Now, the question you may ask is this : could we have chosen some other $g$ maybe? Well, we might have : all we needed to ensure, was that $g$ preserved $g(a) > g(b)$, and that the limit which is given to be zero for $f$, should be strictly positive on $[a,b]$. A linear addition does this best : I am sure we can find other candidates that may have done the same.

With this , I hope you understand the following :

Why, having taking $a < b$ and $f(a) > f(b)$, the suitable contradiction is to contradict the existence of a maximum/minimum of $f$ on $[a,b]$.

Why $f$ stumbles in its pursuit to contradict the maximum existence.

How $g$ was created to do the job that $f$ failed in.

Why $g$, of course, does the job, and there is, at least at first glance, no reason why $g$ is unique.

For once, we sat in anguish at seeing a limit being zero : the inability to assert the sign of the differential quotient was giving trouble. This is sorted out rather elegantly by $g$.

For an alternate approach, I am sure that there must be some theorem involving lots of little lemmas, that will show that $f$ is differentiable. But when I think of it, I will get back to you.