25 - Taylor's theorem
Where we left off last time
What are some applications of the derivative? Where did we leave off last time and where were we going?
-
Recap: We have been talking about the derivative. Last time we defined the derivative. We said what it means for a function to be differentiable, which basically means that a certain limit exists. The derivative is the limit of a difference quotient, where the difference quotient represents the slope of a secant line. We say that a function is differentiable at a point if the limiting slope of a secant line actually exists. We also talked about a very important theorem which basically is the most important theorem when it comes to derivatives. It connects the value of the function to the value of the derivative. It was the mean value theorem. Today we'll see how it's important and talk about a generalization known as Taylor's theorem. (Taylor's theorem is really just a generalization of the mean value theorem.) That will be the first part of this lesson. The second part we will discuss sequences of functions, which is a different topic.
-
Taylor's theorem (motivation): Suppose you know something about a function and what it is doing at a particular point. The idea of Taylor's theorem is, well, if you know about what's happening at a point, then you know what's happening near a point, if it's differentiable, if it's twice-differentiable, or if you have a number of derivatives. In the simplest case, suppose we know what a function is doing at a point , and we want to approximate what's happening near , say at a point . That is, given we know something about , suppose we want to approximate , where is near .
Well, the mean value theorem tells us something about if we know and the derivative nearby. So the mean value theorem basically says I can figure out what is if I know and I know something about the derivative nearby:
Observe how this is a restatement of the mean value theorem. And the thing to notice here is that there is a mysterious point , and all we know about is that it lives between and . That is, we have for some . So we have no idea what point is. The way to think about this is that we will know if we know and , a term we have little control over or know much about. What we do know is that this term is something like an error term. It tells us how far off is from . So we can think of the formulation like so:
This "error term" is often not precisely known. But it has something to do with the derivative nearby. So if we can bound the derivative nearby, then maybe we can actually say something about that error. That is one way of thinking about the mean value theorem.
This actually suggests that we might be able to do better with our approximation. We don't know what is doing at . We don't even know where is. But what if we knew what the derivative was doing at ? Then maybe we could do better. This suggests that maybe
where this "error" term is hopefully smaller than the error term before. So this suggests there might be a theorem like this: can we fill in this process where the error terms get smaller and smaller? And the answer is yes. If our function were twice differentiable, then we get as our error term, where this is not the same as that in the first statement but just some . Notice that the error term is a little bit better than what we originally had because if is really small, then is really small. But once again, we have no control over where is. We just know .
So this is sort of the direction Taylor's theorem goes in. It gives us some sort of idea of how good this particular error term is. What's the moral of this story? The moral is that if we know a function value and its derivative , and we know its second derivative in some neighborhood of , then we will have a good handle on what is.
Of course, we could continue the process above. We could say the error term in
is kind of like a second derivative. So maybe if we thrown in a term like , where is replaced by (i.e., we would have), then we will yet still get a better error term. That's really the direction that Taylor's theorem goes into.
-
Taylor's theorem (statement): Generally, if
Here's what we should notice: where do the 's appear in the expression above? The first term is just a constant with respect to . There's an in the next term, an in the following term, and so on. So we see that is a polynomial in (or really a polynomial in ) and its degree is . What is the dependence on ? To determine the coefficients, note that is involved with every single one of them. So is a polynomial, and it is often called a Taylor polynomial.
-
Taylor's theorem (statement): If is continuous on and exists on , then approximates , and
where .
Note that the error term here looks a lot like the terms of the Taylor polynomial except that the th derivative is evaluated somewhere between and . What does this theorem say in the case that ? Can we prove this theorem in the case where ? Yes, what is it? (Note that when .) Well, the function is just saying "tell me what's going on at ," and add what? We'd add . That is, for , we should have
Is that true for some ? Yes, by the mean value theorem. When , this literally is the mean value theorem. So Taylor's theorem really is just a generalization of the mean value theorem.
Now, is the "best" approximation of order at . What do we mean by "best" here? Well, what we mean is we've constructed a polynomial that has all the same value in derivatives up to the th; that is, has the same value of at . That is, we have the following correspondence (at ):
So the first derivative of the polynomial and the first derivative of both have the same value at . Let's just convince ourselves of that. If we look at the polynomial
at , then we see that we have
because all of the other terms go to zero.
What happens if we take the first derivative of ? What happens to the term? It disappears because the derivative with respect to is gone. What happens when we take the derivative of with respect to ? We just get . And the rest of the terms have positive powers of in them and hence evaluate to zero at . So .
Similarly, if we look at , then we have the following:
Now it should be clear where the factorials are coming from. They are needed to cancel the powers that are coming down from the differentiation process. So if you want the th derivatives to match up, then you will need the underneath. So how do we prove this?
-
Taylor's theorem (proof): For some number , the statement
is true (for some suitable choice of ). Let , where we hope is small. Equal? Not necessarily. So let's amend our definition of using the -value above we claimed existed:
Why would we do this? What are we hoping for? Where is this headed? Well, it sure would be nice if we could show that were 0 somewhere and maybe were the coefficient . What happens if we take the th derivative of ? We'll have
We claim it is enough to show that for some . Why would that be enough for what we are interested in? If we could show that
was zero somewhere, then that would make . That would establish that is, in fact, the th derivative of at some divided by ; that is, we would have
Let's check a few things about . We would have the following: