4.1 The Derivative: Definition and Elementary Consequences - Zooming in on Change
We now come to the heart of differential calculus: the derivative. The derivative captures the instantaneous rate of change of a function at a point. Geometrically, it represents the slope of the tangent line to the graph of the function at that point.
Imagine zooming in on the graph of a function at a particular point . If the function is “smooth” at that point, as you zoom in closer and closer, the graph will start to look more and more like a straight line. The slope of this line is the derivative at .
To make this precise, we define the derivative using a limit:
Definition 4.1.1: Differentiability and the Derivative
Let , , and be a limit point of . We say is differentiable at if the following limit exists:
If this limit exists, we call it the derivative of at , denoted by or .
The expression represents the slope of the secant line through the points and . As approaches , this secant line approaches the tangent line, and its slope approaches the derivative .
Alternative Form of the Derivative:
Sometimes it’s convenient to express the derivative using a change in , denoted by , so . As , we have . This gives us an equivalent definition:
Remark 4.1.2: Alternative Definition of the Derivative
This form is often useful in calculations and proofs.
Linear Approximation: Zooming in and Seeing a Line
Differentiability at a point means we can approximate the function near by a linear function. This is the essence of the “zooming in” idea.
Theorem 4.1.3: Linear Approximation Characterization of Differentiability
A function is differentiable at if and only if there exists a real number and a function such that:
- Linear Approximation Formula: For all , .
- Remainder Term Vanishes Near : , and we can define to make continuous at .
If these conditions hold, then is uniquely determined and .
This theorem provides a different perspective on differentiability. It says is differentiable at if we can write as:
for close to . The term is the equation of the tangent line at . The remainder term represents the “error” in this linear approximation, and it becomes negligible as approaches (since ).
Proof of Theorem 4.1.3 (Sketch)
() If is differentiable, then linear approximation exists: Define and for , and . Rearranging this gives the linear approximation formula. Taking the limit as shows , so is continuous at .
() If linear approximation exists, then is differentiable: Given with . Rearranging: . Taking the limit as : . So the derivative exists and is equal to .
This linear approximation view is fundamental for understanding applications of derivatives in physics, engineering, and other fields, where we often approximate nonlinear functions with linear ones locally.
Differentiability Implies Continuity: Smoothness Requires No Jumps
If a function is differentiable at a point, it must also be continuous at that point. Smoothness is a stronger condition than just being “connected”.
Corollary 4.1.5: Differentiability Implies Continuity
If a function is differentiable at , then is also continuous at .
Proof of Differentiability Implies Continuity
Using the linear approximation form (Theorem 4.1.4): , where is continuous at .
Taking the limit as : .
Thus , which is the definition of continuity at .
The converse is not true: continuity does not imply differentiability. A function can be continuous but have “corners” or “kinks” where the derivative does not exist (e.g., the absolute value function is continuous at , but not differentiable there).
Examples of Derivatives: Basic Building Blocks
Let’s calculate derivatives of some basic functions to get a feel for the definition.
Example 4.1.6 (1): Derivative of a Constant Function
Let (the constant function). Then for all . The slope of a horizontal line is always zero.
Example 4.1.6 (2): Derivative of the Identity Function
Let (the identity function). Then for all . The slope of the line is always 1.
Example 4.1.6 (3): Derivative of
Let . Then for all .
Proof that
Using the definition of the derivative:
Example 4.1.6 (4): Absolute Value Function is Not Differentiable at 0
Let . The derivative does not exist. The function has a “corner” at .
Example 4.1.6 (5): A Continuous Function Nowhere Differentiable
The function (a “sawtooth” function) is continuous everywhere but differentiable nowhere. Even more surprisingly, the function (Van der Waerden function) is continuous everywhere but differentiable nowhere. This shows that there exist continuous functions that are “extremely non-smooth”.
These examples illustrate the definition of the derivative and highlight the relationship between differentiability and smoothness of functions. In the next sections, we will develop rules for computing derivatives more efficiently and explore the fundamental theorems that connect derivatives to the behavior of functions.
4.2 Central Theorems about the (First) Derivative: Making Differentiation Easier
Calculating derivatives directly from the definition can be tedious. Fortunately, there are rules that make differentiation much easier, especially for combinations of functions. These are the Differentiation Rules or “Rechenregeln” in the original script.
Theorem 4.1.9: Differentiation Rules
Let , be a limit point of , and let and be differentiable at . Then:
- Sum Rule: The sum is differentiable at , and .
- Product Rule (Leibniz Rule): The product is differentiable at , and .
- Quotient Rule: If , then the quotient is differentiable at , and
These rules allow us to differentiate sums, products, and quotients of functions by differentiating the individual functions and combining the results algebraically.
Proof of Differentiation Rules (using Linear Approximation - Theorem 4.1.3)
We use the linear approximation characterization of differentiability: and , where and are continuous at and .
1. Sum Rule: . Let . Since and are continuous at , is also continuous at by the sum rule for continuous functions. And . By Theorem 4.1.4, is differentiable at and .
2. Product Rule: . Let . Since are continuous at , is also continuous at . And . By Theorem 4.1.4, is differentiable at and .
3. Quotient Rule: (Proof involves similar but slightly more involved algebraic manipulation using the linear approximation form and showing the quotient fits the required form for differentiability).
Examples using Differentiation Rules:
Example 4.1.10 (1): Derivative of
Using induction and the product rule, we can show that for , .
Example 4.1.10 (2): Derivative of Tangent Function
Using the quotient rule and the derivatives of sine and cosine, we can derive the derivative of the tangent function:
4.3 The Chain Rule: Derivative of Composite Functions - Functions Inside Functions
The Chain Rule tells us how to differentiate a composite function, a function inside another function. If we have , the chain rule tells us how to find its derivative in terms of the derivatives of and .
Theorem 4.1.11: Chain Rule
Let , , . Let and assume is differentiable at , and is differentiable at . Then the composite function , defined by , is differentiable at , and its derivative is:
In words: The derivative of the outside function (evaluated at the inside function) times the derivative of the inside function.
Proof of Chain Rule (using Linear Approximation)
Let , . Since is differentiable at , with and continuous at . Since is differentiable at , with and continuous at .
Substitute and into the expression for : .
Let . Since is continuous at (differentiability implies continuity), and and are continuous at and respectively, the composition and the product are continuous at .
And .
By Theorem 4.1.4, is differentiable at and .
Example: Derivative of .
Let and . Then . Using the chain rule:
4.4 Inverse Functions and Their Derivatives
If a function is bijective (one-to-one and onto) and differentiable, and its inverse function is also continuous, we can find a formula for the derivative of the inverse function.
Corollary 4.1.12: Derivative of Inverse Function
Let be a bijective function, a limit point of , and . Assume is differentiable at with , and the inverse function is continuous at . Then is differentiable at , and its derivative is:
In words: The derivative of the inverse function at is the reciprocal of the derivative of the original function at .
Proof of Derivative of Inverse Function
Let and . Since is differentiable at , with and continuous at .
We want to find the derivative of at . Let . Since is continuous at , .
Rearranging the linear approximation formula: .
So, .
Let . Since and are continuous at and respectively, and , the function is continuous at . And .
By Theorem 4.1.4, is differentiable at and .
Examples: Derivatives of Inverse Functions
-
Derivative of Natural Logarithm: Since the exponential function has derivative and inverse function , we can use the inverse function theorem to find the derivative of :
-
Derivatives of Inverse Trigonometric Functions: We can use the inverse function theorem to find derivatives of arcsin, arccos, arctan, arccot, arcosh, arsinh, artanh, arccot, as shown in the script examples (Example 4.2.6 and 4.1.13).
These differentiation rules, including the Chain Rule and the Inverse Function Theorem, are the workhorses of differential calculus. They enable us to differentiate a vast range of functions and are essential tools for applications in mathematics, physics, engineering, and beyond.
4.5 Central Theorems About the First Derivative: Unlocking Function Behavior
Now we move to some of the most powerful theorems in differential calculus, theorems that reveal deep connections between the derivative of a function and its overall behavior. These are the Central Theorems about the (First) Derivative.
Rolle’s Theorem: Leveling Out
Rolle’s Theorem is a cornerstone result that sets the stage for the Mean Value Theorem. It deals with a function that starts and ends at the same height.
Theorem 4.2.3: Rolle's Theorem (1690)
Let be continuous on and differentiable on . If , then there exists at least one point such that .
Imagine a smooth curve that starts and ends at the same y-value. Somewhere in between, the curve must “level out” – it must have a horizontal tangent line, meaning the derivative is zero.
Proof of Rolle's Theorem
Since is continuous on , by the Min-Max Theorem (Extreme Value Theorem, Theorem 3.4.5), attains a maximum and a minimum value on . Let be points where the minimum and maximum are attained, respectively.
Case 1: is constant. If is constant on , then for all , so any works.
Case 2: is not constant. Since and is not constant, either the maximum value or the minimum value (or both) must be attained at some point inside the interval , not at the endpoints or .
Suppose the maximum is attained at . Then has a local maximum at . By Theorem 4.2.2(3), if has a local extremum (maximum or minimum) at an interior point where it’s differentiable, then . The same logic applies if the minimum is attained at an interior point.
Therefore, in either case, there exists a point such that .
The Mean Value Theorem (MVT): The Average Slope
The Mean Value Theorem (MVT) is a generalization of Rolle’s Theorem. It says that for any smooth curve segment, there’s always a point where the tangent line is parallel to the secant line connecting the endpoints.
Theorem 4.2.4: Mean Value Theorem (Lagrange, 1797)
Let be continuous on and differentiable on . Then there exists at least one point such that:
Rearranging, we get:
The right side is the slope of the secant line connecting and – the average rate of change of over . The MVT says that at some point , the instantaneous rate of change is equal to this average rate of change.
Proof of Mean Value Theorem (Using Rolle's Theorem)
Define a new function which represents the vertical difference between the function and the secant line connecting and :
- Check that : By construction, the secant line passes through and , so the vertical difference is zero at the endpoints.
- Apply Rolle’s Theorem to : Since and the linear function are continuous on and differentiable on , so is . And . Thus, Rolle’s Theorem applies to .
- Conclude for some : There exists such that .
- Compute : .
- Set and rearrange: , which gives , or .
Consequences of the Mean Value Theorem: Linking Derivative to Function Behavior
The MVT is not just a theoretical curiosity; it’s a powerful tool with many important consequences. It allows us to relate the derivative of a function (local information) to its global behavior (monotonicity, boundedness, etc.).
Corollary 4.2.5: Consequences of the Mean Value Theorem
Let be continuous on and differentiable on .
- Zero Derivative Implies Constant Function: If for all , then is constant on .
- Equal Derivatives Imply Constant Difference: If for all , then there exists a constant such that for all .
- Non-negative Derivative Implies Increasing Function: If for all , then is monotonically increasing on .
- Positive Derivative Implies Strictly Increasing Function: If for all , then is strictly monotonically increasing on .
- Non-positive Derivative Implies Decreasing Function: If for all , then is monotonically decreasing on .
- Negative Derivative Implies Strictly Decreasing Function: If for all , then is strictly monotonically decreasing on .
- Bounded Derivative Implies Lipschitz Continuity: If there exists such that for all , then for all , . (Lipschitz continuity implies uniform continuity).
These corollaries are incredibly useful. They allow us to determine if a function is constant, increasing, or decreasing simply by examining the sign of its derivative. They also provide bounds on how much a function can change based on the boundedness of its derivative.
4.6 Higher Derivatives: Rate of Change of Rate of Change
We can differentiate a function more than once. The second derivative is the derivative of the first derivative, the third derivative is the derivative of the second derivative, and so on. These higher derivatives give us information about the rate of change of the rate of change, and so on, capturing more subtle aspects of a function’s behavior, like concavity and inflection points (which we’ll explore later).
Definition 4.3.1: Higher Derivatives
- For , a function is -times differentiable on if its -th derivative is differentiable on . The -th derivative is denoted by . We also write and .
- A function is -times continuously differentiable on if it is -times differentiable on and its -th derivative is continuous on .
- A function is smooth (or infinitely differentiable) on if it is -times differentiable for all .
Notation:
- for the first, second, third, … derivatives.
- for the first, second, third, … derivatives.
- for the -th derivative.
Rules for Higher Derivatives:
We can extend the differentiation rules (Sum Rule, Product Rule, Chain Rule, Quotient Rule) to higher derivatives. For example, Leibniz’s Rule generalizes the Product Rule to higher derivatives:
Theorem 4.3.3: Leibniz Rule for Higher Derivatives of a Product
Let be -times differentiable on . Then their product is also -times differentiable on , and its -th derivative is given by:
This formula looks similar to the Binomial Theorem and is proved using induction.
Examples of Higher Derivatives:
- Exponential Function: The exponential function is infinitely differentiable, and all its derivatives are equal to itself: for all .
- Polynomials: Polynomials are also infinitely differentiable. However, their higher derivatives eventually become zero. For a polynomial of degree , its -th derivative and all higher derivatives are zero.
- Natural Logarithm: The natural logarithm (for ) is infinitely differentiable, with derivatives given by a pattern: for .
4.7 Taylor Approximation: Polynomials That Mimic Functions
Taylor approximation is a powerful technique that uses polynomials to approximate functions. The idea is to find a polynomial that “mimics” the behavior of a function near a particular point, matching its value and its derivatives at that point.
Theorem 4.4.5: Taylor's Theorem (with Remainder in Lagrange Form)
Let be continuous on and -times differentiable on . For any and a fixed point , there exists a point between and such that:
where the remainder term is given by:
The polynomial is called the Taylor polynomial of degree for centered at . It’s the best polynomial approximation of near in the sense that it matches the function’s value and first derivatives at . The remainder term quantifies the error in this approximation.
Taylor Series:
If a function is infinitely differentiable, we can consider the Taylor series:
This is an infinite series that is constructed from the function’s derivatives at . For many important functions (like ), the Taylor series actually converges to the function itself within its radius of convergence. This means we can represent these functions exactly as infinite polynomials (power series).
Taylor approximation and Taylor series are fundamental tools in analysis, approximation theory, numerical analysis, and many other areas. They allow us to approximate complicated functions with simpler polynomials, analyze function behavior, and solve problems that would be intractable otherwise.
This concludes Chapter 4. We’ve explored the derivative, its rules, its connection to function behavior through the Mean Value Theorem, higher derivatives, and the powerful tool of Taylor approximation. These concepts are essential for understanding and applying calculus in a wide range of contexts.