4.1 The Derivative: Definition and Elementary Consequences - Zooming in on Change

We now come to the heart of differential calculus: the derivative. The derivative captures the instantaneous rate of change of a function at a point. Geometrically, it represents the slope of the tangent line to the graph of the function at that point.

Imagine zooming in on the graph of a function at a particular point . If the function is “smooth” at that point, as you zoom in closer and closer, the graph will start to look more and more like a straight line. The slope of this line is the derivative at .

To make this precise, we define the derivative using a limit:

The expression represents the slope of the secant line through the points and . As approaches , this secant line approaches the tangent line, and its slope approaches the derivative .

Alternative Form of the Derivative:

Sometimes it’s convenient to express the derivative using a change in , denoted by , so . As , we have . This gives us an equivalent definition:

This form is often useful in calculations and proofs.

Linear Approximation: Zooming in and Seeing a Line

Differentiability at a point means we can approximate the function near by a linear function. This is the essence of the “zooming in” idea.

This theorem provides a different perspective on differentiability. It says is differentiable at if we can write as:

for close to . The term is the equation of the tangent line at . The remainder term represents the “error” in this linear approximation, and it becomes negligible as approaches (since ).

This linear approximation view is fundamental for understanding applications of derivatives in physics, engineering, and other fields, where we often approximate nonlinear functions with linear ones locally.

Differentiability Implies Continuity: Smoothness Requires No Jumps

If a function is differentiable at a point, it must also be continuous at that point. Smoothness is a stronger condition than just being “connected”.

The converse is not true: continuity does not imply differentiability. A function can be continuous but have “corners” or “kinks” where the derivative does not exist (e.g., the absolute value function is continuous at , but not differentiable there).

Examples of Derivatives: Basic Building Blocks

Let’s calculate derivatives of some basic functions to get a feel for the definition.

These examples illustrate the definition of the derivative and highlight the relationship between differentiability and smoothness of functions. In the next sections, we will develop rules for computing derivatives more efficiently and explore the fundamental theorems that connect derivatives to the behavior of functions.

4.2 Central Theorems about the (First) Derivative: Making Differentiation Easier

Calculating derivatives directly from the definition can be tedious. Fortunately, there are rules that make differentiation much easier, especially for combinations of functions. These are the Differentiation Rules or “Rechenregeln” in the original script.

These rules allow us to differentiate sums, products, and quotients of functions by differentiating the individual functions and combining the results algebraically.

Examples using Differentiation Rules:

4.3 The Chain Rule: Derivative of Composite Functions - Functions Inside Functions

The Chain Rule tells us how to differentiate a composite function, a function inside another function. If we have , the chain rule tells us how to find its derivative in terms of the derivatives of and .

In words: The derivative of the outside function (evaluated at the inside function) times the derivative of the inside function.

Example: Derivative of .

Let and . Then . Using the chain rule:

4.4 Inverse Functions and Their Derivatives

If a function is bijective (one-to-one and onto) and differentiable, and its inverse function is also continuous, we can find a formula for the derivative of the inverse function.

In words: The derivative of the inverse function at is the reciprocal of the derivative of the original function at .

Examples: Derivatives of Inverse Functions

  • Derivative of Natural Logarithm: Since the exponential function has derivative and inverse function , we can use the inverse function theorem to find the derivative of :

  • Derivatives of Inverse Trigonometric Functions: We can use the inverse function theorem to find derivatives of arcsin, arccos, arctan, arccot, arcosh, arsinh, artanh, arccot, as shown in the script examples (Example 4.2.6 and 4.1.13).

These differentiation rules, including the Chain Rule and the Inverse Function Theorem, are the workhorses of differential calculus. They enable us to differentiate a vast range of functions and are essential tools for applications in mathematics, physics, engineering, and beyond.

4.5 Central Theorems About the First Derivative: Unlocking Function Behavior

Now we move to some of the most powerful theorems in differential calculus, theorems that reveal deep connections between the derivative of a function and its overall behavior. These are the Central Theorems about the (First) Derivative.

Rolle’s Theorem: Leveling Out

Rolle’s Theorem is a cornerstone result that sets the stage for the Mean Value Theorem. It deals with a function that starts and ends at the same height.

Imagine a smooth curve that starts and ends at the same y-value. Somewhere in between, the curve must “level out” – it must have a horizontal tangent line, meaning the derivative is zero.

The Mean Value Theorem (MVT): The Average Slope

The Mean Value Theorem (MVT) is a generalization of Rolle’s Theorem. It says that for any smooth curve segment, there’s always a point where the tangent line is parallel to the secant line connecting the endpoints.

Rearranging, we get:

The right side is the slope of the secant line connecting and – the average rate of change of over . The MVT says that at some point , the instantaneous rate of change is equal to this average rate of change.

Consequences of the Mean Value Theorem: Linking Derivative to Function Behavior

The MVT is not just a theoretical curiosity; it’s a powerful tool with many important consequences. It allows us to relate the derivative of a function (local information) to its global behavior (monotonicity, boundedness, etc.).

These corollaries are incredibly useful. They allow us to determine if a function is constant, increasing, or decreasing simply by examining the sign of its derivative. They also provide bounds on how much a function can change based on the boundedness of its derivative.

4.6 Higher Derivatives: Rate of Change of Rate of Change

We can differentiate a function more than once. The second derivative is the derivative of the first derivative, the third derivative is the derivative of the second derivative, and so on. These higher derivatives give us information about the rate of change of the rate of change, and so on, capturing more subtle aspects of a function’s behavior, like concavity and inflection points (which we’ll explore later).

Notation:

  • for the first, second, third, … derivatives.
  • for the first, second, third, … derivatives.
  • for the -th derivative.

Rules for Higher Derivatives:

We can extend the differentiation rules (Sum Rule, Product Rule, Chain Rule, Quotient Rule) to higher derivatives. For example, Leibniz’s Rule generalizes the Product Rule to higher derivatives:

This formula looks similar to the Binomial Theorem and is proved using induction.

Examples of Higher Derivatives:

  • Exponential Function: The exponential function is infinitely differentiable, and all its derivatives are equal to itself: for all .
  • Polynomials: Polynomials are also infinitely differentiable. However, their higher derivatives eventually become zero. For a polynomial of degree , its -th derivative and all higher derivatives are zero.
  • Natural Logarithm: The natural logarithm (for ) is infinitely differentiable, with derivatives given by a pattern: for .

4.7 Taylor Approximation: Polynomials That Mimic Functions

Taylor approximation is a powerful technique that uses polynomials to approximate functions. The idea is to find a polynomial that “mimics” the behavior of a function near a particular point, matching its value and its derivatives at that point.

The polynomial is called the Taylor polynomial of degree for centered at . It’s the best polynomial approximation of near in the sense that it matches the function’s value and first derivatives at . The remainder term quantifies the error in this approximation.

Taylor Series:

If a function is infinitely differentiable, we can consider the Taylor series:

This is an infinite series that is constructed from the function’s derivatives at . For many important functions (like ), the Taylor series actually converges to the function itself within its radius of convergence. This means we can represent these functions exactly as infinite polynomials (power series).

Taylor approximation and Taylor series are fundamental tools in analysis, approximation theory, numerical analysis, and many other areas. They allow us to approximate complicated functions with simpler polynomials, analyze function behavior, and solve problems that would be intractable otherwise.

This concludes Chapter 4. We’ve explored the derivative, its rules, its connection to function behavior through the Mean Value Theorem, higher derivatives, and the powerful tool of Taylor approximation. These concepts are essential for understanding and applying calculus in a wide range of contexts.

Previous Chapter: Chapter 3 - Continuous Functions, Smoothness and Limits, Continuity, Functions Without Jumps, Combining Continuous Functions, Fundamental Theorems of Continuity, Exponential and Trigonometric Functions

Next Chapter: Chapter 5 - The Riemann Integral, Measuring Areas Under Curves, Properties and Classes of Integrable Functions, Properties of Integrals, Linearity, Monotonicity, and the Mean Value Theorem for Integrals, Fundamental Theorem and Applications