Partial Derivatives

#math

Chapter Overview

This chapter covers multivariable calculus fundamentals: functions of several variables, limits, continuity, partial derivatives, chain rule, directional derivatives, gradients, optimization, and Lagrange multipliers. The concepts build systematically from basic definitions to complex optimization techniques.


14.1 Functions of Several Variables

Core Definitions

Function of Several Variables

A function of several variables is a rule that assigns a single value in the range to each point in the domain. For a function f(x,y), we write:

z=f(x,y)

where x and y are independent variables and z is the dependent variable.

Intuitive Understanding: Just as y=f(x) describes a curve in 2D, z=f(x,y) describes a surface in 3D space. Each point (x,y) in the domain maps to exactly one height z.

Domain and Range

  • Domain: The set of all possible input values (x,y) for which f(x,y) is defined
  • Range: The set of all possible output values f(x,y) as (x,y) varies throughout the domain

Example: For f(x,y)=9x2y2, the domain is the disk x2+y29 and the range is [0,3].

Geometric Concepts

Interior Point, Boundary Point, Limit Point

For a region R in the xy-plane:

  • Interior point: A point P in R such that some disk centered at P lies entirely in R
  • Boundary point: A point P such that every disk centered at P contains both points in R and points not in R
  • Limit point: A point P such that every deleted neighborhood of P contains points of R

Visualization: Think of a filled circle. Interior points are "safely inside," boundary points are on the edge, and limit points include both interior and boundary points.

Open Region, Closed Region, Closure

  • Open region: A region consisting entirely of interior points
  • Closed region: A region that contains all its boundary points
  • Closure of a region: The union of the region and all its boundary points
  • Bounded region: A region that lies inside some disk
  • Unbounded region: A region that is not bounded

Key Insight: Open regions are like "intervals without endpoints" - they don't include their boundaries. Closed regions include their boundaries.

Level Curve and Level Surface

  • Level curve: For f(x,y), the curve f(x,y)=c for constant c
  • Level surface: For f(x,y,z), the surface f(x,y,z)=c for constant c

Physical Interpretation: Level curves are like contour lines on a topographic map - they connect points of equal elevation. For temperature functions, they're isotherms.


14.2 Limits and Continuity in Higher Dimensions

Limit Definitions

Thomas' Definition of Limit

We say lim(x,y)(x0,y0)f(x,y)=L if for every ε>0, there exists a δ>0 such that whenever

0<(xx0)2+(yy0)2<δ

we have |f(x,y)L|<ε.

Key Difference from Single Variable: We must approach the point from ALL possible directions in the plane, not just left and right.

Path-Dependent Approach

lim(x,y)(x0,y0)f(x,y)=L if f(x,y) approaches L along every possible path to (x0,y0).

Critical Insight: If different paths give different limits, the limit does not exist. This is the basis of the two-path test.

Key Theorems

Properties of Limits (Theorem 1)

If lim(x,y)(x0,y0)f(x,y)=L and lim(x,y)(x0,y0)g(x,y)=M, then:

  1. lim(x,y)(x0,y0)[f(x,y)+g(x,y)]=L+M
  2. lim(x,y)(x0,y0)[f(x,y)g(x,y)]=LM
  3. lim(x,y)(x0,y0)f(x,y)g(x,y)=LM if M0
  4. Root Rule: lim(x,y)(x0,y0)f(x,y)n=Ln if n is odd, or if n is even and L>0

Test Methods

Two-Path Test for Nonexistence of a Limit

If f(x,y) approaches different values along two different paths to (x0,y0), then lim(x,y)(x0,y0)f(x,y) does not exist.

Strategy: Try paths like y=mx (straight lines), y=mx2 (parabolas), and x=0,y=0 (coordinate axes).

Continuity

A function f(x,y) is continuous at (x0,y0) if:

lim(x,y)(x0,y0)f(x,y)=f(x0,y0)

Practical Test: All polynomial and rational functions are continuous where they're defined. Compositions of continuous functions are continuous.


14.3 Partial Derivatives

Basic Definitions

Partial Derivative

The partial derivative of f(x,y) with respect to x at (x0,y0) is:

fx|(x0,y0)=limh0f(x0+h,y0)f(x0,y0)h

Similarly, the partial derivative with respect to y is:

fy|(x0,y0)=limh0f(x0,y0+h)f(x0,y0)h

Geometric Interpretation: fx is the slope of the curve formed by intersecting the surface z=f(x,y) with the plane y=y0. It measures how steeply the surface rises in the x-direction while keeping y fixed.

Computational Rule

To find fx: Treat all other variables as constants and differentiate normally with respect to x.

Example: If f(x,y)=x3y2+5xy+sin(x), then:

  • fx=3x2y2+5y+cos(x) (treat y as constant)
  • fy=2x3y+5x (treat x as constant)

Higher Order Partial Derivatives

Second-Order Partial Derivatives

For z=f(x,y):

  • fxx=2fx2=x(fx)
  • fyy=2fy2=y(fy)
  • fxy=2fyx=y(fx) (mixed partial)
  • fyx=2fxy=x(fy) (mixed partial)

Order of Operations: In fxy, differentiate first with respect to x, then with respect to y.

Key Theorems

Clairaut's Theorem (Mixed Derivative Theorem) - Theorem 2

If f(x,y) and its partial derivatives fx, fy, fxy, and fyx are all continuous in an open region containing (a,b), then:

fxy(a,b)=fyx(a,b)

Practical Impact: For "nice" functions, the order of mixed partial differentiation doesn't matter. This dramatically simplifies calculations.

Differentiability

A function f(x,y) is differentiable at (x0,y0) if:

f(x0+Δx,y0+Δy)f(x0,y0)=fx(x0,y0)Δx+fy(x0,y0)Δy+ε1Δx+ε2Δy

where ε1,ε20 as (Δx,Δy)(0,0).

Geometric Meaning: The surface has a well-defined tangent plane at the point, and the function is well-approximated by this plane near the point.
If the partial derivatives of a function f(x,y) exist at (x0,y0) and are continuous there, then the function is differentiable there. But the converse is not true!!

Differentiability Implies Continuity (Theorem 4)

If f(x,y) is differentiable at (x0,y0), then f is continuous at (x0,y0).

Warning: The converse is false! A function can be continuous but not differentiable (like f(x,y)=x2+y2 at the origin).


14.4 The Chain Rule

The chain rule extends to multivariable functions but becomes more complex due to multiple paths of dependence. Understanding the "dependency tree" is crucial.

Chain Rule Formulas

Chain Rule for One Parameter (Theorem 5)

If z=f(x,y) is differentiable and x=g(t), y=h(t) are differentiable, then:

dzdt=zxdxdt+zydydt

Conceptual Understanding: The rate of change of z with respect to t equals the sum of:

  • How z changes with x times how x changes with t
  • How z changes with y times how y changes with t
Chain Rule for Two Parameters (Theorem 6)

If z=f(x,y) where x=g(u,v), y=h(u,v), then:

zu=zxxu+zyyuzv=zxxv+zyyv
General Chain Rule (Theorem 7)

If w=f(x,y,z) where x=g(u,v), y=h(u,v), z=k(u,v), then:

wu=wxxu+wyyu+wzzu

Tools and Applications

Branch Diagrams (Tree Diagrams)

Tree diagrams show the dependency relationships between variables. Each path from the top variable to a bottom variable represents a term in the chain rule.

How to Use:

  1. Draw the dependent variable at the top
  2. Draw intermediate variables in the middle
  3. Draw independent variables at the bottom
  4. Connect variables with arrows showing dependencies
  5. Each complete path gives one term in the chain rule
Implicit Differentiation (Theorem 8)

If F(x,y)=0 defines y implicitly as a function of x, and Fy0, then:

dydx=FxFy

Extension to More Variables: If F(x,y,z)=0 defines z implicitly, then:

zx=FxFz,zy=FyFz

Challenging Chain Rule Problems

Challenging Problem Type

Problem: If w=f(x2+y2,xy) where x=rcosθ and y=rsinθ, find wr and wθ.

Strategy: Let u=x2+y2, v=xy. Then w=f(u,v).

  • wr=wuur+wvvr
  • Need: ur=r(r2)=2r
  • Need: vr=r(r2cosθsinθ)=2rcosθsinθ

14.5 Directional Derivatives and Gradient Vectors

This section introduces the crucial concept of the gradient - arguably the most important concept in multivariable calculus for applications.

Core Definitions

Directional Derivative

The directional derivative of f(x,y) at point P0(x0,y0) in the direction of unit vector u=a,b is:

Duf(x0,y0)=limh0f(x0+ha,y0+hb)f(x0,y0)h

Physical Interpretation: If you're standing at point (x0,y0) on a hill represented by z=f(x,y), the directional derivative tells you how steeply the hill rises if you walk in direction u.

Gradient Vector

The gradient of f(x,y) is the vector:

f(x,y)=fx,fy=fxi+fyj

For functions of three variables:

f(x,y,z)=fx,fy,fz

Fundamental Insight: The gradient is a vector field that points in the direction of steepest increase of the function at each point.

Key Theorems and Properties

Computation of Directional Derivatives (Theorem 9)

If f is differentiable at P0, then the directional derivative in the direction of unit vector u is:

Duf(P0)=f(P0)u

This is the most important formula in the chapter! It connects directional derivatives to gradients through the dot product.

Properties of the Gradient

  1. Maximum rate of increase: f points in the direction of maximum rate of increase
  2. Magnitude: |f| gives the maximum directional derivative
  3. Minimum rate: f points in direction of maximum rate of decrease (steepest descent)
  4. Zero directional derivative: Directions perpendicular to f have zero directional derivative
  5. Level curve relationship: f level curves

Gradient Normal to Level Curves

At any point, the gradient vector f is perpendicular to the level curve f(x,y)=c passing through that point.

Application: This is why water flows perpendicular to contour lines on topographic maps - it follows the negative gradient (steepest descent).

Algebraic Rules for Gradients

Gradient Rules

  1. (af+bg)=af+bg (linearity)
  2. (fg)=fg+gf (product rule)
  3. (fg)=gffgg2 (quotient rule)
  4. (fn)=nfn1f (power rule)
  5. (f(g))=f(g)g (chain rule)

Challenging Directional Derivative Problems

Challenging Problem (Based on Exercise 14.5)

Problem: Find the directional derivative of f(x,y,z)=xeyz+ln(x2+y2) at point (1,0,2) in the direction from (1,0,2) to (3,1,1).

Solution Strategy:

  1. Find f=eyz+2xx2+y2,xeyzz+2yx2+y2,xeyzy
  2. Evaluate at (1,0,2): f(1,0,2)=1+2,0,0=3,0,0
  3. Direction vector: v=(3,1,1)(1,0,2)=2,1,1
  4. Unit vector: u=2,1,16
  5. Duf=fu=3,0,02,1,16=66=6

14.6 Tangent Planes and Differentials

This section provides the multivariable analog of linear approximation, essential for error analysis and optimization.

Tangent Planes

Tangent Plane Equation

The tangent plane to the surface z=f(x,y) at point (x0,y0,z0) has equation:

$$z - z_0 = f_x(x_0,y_0)(x-x_0) + f_y(x_0,y_0)(y-y_0)$$

Geometric Understanding: The tangent plane is the "best linear approximation" to the surface at the given point. It's the 3D analog of a tangent line.

Normal Vector

The normal vector to the surface z=f(x,y) at (x0,y0,z0) is:

n=fx(x0,y0),fy(x0,y0),1

Alternative Form: For surfaces defined by F(x,y,z)=0, the normal vector is F.

Linear Approximations and Differentials

Linearization

The linearization of f(x,y) at (x0,y0) is:

L(x,y)=f(x0,y0)+fx(x0,y0)(xx0)+fy(x0,y0)(yy0)

Approximation: f(x,y)L(x,y) for (x,y) near (x0,y0)

For a function dependant on three or more variables, it's linearization can be represented as $$L(x,y,z, \dots ) = f(P_{0}) + f_{x}(P_{0})\dd x + f_{y}(P_{0})\dd y + f_{z}(P_{0})\dd z + \dots $$

Total Differential

The total differential of z=f(x,y) is:

dz=fxdx+fydy=fxdx+fydy

Error Estimation: For small changes,

Δzdz=fx(x0,y0)Δx+fy(x0,y0)Δy

Applications of Differentials

Error Propagation

Problem: The radius and height of a cylinder are measured as r=5±0.1 cm and h=12±0.2 cm. Estimate the maximum error in the calculated volume V=πr2h.

Solution:

  • Vr=2πrh, Vh=πr2
  • dV=2πrhdr+πr2dh
  • At (r,h)=(5,12) with dr=±0.1, dh=±0.2:
  • dV=2π(5)(12)(0.1)+π(5)2(0.2)=12π+5π=17π cm³

14.7 Extreme Values and Saddle Points

This section extends single-variable optimization to functions of several variables, introducing the crucial second derivative test.

Critical Points and Classifications

Critical Point

A point (a,b) is a critical point of f(x,y) if either:

  1. fx(a,b)=0 and fy(a,b)=0, or
  2. At least one partial derivative does not exist at (a,b)

Geometric Interpretation: At interior critical points, the tangent plane is horizontal (if it exists).

Types of Critical Points

  • Local maximum: f(a,b)f(x,y) for all (x,y) near (a,b)
  • Local minimum: f(a,b)f(x,y) for all (x,y) near (a,b)
  • Saddle point: Neither a local max nor min - the surface "saddles" at this point

The Second Derivative Test

Second Derivative Test (Theorem 11)

Let (a,b) be a critical point where fx(a,b)=fy(a,b)=0, and suppose f has continuous second partial derivatives. Define the discriminant:

D=fxx(a,b)fyy(a,b)[fxy(a,b)]2

Then:

  1. D>0 and fxx(a,b)>0: Local minimum
  2. D>0 and fxx(a,b)<0: Local maximum
  3. D<0: Saddle point
  4. D=0: Test inconclusive
Hessian Matrix and Discriminant

The Hessian matrix of f(x,y) is:

H=(fxxfxyfxyfyy)

The discriminant is: D=det(H)=fxxfyy(fxy)2

Connection to Linear Algebra: The second derivative test is really about the eigenvalues of the Hessian matrix. D>0 means both eigenvalues have the same sign.

Optimization Strategy

Finding Absolute Extrema on Closed Bounded Regions

  1. Find interior critical points: Solve f=0
  2. Find boundary extrema: Use techniques like parameterization or Lagrange multipliers
  3. Evaluate f at all candidates: Compare values to find absolute max/min

Extreme Value Theorem: Continuous functions on closed, bounded regions always attain their maximum and minimum values.

Challenging Optimization Problems

Difficult Second Derivative Test Problem

Problem: Classify all critical points of f(x,y)=x4+y44x2y+2y2.

Solution Process:

  1. fx=4x38xy=4x(x22y)=0
  2. fy=4y34x2+4y=4y(y2+1)4x2=0
  3. From first: x=0 or x2=2y
  4. Case analysis leads to critical points requiring careful algebraic manipulation
  5. Second derivative test with D=fxxfyy(fxy)2 at each point

14.8 Lagrange Multipliers

This powerful technique handles constrained optimization - finding extrema when variables are restricted by constraint equations.

The Method of Lagrange Multipliers

Lagrange Multiplier Method

To find the extreme values of f(x,y,z) subject to constraint g(x,y,z)=k, solve the system:

f=λgandg(x,y,z)=k

where λ is the Lagrange multiplier.

Geometric Principle: At an extremum, the gradient of the objective function must be parallel to the gradient of the constraint. This happens because both gradients are perpendicular to the constraint surface.

The Orthogonal Gradient Theorem

Suppose that f(x,y,z) is differentiable in a region whose interior consists of the smooth curve C given by $$ \mathbf{r}(t) = x(t)\mathbf{i} + y(t)\mathbf{j} + z(t)\mathbf{k} $$

If P0 is a point on C where f has an extremum relative to it's values on C, then f is orthogonal to C at P0

Multiple Constraints

Two Constraints

To find extrema of f(x,y,z) subject to g(x,y,z)=0 and h(x,y,z)=0:

f=λg+μhg(x,y,z)=0,h(x,y,z)=0

Geometric Picture: The constraints define a curve (intersection of two surfaces), and we want the extrema of f along this curve.
Invoking Orthogonal Gradient theorem, we know f must be perpendicular to this curve C, but since this curve itself is an intersection of both g and h, both g and h are perpendicular to it, meaning, f lies in the same plane as g and h

Challenging Lagrange Multiplier Problems

Advanced Lagrange Problem (Similar to Thomas Exercise 14.8, No. 15-20)

Problem: Find the minimum distance from the origin to the curve of intersection of the cylinder x2+y2=1 and the plane x+y+z=2.

Setup: Minimize f(x,y,z)=x2+y2+z2 subject to:

  • g(x,y,z)=x2+y21=0
  • h(x,y,z)=x+y+z2=0

Solution Strategy:

  1. f=2x,2y,2z
  2. g=2x,2y,0
  3. h=1,1,1
  4. System: 2x,2y,2z=λ2x,2y,0+μ1,1,1
  5. This gives: 2x=2λx+μ, 2y=2λy+μ, 2z=μ
  6. Combined with constraints, solve for critical points
Economic Application

Problem: A company produces two goods with production function f(L,K)=L0.6K0.4 where L is labor and K is capital. If labor costs w per unit and capital costs r per unit with total budget B, find the optimal allocation.

Mathematical Setup: Maximize f(L,K)=L0.6K0.4 subject to wL+rK=B

Key Insight: The Lagrange multiplier λ represents the marginal utility of money - how much additional output we get per additional dollar of budget.


14.9 Taylor's Formula for Two Variables

Second-Order Taylor Formula

For a function f(x,y) with continuous partial derivatives up to order 2:

f(x0+h,y0+k)=f(x0,y0)+fx(x0,y0)h+fy(x0,y0)k+12[fxx(x0,y0)h2+2fxy(x0,y0)hk+fyy(x0,y0)k2]+R2

Applications:

  • Error analysis for linear approximations
  • Theoretical foundation for the second derivative test
  • Quadratic approximations of functions

14.10 Constrained Partial Derivatives

Constrained Variables

When variables are related by constraints, we must account for these relationships when computing partial derivatives.

Example: If w=f(x,y,z) and g(x,y,z)=0, then when we compute wx, we can't hold y and z independent - they're related by the constraint.

Computing Constrained Partials

  1. Method 1: Use the constraint to eliminate one variable, then differentiate
  2. Method 2: Use implicit differentiation on the constraint combined with chain rule
  3. Method 3: Use the formula involving gradients of the constraint


Practice Problems and Common Pitfalls

Hardest Problem Types Students Face

Common Difficulties

  1. Chain Rule with Multiple Paths: Students often miss terms or get confused about dependency relationships
  2. Second Derivative Test Calculations: Computing the discriminant correctly, especially with mixed partials
  3. Lagrange Multipliers Setup: Identifying constraints correctly and setting up the gradient equations
  4. Directional Derivatives: Forgetting to use unit vectors or miscomputing gradients

Challenge Problems (Based on End-of-Chapter Exercises)

Challenging Chain Rule (Similar to Exercise 14.4 No. 25-30)

If w=f(u,v) where u=x2y2, v=2xy, and f is differentiable, show that:

xwxywy=wu(x2+y2)
Advanced Optimization (Similar to Exercise 14.7 No. 35-40)

Find the points on the surface xyz=1 that are closest to the origin. (This requires Lagrange multipliers and careful analysis of the constraint.)

Complex Directional Derivative (Similar to Exercise 14.5 No. 25-30)

Find the direction in which f(x,y)=xeyyex increases most rapidly at (1,1), and find the rate of increase in that direction.


Chapter 14 Key Insights

  • Partial derivatives extend calculus to multivariable functions by examining rates of change in coordinate directions
  • The gradient f is the most important concept - it points toward maximum increase and is perpendicular to level curves
  • Chain rule becomes more complex but follows systematic patterns using dependency trees
  • Optimization uses critical points and the second derivative test, with the Hessian matrix providing geometric insight
  • Lagrange multipliers handle constrained optimization by ensuring gradients are parallel at extrema
  • Applications span physics, engineering, economics, and all fields involving optimization and rates of change

Powered by Forestry.md