Status:
14.8 Lagrange Multipliers
To find the extreme values of a continuous function defined on a bounded closed region, we need to identify its interior critical points and its possible extreme values on the boundary. In one dimension, a closed interval has only two boundary points. In higher dimensions, however, the boundary of a region most of the time consists of infinitely many points. In two dimensions, the boundary of a region is typically the union of a collection of curves and points; in three dimensions, boundary of a region is typically the union of a collection of surfaces, curves, and points. Finding extreme values of a function on such a boundary is much harder than in one dimension. In some situations, such as shown in Example 6 of 14.7, one can reduce the problem of finding extreme values of a function on a boundary to a one variable calculus problem, when the
boundary can be represented as a graph of a function, so one of the variables can be substituted
in terms of the rest. But this approach has very limited usage.
The method of Lagrangian Multipliers is used to address this problem. The basic idea of Lagrange multipliers is this: if you want to maximize a function, you want to move in the direction of the gradient from wherever you are. If you are subject to a constraint (i.e. you can only move along a boundary curve or surface), you may not be able to move in the direction of the gradient while staying “in bounds” for the constraint. So you do the best you can and move along your constraint curve (or surface) in the direction “closest” to the gradient direction, namely, the orthogonal projection of the gradient of the function at that point in the tangent line/plane of the constraint. When you get to a place where this projection is zero, namely, when the gradient is orthogonal to the constraint where you are (more accurately, orthogonal to the tangent plane of the constraint), moving in any direction in the constraint won’t increase the function value. So where you are is at least a local maximum subject to the constraint.
We have two vectors at play here: the gradient of the function f whose extreme values we are looking for and the gradient of the function g whose level set is our constraint. When the gradient of is orthogonal to the tangent plane of the level set of at point , we know that is parallel to , hence the criterion of Lagrangian Multipliers. Algebraically, this relation is expressed as
Of course, the point also needs to satisfy the constraint equation .
Find the following definitions/concepts/formulas/theorems:
- constraint (you have seen this before)
- A specific region/domain of a function
- Lagrange multiplier- general concept
- A scalar multiplier of a gradient vector
- Orthogonal Gradient Theorem
- Suppose that , is differentiable in a region whose interior contains a smooth curve . If is a point on where has a local maximum or minimum relative to its values on , then is orthogonal to the curve’s tangent vector at .
- Lagrange conditions (in the book this is called The Method of Lagrange Multipliers)
- Suppose that and are differentiable and when . To find the local maximum and minimum values of subject to the constraint (if these exist), find the values of that simultaneously satisfy the equations If they exist, absolute extrema can be found by comparing these values of at each critical point satisfying the aforementioned two equations. For functions of two independent variables, the condition is similar, but without the variable .
- critical point/critical value (specifically for optimization with a constraint)
- TK
- Lagrange multipliers with two constraints
- Many problems require us to find the extreme values of a differentiable function whose variables are subject to two constraints. If the constraints are and and are differentiable, with not parallel to , we find the constrained local maxima and minima of by introducing two Lagrange multipliers and . That is, we locate the points where takes on its constrained extreme values by finding the values of that simultaneously satisfy the three equations
The first two examples are very long, because they are trying to solve a constrained optimization problem without using Lagrange multipliers. Try to follow along all the way through the examples, but understand that the main purpose of these examples is to motivate the technique about to be introduced. Lagrange multipliers take care of many of the considerations in these examples without having to think deeply into the geometry on every problem. The proof of the Orthogonal Gradient Theorem is just an application of the chain rule for paths. Examples 3 and 4 are both long, but they are about as simple as Lagrange Multiplier examples can be. The setup isn’t too bad in either example, and they perform the same steps in the same order. You should certainly try to work your way through them.
Reading Questions/Quizzes
- In the problem of finding the point on the plane that is closest to the origin, there are two functions: related to the plane and as the distance from to the origin. Which is the function to be minimized and which gives the constraint?
- Use a software tool to plot the ellipse and a set of level curves of the function . Adjust the choice of level curves to find those that are tangent to the ellipse. How do those tangent points relate to Example 3?
- In Example 3, write out the equations explicitly given by the method of Lagrange Multipliers. There should be three such equations. What are the unknowns to be solved?
- In Example 3, it is conceivable to solve for in terms of for every point on the ellipse , e.g. , then treat as a function of . What would be the domain of this pair of functions? What do you make of having to work with two functions? Can you use them to find an alternative solution to this problem? What would be the domain of this pair of functions? What do you make of having to work with two functions? Can you use them to find an alternative solution to this problem?
- Use the method of Lagrange multipliers to find
- Minimum on a hyperbola: The minimum value of , subject to the constraints .
- Maximum on a line: The maximum value of , subject to the constraint . Comment on the geometry of each solution.