If you have a few years of experience in Computer Science or research, and you’re interested in sharing that experience with the community, have a look at our **Contribution Guidelines**.

# What Is Constrained Optimization?

Last modified: October 21, 2022

## 1. Introduction

Constrained optimization, also known as constraint optimization, is the process of **optimizing an objective function with respect to a set of decision variables while imposing constraints on those variables**.

In this tutorial, we’ll provide a brief introduction to constrained optimization, explore some examples, and introduce some methods to solve constrained optimization problems.

## 2. Constrained Optimization

In a constrained optimization problem, the objective function is either a cost function (to be minimized) or a reward/utility function (to be maximized).

Now, let’s examine situations where a function is constrained. **A constraint is a hard limit we place on a variable’s value** to prevent the objective function from going forever in certain directions:

### 2.1. Optimum’s Location

With nonlinear functions, the optima (maximum or minimum values) can either occur at the boundaries or between them:

As we can see, the maximum or minimum values can occur in the interior or at boundaries.

In contrast, with linear functions, the maximum or minimum values occur only at boundaries:

## 3. Forms of Constrained Optimization Problems

The following formulation describes the general (canonical) form of a constrained minimization problem:

where is the vector of decision variables; is the objective function that needs to be optimized; for and for are constraints that are required to be satisfied.

If we need to maximize the objective function, e.g., the utilization efficiency of network resources, we can negate it to make it fit the general form:

### 3.1. Types of Constrained Optimization Problems

Depending on the objective function and constraints, there are several types of constrained optimization problems.

**Linear Programming (LP) covers the cases in which the objective function is linear and all constraints are also linear.** If all decision variables are integers, then we’ll have an Integer Linear Programming (ILP) problem. If only some of the decision variables are integers (the remaining are continuous), that’s a Mixed Integer Linear Programming (MILP) problem.

For example:

**Nonlinear Programming (NLP) refers to the problems where** **the objective function or some constraints are nonlinear.** Similar to linear programming, there are Integer Nonlinear Programming (INLP) and Mixed Integer Nonlinear Programming (MINLP).

Here’s an example of a problem with nonlinear constraints:

**Finally, Quadratic Programming (QP) problems are those with linear constraints but the objective function is quadratic.** Similar to linear programming and nonlinear programming problems, we also have Integer Quadratic Programming (IQP) and Mixed Integer Quadratic Programming (MIQP) problems.

For instance:

## 4. Real-World Example

Now let’s see how we can apply constrained optimization to real-world problems. Let’s suppose we want to maximize the area of a rectangle while keeping its perimeter equal to a predefined value :

As we see, the area is , and that will be our objective function. Further, since the perimeter needs to be equal to , we have a constrained optimization problem:

This problem has a quadratic objective function () and a linear constraint, so it falls into the QP category.

## 5. Solving Constrained Problems

Let’s now see how we can solve constrained optimization problems.

### 5.1. Mathematical Methods

**There is, unfortunately, no universal method for solving all types of optimization problems.**

Different proposed methods are tailored to particular types of optimization problems. The following table lists some classical mathematical methods for solving constrained optimization problems:

Most of these methods, such as the branch-and-bound algorithm, are exact. Exact optimization techniques are guaranteed to find the optimal solution. The price for such guarantees is that they can take a lot of time to complete. In contrast, inexact methods don’t guarantee finding the optimum but usually produce a sufficiently good solution quickly. Metaheuristics such as the Grey-Wolf Optimization algorithm fall into the category of inexact methods and are frequently used when we need to find any solution quickly.

### 5.2. The Substitution Method

Let’s explain the substitution method using a simple example with an equality constraint:

As we see, that’s the same problem as in the above example with a rectangle. The only difference is that here, isn’t a free parameter but a known value :

**Since we have a simple equality constraint, we can use the substitution method**. We use this method mostly for problems with an equality constraint from which it’s easy to express one variable in terms of others. If we’re lucky, we can substitute the variable in the objective function and get a form that isn’t hard to solve by hand.

In our example, the constraint gives us a simple substitution:

As a result, we can rewrite the objective function:

The maximum value of should satisfy the first-order derivative condition:

Consequently, , and the maximum value of is .

### 5.3. Optimization Solvers

The following table summarizes some well-known software tools for constrained optimization problems**:**

**They all perform built-in exact methods (e.g., simplex) and usually combine them with inexact algorithms to reach a solution faster**. For instance, CPLEX uses a node heuristic along with the branch-and-cut algorithm.

## 6. Conclusion

In this article, we covered the principles of constrained optimization. There are several types of constrained optimization problems and many methods for solving them. Which one is the best depends on the particular problem at hand.