Table of Contents
Open Table of Contents
Description
Algorithms to find the minimum of a function, where order denotes the number of derivatives used.
- First-order methods: gradient-based.
- Zero-order methods: derivative-free.
Cost Functions
Let
where is a scalar and is an -dimensional vector.
Example:
Typical optimization problem:
Optimality Conditions
-
is a global minimum if
-
is a global maximum if
A Naïve Approach
Evaluate at many points (uniformly or randomly sampled) and choose the smallest value.
- Problem: in high dimensions, sampling requires exponentially more points to cover the same volume (the “curse of dimensionality”).
General Framework
- Start at (random selection in domain).
- Find a descent direction .
- Update: .
- Repeat until convergence.
- Exploration vs. exploitation tradeoff:
- Start with large steps to explore.
- Gradually shrink step size to converge.
Example: Random Search
At step , pick random directions .
-
Evaluate
-
If
then update
.
- Variants: coordinate search vs. descent search.