Skip to content
On this page
On this page

No Free Lunch

Let's start by thinking about search algorithms

Notation:

  • XX is countable search space (e.g. set of parameters)
  • objective function f:XYf: X \mapsto Y that evaluates a particular choice
  • dataset dm={dXm,dYm}d^m = \{ d_X^m, d_Y^m \} is a set of mm pairs of (x,f(x))(x, f(x)) (i.e. what you've tried so far)
  • search algorithm A\mathcal{A} simply appends to this dataset (i.e. continues the search, ideally finding xx's that lead to smaller f(x)f(x))
  • arbitrary performance measure Φ:dYmR\Phi: d_Y^m \to \mathbb{R} (i.e. evaluate how good a search algorithm is)

Suppose we want to find the minimum of ff. Then, the search algorithm is exactly gradient descent, as in a proposal for the next step in the search.

Key Points:

Edit this page
Last updated on 2/23/2022