I'm doing a presentation on the No-Free-Lunch theorem, since I've found that the best way to learn about a topic is to try and teach it...In order to get an idea what I'm talking about, I started reading the Wolpert & Macready Paper.
In this paper, they define a trace of length $m$ as the sequence of $(x_i,y_i)$ pairs where $x_i\in X$ is some setting of your model parameters $y_i\in Y$ is some value dictating how desirable $x_i$ is according to some objective function $f$.
They then describe a "black-box search algorithm" as a function which maps traces to $X$ (yielding the next set of parameters to try.)
Since their description of a "black-box search algorithm" is strictly a function of a trace, is gradient-descent considered a black-box search algorithm? Isn't it cheating to look at the gradient? Aren't you making your algorithm a function of the objective function as well in that case?