Taking the GP Out of the Loop
arXiv:2506.12818v2 Announce Type: replace-cross Abstract: Bayesian optimization (BO) has traditionally solved black-box problems where function evaluation is expensive and, therefore, observations are few. Recently, however, there has been growing interest in applying BO to problems where function evaluation is cheaper and observations are more plentiful. In this regime, scaling to many observations $N$ is impeded by Gaussian-process (GP) surrogates: GP hyperparameter fitting scales as $mathcal{O}(N^3)$ (reduced to roughly $mathcal{O}(N^2)$ in modern implementations), and it is repeated at every […]