Skip to content

Optimizing with Low Budgets: A Comparison on the Black-Box Optimization Benchmarking Suite and OpenAI Gym

Authors: Elena Raponi, Nathanael Carraz Rakotonirina, Jeremy Rapin, Carola Doerr, Olivier Teytaud

Published: 2023 (Journal Paper)

Source: IEEE Transactions on Evolutionary Computation

Algorithm: Low-budget black-box optimizer benchmarking

arXiv: 2310.00077

DOI: 10.1109/TEVC.2023.3346788

Summary

Compares Bayesian-optimization-oriented ML tools with classical derivative-free optimizers under limited evaluation budgets. The study updates earlier benchmarking work by testing both COCO/BBOB functions and direct policy search in OpenAI Gym, showing that BO methods are strong at small budgets but can be overtaken by other black-box optimizers as budgets grow.

Abstract

The growing ubiquity of machine learning (ML) has led it to enter various areas of computer science, including black-box optimization (BBO). Recent research is particularly concerned with Bayesian optimization (BO). BO-based algorithms are popular in the ML community, as they are used for hyperparameter optimization and more generally for algorithm configuration. However, their efficiency decreases as the dimensionality of the problem and the budget of evaluations increase. Meanwhile, derivative-free optimization methods have evolved independently in the optimization community. Therefore, we urge to understand whether cross-fertilization is possible between the two communities, ML and BBO, i.e., whether algorithms that are heavily used in ML also work well in BBO and vice versa. Comparative experiments often involve rather small benchmarks and show visible problems in the experimental setup, such as poor initialization of baselines, overfitting due to problem-specific setting of hyperparameters, and low statistical significance. With this paper, we update and extend a comparative study presented by Hutter et al. in 2013. We compare BBO tools for ML with more classical heuristics, first on the well-known BBOB benchmark suite from the COCO environment and then on Direct Policy Search for OpenAI Gym, a reinforcement learning benchmark. Our results confirm that BO-based optimizers perform well on both benchmarks when budgets are limited, albeit with a higher computational cost, while they are often outperformed by algorithms from other families when the evaluation budget becomes larger. We also show that some algorithms from the BBO community perform surprisingly well on ML tasks.

Tags

  • Black-box optimization

  • Low-budget optimization

  • Bayesian optimization

  • Derivative-free optimization

  • BBOB

  • COCO

  • OpenAI Gym

  • Benchmarking

  • Reinforcement learning