28
0

Bencher: Simple and Reproducible Benchmarking for Black-Box Optimization

Main:4 Pages
2 Figures
Bibliography:3 Pages
Abstract

We present Bencher, a modular benchmarking framework for black-box optimization that fundamentally decouples benchmark execution from optimization logic. Unlike prior suites that focus on combining many benchmarks in a single project, Bencher introduces a clean abstraction boundary: each benchmark is isolated in its own virtual Python environment and accessed via a unified, version-agnostic remote procedure call (RPC) interface. This design eliminates dependency conflicts and simplifies the integration of diverse, real-world benchmarks, which often have complex and conflicting software requirements. Bencher can be deployed locally or remotely via Docker or on high-performance computing (HPC) clusters via Singularity, providing a containerized, reproducible runtime for any benchmark. Its lightweight client requires minimal setup and supports drop-in evaluation of 80 benchmarks across continuous, categorical, and binary domains.

View on arXiv
@article{papenmeier2025_2505.21321,
  title={ Bencher: Simple and Reproducible Benchmarking for Black-Box Optimization },
  author={ Leonard Papenmeier and Luigi Nardi },
  journal={arXiv preprint arXiv:2505.21321},
  year={ 2025 }
}
Comments on this paper