ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.04261
53
0

Simulating, Fast and Slow: Learning Policies for Black-Box Optimization

6 June 2024
F. V. Massoli
Tim Bakker
Thomas M. Hehn
Tribhuvanesh Orekondy
Arash Behboodi
ArXivPDFHTML
Abstract

In recent years, solving optimization problems involving black-box simulators has become a point of focus for the machine learning community due to their ubiquity in science and engineering. The simulators describe a forward process fsim:(ψ,x)→yf_{\mathrm{sim}}: (\psi, x) \rightarrow yfsim​:(ψ,x)→y from simulation parameters ψ\psiψ and input data xxx to observations yyy, and the goal of the optimization problem is to find parameters ψ\psiψ that minimize a desired loss function. Sophisticated optimization algorithms typically require gradient information regarding the forward process, fsimf_{\mathrm{sim}}fsim​, with respect to the parameters ψ\psiψ. However, obtaining gradients from black-box simulators can often be prohibitively expensive or, in some cases, impossible. Furthermore, in many applications, practitioners aim to solve a set of related problems. Thus, starting the optimization ``ab initio", i.e. from scratch, each time might be inefficient if the forward model is expensive to evaluate. To address those challenges, this paper introduces a novel method for solving classes of similar black-box optimization problems by learning an active learning policy that guides a differentiable surrogate's training and uses the surrogate's gradients to optimize the simulation parameters with gradient descent. After training the policy, downstream optimization of problems involving black-box simulators requires up to ∼\sim∼90\% fewer expensive simulator calls compared to baselines such as local surrogate-based approaches, numerical optimization, and Bayesian methods.

View on arXiv
Comments on this paper