ICLR 2026 Orals

Task-free Adaptive Meta Black-box Optimization

Chao Wang, Licheng Jiao, Lingling Li, Jiaxuan Zhao, Guanchun Wang, Fang Liu, Shuyuan Yang

Reinforcement Learning & Agents Sat, Apr 25 · 3:15 PM–3:25 PM · 202 A/B Avg rating: 5.50 (2–8)

Abstract

Handcrafted optimizers become prohibitively inefficient for complex black-box optimization (BBO) tasks. MetaBBO addresses this challenge by meta-learning to automatically configure optimizers for low-level BBO tasks, thereby eliminating heuristic dependencies. However, existing methods typically require extensive handcrafted training tasks to learn meta-strategies that generalize to target tasks, which poses a critical limitation for realistic applications with unknown task distributions. To overcome the issue, we propose the Adaptive meta Black-box Optimization Model (ABOM), which performs online parameter adaptation using solely optimization data from the target task, obviating the need for predefined task distributions. Unlike conventional metaBBO frameworks that decouple meta-training and optimization phases, ABOM introduces a closed-loop adaptive parameter learning mechanism, where parameterized evolutionary operators continuously self-update by leveraging generated populations during optimization. This paradigm shift enables zero-shot optimization: ABOM achieves competitive performance on synthetic BBO benchmarks and realistic unmanned aerial vehicle path planning problems without any handcrafted training tasks. Visualization studies reveal that parameterized evolutionary operators exhibit statistically significant search patterns, including natural selection and genetic recombination.

One-sentence summary·Auto-generated by claude-haiku-4-5-20251001(?)

ABOM performs task-free adaptive meta black-box optimization using online parameter adaptation without predefined task distributions.

Contributions·Auto-generated by claude-haiku-4-5-20251001(?)
  • Introduces closed-loop adaptive parameter learning mechanism where evolutionary operators continuously self-update during optimization
  • Eliminates dependency on handcrafted training tasks by performing online adaptation using only target task optimization data
  • Enables zero-shot optimization through parameterized evolutionary operators updated via gradient descent
Methods used·Auto-generated by claude-haiku-4-5-20251001(?)
  • Meta black-box optimization
  • Evolutionary algorithms
  • Online parameter adaptation
Datasets used·Auto-generated by claude-haiku-4-5-20251001(?)
  • Synthetic benchmarks
  • UAV path planning
Limitations (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)
  • Cubic computational bottleneck O(d3) from attention mechanisms
    from the paper
  • Limited exploration of hybrid training paradigms integrating pretraining with online adaptation
    from the paper
  • Convergence rate analysis needed for theoretical examination of adaptive parameter learning
    from the paper
Future work (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)
  • Address cubic computational bottleneck through sparse or low-rank attention mechanisms
    from the paper
  • Dynamically adapt population size and model capacity during optimization
    from the paper
  • Conduct convergence rate analysis for adaptive parameter learning
    from the paper
  • Explore hybrid training paradigms integrating pretraining with online adaptation
    from the paper

Author keywords

  • Meta Black-box Optimization
  • Evolutionary Algorithms

Related orals

Something off? Let us know →