Timezone: »
This article develops a Bayesian optimization (BO) method which acts directly over raw strings, proposing the first uses of string kernels and genetic algorithms within BO loops. Recent applications of BO over strings have been hindered by the need to map inputs into a smooth and unconstrained latent space. Learning this projection is computationally and data-intensive. Our approach instead builds a powerful Gaussian process surrogate model based on string kernels, naturally supporting variable length inputs, and performs efficient acquisition function maximization for spaces with syntactic constraints. Experiments demonstrate considerably improved optimization over existing approaches across a broad range of constraints, including the popular setting where syntax is governed by a context-free grammar.
Author Information
Henry Moss (Lancaster University)
David Leslie (Lancaster University and PROWLER.io)
Daniel Beck (University of Melbourne)
Javier González (Microsoft Research Cambridge)
Paul Rayson (Lancaster University)
Related Events (a corresponding poster, oral, or spotlight)
-
2020 Poster: BOSS: Bayesian Optimization over String Spaces »
Wed Dec 9th 05:00 -- 07:00 PM Room Poster Session 3
More from the Same Authors
-
2020 Poster: Multi-task Causal Learning with Gaussian Processes »
Virginia Aglietti · Theodoros Damoulas · Mauricio Álvarez · Javier González -
2019 Poster: Meta-Surrogate Benchmarking for Hyperparameter Optimization »
Aaron Klein · Zhenwen Dai · Frank Hutter · Neil Lawrence · Javier González -
2018 Poster: Bandit Learning in Concave N-Person Games »
Mario Bravo · David Leslie · Panayotis Mertikopoulos