Timezone: »

Bayesian optimization, experimental design and bandits: Theory and applications
Nando de Freitas · Roman Garnett · Frank R Hutter · Michael A Osborne

Thu Dec 15 10:30 PM -- 11:00 AM (PST) @ Melia Sierra Nevada: Hotel Bar
Event URL: http://www.cs.ubc.ca/~hutter/nips2011workshop/index.html »
Recently, we have witnessed many important advances in learning approaches for sequential decision making. These advances have occurred in different communities, who refer to the problem using different terminology: Bayesian optimization, experimental design, bandits ($$x$$-armed bandits, contextual bandits, Gaussian process bandits), active sensing, personalized recommender systems, automatic algorithm configuration, reinforcement learning and so on. These communities tend to use different methodologies too. Some focus more on practical performance while others are more concerned with theoretical aspects of the problem. As a result, they have derived and engineered a diverse range of methods for trading off exploration and exploitation in learning. For these reasons, it is timely and important to bring these communities together to identify differences and commonalities, to propose common benchmarks, to review the many practical applications (interactive user interfaces, automatic tuning of parameters and architectures, robotics, recommender systems, active vision, and more), to narrow the gap between theory and practice and to identify strategies for attacking high dimensionality.

Author Information

Nando de Freitas (University of Oxford)
Roman Garnett (Washington University in St. Louis)
Frank R Hutter (Freiburg University)
Michael A Osborne (U Oxford)

More from the Same Authors