Overview
Recently, we have witnessed many important advances in learning approaches for sequential decision making. These advances have occurred in different communities, who refer to the problem using different terminology: Bayesian optimization, experimental design, bandits (x-armed bandits, contextual bandits, Gaussian process bandits), active sensing, personalized recommender systems, automatic algorithm configuration, reinforcement learning and so on. These communities tend to use different methodologies too. Some focus more on practical performance while others are more concerned with theoretical aspects of the problem. As a result, they have derived and engineered a diverse range of methods for trading off exploration and exploitation in learning. For these reasons, it is timely and important to bring these communities together to identify differences and commonalities, to propose common benchmarks, to review the many practical applications (interactive user interfaces, automatic tuning of parameters and architectures, robotics, recommender systems, active vision, and more), to narrow the gap between theory and practice and to identify strategies for attacking high dimensionality.
Topics
- Bayesian optimization
- Sequential experimental design
- Bandits
- Exploration-exploitation trade-off
We welcome contributions on theoretical models, empirical studies, and applications of the above. The list is not exhaustive, and we also welcome submissions on highly related topics. Accepted papers will be made available online at the workshop website.
General Information
- Workshop date: December 16, 2011
- Location: Sierra Nevada, Spain (held at NIPS 2011)
- Submission deadline: October 21, 2011
- Organizers: Nando de Freitas, Roman Garnett, Frank Hutter and Michael Osborne
Confirmed Invited Speakers
- Andreas Krause, ETH Zurich and Caltech
- Csaba Szepesvari, University of Alberta
- Louis Dorard, University College London
- Remi Munos, INRIA Lille - Nord Europe