RESEARCH

Differentiable Expected Hypervolume Improvement for Parallel Multi-Objective Bayesian Optimization

December 10, 2020

Abstract

In many real-world scenarios, decision makers seek to efficiently optimize multiple competing objectives in a sample-efficient fashion. Multi-objective Bayesian optimization (BO) is a common approach, but many of the best-performing acquisition functions do not have known analytic gradients and suffer from high computational overhead. We leverage recent advances in programming models and hardware acceleration for multi-objective BO using Expected Hypervolume Improvement (EHVI)---an algorithm notorious for its high computational complexity. We derive a novel formulation of q-Expected Hypervolume Improvement (qEHVI), an acquisition function that extends EHVI to the parallel, constrained evaluation setting. qEHVI is an exact computation of the joint EHVI of q new candidate points (up to Monte-Carlo (MC) integration error). Whereas previous EHVI formulations rely on gradient-free acquisition optimization or approximated gradients, we compute exact gradients of the MC estimator via auto-differentiation, thereby enabling efficient and effective optimization using first-order and quasi-second-order methods. Our empirical evaluation demonstrates that qEHVI is computationally tractable in many practical scenarios and outperforms state-of-the-art multi-objective BO algorithms at a fraction of their wall time.

Download the Paper

AUTHORS

Written by

Samuel Daulton

Maximilian Balandat

Eytan Bakshy

Publisher

Neural Information Processing Systems (NeurIPS 2020)

Research Topics

Core Machine Learning

Related Publications

June 02, 2019

Simple Attention-Based Representation Learning for Ranking Short Social Media Posts | Facebook AI Research

This paper explores the problem of ranking short social media posts with respect to user queries using neural networks. Instead of starting with a complex architecture, we proceed from the bottom up and examine the effectiveness of a simple,…

Peng Shi, Jinfeng Rao, Jimmy Lin

June 02, 2019

June 09, 2019

THEORY

First-order Adversarial Vulnerability of Neural Networks and Input Dimension | Facebook AI Research

Over the past few years, neural networks were proven vulnerable to adversarial images: targeted but imperceptible image perturbations lead to drastically different predictions. We show that adversarial vulnerability increases with the gradients…

Carl-Johann Simon-Gabriel, Yann Ollivier, Bernhard Scholkopf, Leon Bottou, David Lopez-Paz

June 09, 2019

May 31, 2019

INTEGRITY

Abusive Language Detection with Graph Convolutional Networks | Facebook AI Research

Abuse on the Internet represents a significant societal problem of our time. Previous research on automated abusive language detection in Twitter has shown that community-based profiling of users is a promising technique for this task. However,…

Pushkar Mishra, Marco Del Tredici, Helen Yannakoudakis, Ekaterina Shutova

May 31, 2019

June 01, 2019

Probabilistic Planning with Reduced Models | Facebook AI Research

Reduced models are simplified versions of a given domain, designed to accelerate the planning process. Interest in reduced models has grown since the surprising success of determinization in the first international probabilistic planning…

Luis Pineda, Shlomo Zilberstein

June 01, 2019

Help Us Pioneer The Future of AI

We share our open source frameworks, tools, libraries, and models for everything from research exploration to large-scale production deployment.