论文标题
调整后的累积遗憾最小化噪声贝叶斯优化的预期改善
Adjusted Expected Improvement for Cumulative Regret Minimization in Noisy Bayesian Optimization
论文作者
论文摘要
预期改进(EI)是贝叶斯优化(BO)最受欢迎的采集功能之一,并且在许多应用中表现出良好的经验表现,以最大程度地减少简单的遗憾。但是,在累积遗憾的评估指标下,EI的表现可能没有竞争力,其现有的理论遗憾上限仍然有改进的余地。为了适应EI在累积遗憾下以更好的性能,我们引入了一种称为评估成本的新颖数量,该数量与采集函数进行了比较,并因此发展了预期的改善成本(EIC)算法。在EIC的每次迭代中,只有当该值超过其评估成本时,才对具有最大采集功能值的新点进行采样。如果没有符合此条件,那么当前的最佳点将被重新采样。这种评估成本量化了采样点的潜在缺点,这在累积后悔指标下很重要,因为每次迭代的目标函数值都会影响性能度量。从理论上讲,我们基于最大信息增益建立了EIC的高概率遗憾上限,这比现有基于EI的算法的界限更紧。它也可以与其他流行的BO算法(例如Thompson采样(GP-TS)和上置信度结合(GP-UCB)等其他流行的BO算法的遗憾相提并论。我们进一步执行实验,以说明EIC对几种流行的BO算法的改善。
The expected improvement (EI) is one of the most popular acquisition functions for Bayesian optimization (BO) and has demonstrated good empirical performances in many applications for the minimization of simple regret. However, under the evaluation metric of cumulative regret, the performance of EI may not be competitive, and its existing theoretical regret upper bound still has room for improvement. To adapt the EI for better performance under cumulative regret, we introduce a novel quantity called the evaluation cost which is compared against the acquisition function, and with this, develop the expected improvement-cost (EIC) algorithm. In each iteration of EIC, a new point with the largest acquisition function value is sampled, only if that value exceeds its evaluation cost. If none meets this criteria, the current best point is resampled. This evaluation cost quantifies the potential downside of sampling a point, which is important under the cumulative regret metric as the objective function value in every iteration affects the performance measure. We establish in theory a high-probability regret upper bound of EIC based on the maximum information gain, which is tighter than the bound of existing EI-based algorithms. It is also comparable to the regret bound of other popular BO algorithms such as Thompson sampling (GP-TS) and upper confidence bound (GP-UCB). We further perform experiments to illustrate the improvement of EIC over several popular BO algorithms.