论文标题

端到端预测和优化过程的风险保证

Risk Guarantees for End-to-End Prediction and Optimization Processes

论文作者

Ho-Nguyen, Nam, Kılınç-Karzan, Fatma

论文摘要

预测模型通常用于估计优化模型的参数。尽管在端到端的视图中,真正的目标是实现良好的优化性能,但预测性能是自行衡量的。虽然通常认为在估计参数方面的良好预测性能将导致良好的随后优化性能,但对此有正式的理论保证显然缺乏。在本文中,我们探讨了使我们能够明确描述预测性能如何控制优化性能的条件。我们的弱条件允许产生渐近收敛的结果,而我们更强的条件可以准确地量化预测性能的优化性能。通常,对这些条件的验证是一项非平凡的任务。然而,我们表明我们的状况较弱与学习理论文献中著名的Fisher一致性概念相当。然后,这使我们能够轻松检查较弱的状况是否有多个损失功能。我们还确定平方误差函数满足了我们更强的状况。因此,我们得出了用平方损耗测量的预测性能与一类对称损耗函数与随后的优化性能之间的确切理论关系。在一项有关投资组合优化,分数背包和多类分类问题的计算研究中,我们比较了使用多种预测损失函数的优化性能(有些是Fisher一致的,有些是不一致的),并证明缺乏损失函数的一致性确实可以对性能产生不利影响。

Prediction models are often employed in estimating parameters of optimization models. Despite the fact that in an end-to-end view, the real goal is to achieve good optimization performance, the prediction performance is measured on its own. While it is usually believed that good prediction performance in estimating the parameters will result in good subsequent optimization performance, formal theoretical guarantees on this are notably lacking. In this paper, we explore conditions that allow us to explicitly describe how the prediction performance governs the optimization performance. Our weaker condition allows for an asymptotic convergence result, while our stronger condition allows for exact quantification of the optimization performance in terms of the prediction performance. In general, verification of these conditions is a non-trivial task. Nevertheless, we show that our weaker condition is equivalent to the well-known Fisher consistency concept from the learning theory literature. This then allows us to easily check our weaker condition for several loss functions. We also establish that the squared error loss function satisfies our stronger condition. Consequently, we derive the exact theoretical relationship between prediction performance measured with the squared loss, as well as a class of symmetric loss functions, and the subsequent optimization performance. In a computational study on portfolio optimization, fractional knapsack and multiclass classification problems, we compare the optimization performance of using of several prediction loss functions (some that are Fisher consistent and some that are not) and demonstrate that lack of consistency of the loss function can indeed have a detrimental effect on performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源