论文标题

算法配置的精制边界:双类近似性的刀具边缘

Refined bounds for algorithm configuration: The knife-edge of dual class approximability

论文作者

Balcan, Maria-Florina, Sandholm, Tuomas, Vitercik, Ellen

论文摘要

随着算法带有越来越多的可调参数,自动化算法配置的增长越来越必要。使用机器学习来调整参数,优化性能指标,例如运行时和解决方案质量。培训集由手头特定领域的问题实例组成。我们研究了一个有关这些技术的基本问题:培训设置应该多大以确保参数在训练集中的平均经验表现接近其预期的未来表现?我们为表现出广泛应用结构的算法配置问题回答了这个问题:算法的性能作为其参数的函数可以通过“简单”功能近似。我们表明,如果此近似值在L-内度标准下存在,则可以提供强大的样本复杂性界限。另一方面,如果仅在p小于无穷大的L-P标准下近似保留,则不可能在最坏情况下提供有意义的样品复杂性界限。我们在Integer编程的背景下经验评估了我们的界限,这是计算机科学中最强大的工具之一。通过实验,我们获得的样品复杂性边界比以前最著名的边界小700倍。

Automating algorithm configuration is growing increasingly necessary as algorithms come with more and more tunable parameters. It is common to tune parameters using machine learning, optimizing performance metrics such as runtime and solution quality. The training set consists of problem instances from the specific domain at hand. We investigate a fundamental question about these techniques: how large should the training set be to ensure that a parameter's average empirical performance over the training set is close to its expected, future performance? We answer this question for algorithm configuration problems that exhibit a widely-applicable structure: the algorithm's performance as a function of its parameters can be approximated by a "simple" function. We show that if this approximation holds under the L-infinity norm, we can provide strong sample complexity bounds. On the flip side, if the approximation holds only under the L-p norm for p smaller than infinity, it is not possible to provide meaningful sample complexity bounds in the worst case. We empirically evaluate our bounds in the context of integer programming, one of the most powerful tools in computer science. Via experiments, we obtain sample complexity bounds that are up to 700 times smaller than the previously best-known bounds.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源