论文标题

温暖的启动CMA-E用于超参数优化

Warm Starting CMA-ES for Hyperparameter Optimization

论文作者

Nomura, Masahiro, Watanabe, Shuhei, Akimoto, Youhei, Ozaki, Yoshihiko, Onishi, Masaki

论文摘要

为黑盒优化(BBO)配制的超参数优化(HPO)被认为对于机器学习方法的自动化和高性能至关重要。 CMA-ES是一种具有高度并行性的有前途的BBO方法,并且已应用于HPO任务,通常是在平行实现下,并且表现出与包括贝叶斯优化(BO)在内的其他方法的卓越性能。但是,如果高参数评估的预算受到严重限制,对于不值得并行计算的最终用户通常是这种情况,则CMA-ES会耗尽预算,而不会改善其长期适应阶段的绩效,从而超过了BO方法的表现。为了解决这个问题,我们建议通过CMA-E的初始化转移有关类似HPO任务的先验知识,从而大大缩短适应时间。知识转移是基于任务相似性的新定义而设计的,在综合问题上证实了所提出方法的性能的相关性。提出的温暖起始CMA-ES(称为WS-CMA-Es)应用于具有一些先验知识的不同HPO任务,显示其优于原始CMA-ES的性能以及在不使用先验知识的情况下进行的BO接近。

Hyperparameter optimization (HPO), formulated as black-box optimization (BBO), is recognized as essential for automation and high performance of machine learning approaches. The CMA-ES is a promising BBO approach with a high degree of parallelism, and has been applied to HPO tasks, often under parallel implementation, and shown superior performance to other approaches including Bayesian optimization (BO). However, if the budget of hyperparameter evaluations is severely limited, which is often the case for end users who do not deserve parallel computing, the CMA-ES exhausts the budget without improving the performance due to its long adaptation phase, resulting in being outperformed by BO approaches. To address this issue, we propose to transfer prior knowledge on similar HPO tasks through the initialization of the CMA-ES, leading to significantly shortening the adaptation time. The knowledge transfer is designed based on the novel definition of task similarity, with which the correlation of the performance of the proposed approach is confirmed on synthetic problems. The proposed warm starting CMA-ES, called WS-CMA-ES, is applied to different HPO tasks where some prior knowledge is available, showing its superior performance over the original CMA-ES as well as BO approaches with or without using the prior knowledge.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源