论文标题
帕累托不变风险最小化:旨在减轻分布概括的优化困境
Pareto Invariant Risk Minimization: Towards Mitigating the Optimization Dilemma in Out-of-Distribution Generalization
论文作者
论文摘要
最近,人们对使机器学习系统能够充分推广到分布(OOD)数据的兴趣越来越大。大多数努力致力于提高优化目标,以使模型捕获基本不变性。但是,在这些OOD目标的优化过程中通常存在妥协:i)许多OOD目标必须放宽,因为经验风险最小化的惩罚条款(ERM),以便于优化,而放松的形式可以削弱原始目标的鲁棒性; ii)由于ERM和OOD目标之间的内在冲突,罚款条款还需要仔细调整罚款权重。因此,这些妥协很容易导致ERM或OOD目标的次优性能。为了解决这些问题,我们引入了多目标优化(MOO)的观点,以了解OOD优化过程,并提出了一种称为Pareto不变风险最小化(PAIR)的新优化方案。对通过与其他OOD目标进行合作优化,提高了OOD目标的鲁棒性,从而弥合了由于放松而引起的差距。然后,对接近帕累托最佳解决方案,该解决方案可以正确交易ERM并正确地实现目标。关于挑战性基准,荒野的广泛实验表明,对减轻了妥协并产生最高的OOD表现。
Recently, there has been a growing surge of interest in enabling machine learning systems to generalize well to Out-of-Distribution (OOD) data. Most efforts are devoted to advancing optimization objectives that regularize models to capture the underlying invariance; however, there often are compromises in the optimization process of these OOD objectives: i) Many OOD objectives have to be relaxed as penalty terms of Empirical Risk Minimization (ERM) for the ease of optimization, while the relaxed forms can weaken the robustness of the original objective; ii) The penalty terms also require careful tuning of the penalty weights due to the intrinsic conflicts between ERM and OOD objectives. Consequently, these compromises could easily lead to suboptimal performance of either the ERM or OOD objective. To address these issues, we introduce a multi-objective optimization (MOO) perspective to understand the OOD optimization process, and propose a new optimization scheme called PAreto Invariant Risk Minimization (PAIR). PAIR improves the robustness of OOD objectives by cooperatively optimizing with other OOD objectives, thereby bridging the gaps caused by the relaxations. Then PAIR approaches a Pareto optimal solution that trades off the ERM and OOD objectives properly. Extensive experiments on challenging benchmarks, WILDS, show that PAIR alleviates the compromises and yields top OOD performances.