论文标题
多阶段转移学习与选择过程的应用
Multi-Stage Transfer Learning with an Application to Selection Process
论文作者
论文摘要
在多阶段过程中,决策以有序的阶段顺序进行。他们中的许多人都有双重漏斗问题的结构:随着样本量从一个阶段到另一个阶段的减小,信息会增加。一个相关的例子是一个选择过程,申请人申请职位,奖励或赠款。在每个阶段,对更多的申请人进行了评估和过滤,从其余的申请人中收集了更多信息。在最后阶段,决策者使用所有可用信息来做出最终决定。训练每个阶段的分类器变得不切实际,因为由于早期阶段的尺寸较低或由于后一个阶段的样本量较小而导致的尺寸较低。在这项工作中,我们提出了一个\ textIt {多阶段转移学习}(MSGTL)方法,该方法使用了在早期阶段接受培训的简单分类器的知识来提高分类器在后期阶段的性能。通过从在较大数据集中训练的简单神经网络中转移权重,我们能够在后一个阶段微调更复杂的神经网络,而不会因样本量较小而过度拟合。我们表明,可以使用简单的概率地图来控制维护知识和微调之间的权衡。使用现实世界数据的实验证明了我们方法的功效,因为它表现出了转移学习和正则化的其他最新方法。
In multi-stage processes, decisions happen in an ordered sequence of stages. Many of them have the structure of dual funnel problem: as the sample size decreases from one stage to the other, the information increases. A related example is a selection process, where applicants apply for a position, prize, or grant. In each stage, more applicants are evaluated and filtered out, and from the remaining ones, more information is collected. In the last stage, decision-makers use all available information to make their final decision. To train a classifier for each stage becomes impracticable as they can underfit due to the low dimensionality in early stages or overfit due to the small sample size in the latter stages. In this work, we proposed a \textit{Multi-StaGe Transfer Learning} (MSGTL) approach that uses knowledge from simple classifiers trained in early stages to improve the performance of classifiers in the latter stages. By transferring weights from simpler neural networks trained in larger datasets, we able to fine-tune more complex neural networks in the latter stages without overfitting due to the small sample size. We show that it is possible to control the trade-off between conserving knowledge and fine-tuning using a simple probabilistic map. Experiments using real-world data demonstrate the efficacy of our approach as it outperforms other state-of-the-art methods for transfer learning and regularization.