论文标题

重新思考防止基于利润的损失的公制学习中的集体批量

Rethinking preventing class-collapsing in metric learning with margin-based losses

论文作者

Levi, Elad, Xiao, Tete, Wang, Xiaolong, Darrell, Trevor

论文摘要

公制学习寻求感知嵌入,在视觉上相似的实例很接近并且不同的实例是分开的,但是当阶级样本的分布不同并且存在独特的子群体时,学到的表示形式可以是次优的。尽管从理论上讲,具有最佳假设,但基于利润的损失(例如三胞胎损失和保证金损失)具有多样的解决方案。从理论上讲,我们证明并从经验上表明,在合理的噪声假设下,基于保证金的损失倾向于将各种模式的所有样本投射到嵌入空间中的单个点上,从而导致类倒塌,通常会导致空间不足以进行分类或检索。为了解决这个问题,我们对嵌入损失进行了简单的修改,以便每个样本在批处理中选择其最接近的同一级对应物作为元组中的正元素。这允许每个类中存在多个子群体。适应性可以集成到广泛的度量学习损失中。提出的抽样方法证明了各种现有损失对各种细粒图像检索数据集的明显好处;定性检索结果表明,具有相似视觉模式的样品在嵌入空间中确实更近。

Metric learning seeks perceptual embeddings where visually similar instances are close and dissimilar instances are apart, but learned representations can be sub-optimal when the distribution of intra-class samples is diverse and distinct sub-clusters are present. Although theoretically with optimal assumptions, margin-based losses such as the triplet loss and margin loss have a diverse family of solutions. We theoretically prove and empirically show that under reasonable noise assumptions, margin-based losses tend to project all samples of a class with various modes onto a single point in the embedding space, resulting in a class collapse that usually renders the space ill-sorted for classification or retrieval. To address this problem, we propose a simple modification to the embedding losses such that each sample selects its nearest same-class counterpart in a batch as the positive element in the tuple. This allows for the presence of multiple sub-clusters within each class. The adaptation can be integrated into a wide range of metric learning losses. The proposed sampling method demonstrates clear benefits on various fine-grained image retrieval datasets over a variety of existing losses; qualitative retrieval results show that samples with similar visual patterns are indeed closer in the embedding space.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源