论文标题
与梯度匹配的数据集冷凝
Dataset Condensation with Gradient Matching
论文作者
论文摘要
由于许多领域的最新机器学习方法依赖于较大的数据集,因此在其上存储数据集和培训模型变得更加昂贵。本文提出了一种用于数据效率学习的培训集合技术,称为数据集凝结,该技术学会了将大数据集凝结到一小部分信息的合成样本中,以从Scratch培训深层神经网络。我们将此目标提出为梯度匹配问题,在对原始数据和我们的合成数据训练的深神经网络权重之间的梯度之间。我们严格评估其在几个计算机视觉基准测试中的性能,并证明它的表现明显优于最先进的方法。最后,我们探讨了我们在持续学习和神经体系结构搜索中使用方法的使用,并报告有限的内存和计算时有希望的收益。
As the state-of-the-art machine learning methods in many fields rely on larger datasets, storing datasets and training models on them become significantly more expensive. This paper proposes a training set synthesis technique for data-efficient learning, called Dataset Condensation, that learns to condense large dataset into a small set of informative synthetic samples for training deep neural networks from scratch. We formulate this goal as a gradient matching problem between the gradients of deep neural network weights that are trained on the original and our synthetic data. We rigorously evaluate its performance in several computer vision benchmarks and demonstrate that it significantly outperforms the state-of-the-art methods. Finally we explore the use of our method in continual learning and neural architecture search and report promising gains when limited memory and computations are available.