论文标题

结构保存深度学习

Structure preserving deep learning

论文作者

Celledoni, Elena, Ehrhardt, Matthias J., Etmann, Christian, McLachlan, Robert I, Owren, Brynjulf, Schönlieb, Carola-Bibiane, Sherry, Ferdia

论文摘要

在过去的几年中,深度学习一直升至前景,这是一个引起人们关注的话题,这主要是由于解决了大规模图像处理任务而获得的成功。应用深度学习涉及多个具有挑战性的数学问题:大多数深度学习方法需要解决硬性优化问题,并且需要对计算工作,数据量和模型复杂性之间的权衡有很好的了解,以成功地为给定问题设计深度学习方法。在深度学习中取得的很大进步是基于启发式探索,但是越来越多的努力来数学上了解现有深度学习方法中的结构,并系统地设计新的深度学习方法,以保留深度学习中的某些类型的结构。在本文中,我们回顾了许多这些方向:可以将一些深神经网络理解为动态系统的离散性,可以设计神经网络具有理想的属性,例如可逆性或群体等效性,以及基于新算法的算法框架,基于同质的汉密尔顿系统和Riemann Systems和Riemannian歧管,以解决优化问题。我们通过讨论一些我们认为是未来研究的有趣方向的开放问题来结束对这些主题的回顾。

Over the past few years, deep learning has risen to the foreground as a topic of massive interest, mainly as a result of successes obtained in solving large-scale image processing tasks. There are multiple challenging mathematical problems involved in applying deep learning: most deep learning methods require the solution of hard optimisation problems, and a good understanding of the tradeoff between computational effort, amount of data and model complexity is required to successfully design a deep learning approach for a given problem. A large amount of progress made in deep learning has been based on heuristic explorations, but there is a growing effort to mathematically understand the structure in existing deep learning methods and to systematically design new deep learning methods to preserve certain types of structure in deep learning. In this article, we review a number of these directions: some deep neural networks can be understood as discretisations of dynamical systems, neural networks can be designed to have desirable properties such as invertibility or group equivariance, and new algorithmic frameworks based on conformal Hamiltonian systems and Riemannian manifolds to solve the optimisation problems have been proposed. We conclude our review of each of these topics by discussing some open problems that we consider to be interesting directions for future research.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源