论文标题
通过拓扑排序进行因果发现的扩散模型
Diffusion Models for Causal Discovery via Topological Ordering
论文作者
论文摘要
从观察数据中发现因果关系是可能的,例如考虑将功能关系视为与添加剂噪声(ANM)的非线性相关的函数关系。即使有很强的假设,因果发现也涉及在有向无环图(DAG)的空间上的昂贵搜索问题。 \ emph {拓扑排序}方法通过搜索置换而不是图形空间来减少因果发现的优化空间。对于ANM,数据对数似然的\ emph {hessian}可用于在因果图中查找叶子节点,从而允许其拓扑排序。但是,获得Hessian的现有计算方法仍然没有随着变量的数量和样本数量的增加而扩展。因此,受到扩散概率模型(DPM)的最新创新的启发,我们提出\ emph {diffan} \ footNote {实现可在\ url {https://github.com/vios.com/vios.com/vios.com/vios-s/diffan}。}中,用于学习Algorithm的拓扑订购,用于学习loeveragian dpms dpms dpms dpms dpms。我们介绍了在不重新培训神经网络的情况下更新学习的Hessian的理论,我们表明,使用样本子集的计算可以准确地近似订购,从而可以将缩放到具有更多示例和变量的数据集中。我们从经验上表明,我们的方法量表非常适合到具有高达$ 500 $节点的数据集,最高$ 10^5 $样本,同时仍在使用最先进的因果发现方法的小型数据集上执行PAR。实施可在https://github.com/vios-s/diffan上获得。
Discovering causal relations from observational data becomes possible with additional assumptions such as considering the functional relations to be constrained as nonlinear with additive noise (ANM). Even with strong assumptions, causal discovery involves an expensive search problem over the space of directed acyclic graphs (DAGs). \emph{Topological ordering} approaches reduce the optimisation space of causal discovery by searching over a permutation rather than graph space. For ANMs, the \emph{Hessian} of the data log-likelihood can be used for finding leaf nodes in a causal graph, allowing its topological ordering. However, existing computational methods for obtaining the Hessian still do not scale as the number of variables and the number of samples increase. Therefore, inspired by recent innovations in diffusion probabilistic models (DPMs), we propose \emph{DiffAN}\footnote{Implementation is available at \url{https://github.com/vios-s/DiffAN} .}, a topological ordering algorithm that leverages DPMs for learning a Hessian function. We introduce theory for updating the learned Hessian without re-training the neural network, and we show that computing with a subset of samples gives an accurate approximation of the ordering, which allows scaling to datasets with more samples and variables. We show empirically that our method scales exceptionally well to datasets with up to $500$ nodes and up to $10^5$ samples while still performing on par over small datasets with state-of-the-art causal discovery methods. Implementation is available at https://github.com/vios-s/DiffAN .