论文标题

深刻的细心信念传播:整合推理和学习以解决约束优化问题

Deep Attentive Belief Propagation: Integrating Reasoning and Learning for Solving Constraint Optimization Problems

论文作者

Deng, Yanchen, Kong, Shufeng, Liu, Caihua, An, Bo

论文摘要

信仰传播(BP)是针对图形模型的各种推理任务的重要消息算法,包括解决约束优化问题(COPS)。已经表明,BP可以通过在发送新消息(即抑制作用)之前混合新消息来实现各种基准测试的最先进性能。但是,对BP调整静态阻尼因子的现有方法不仅在繁琐,而且损害了其性能。此外,现有的BP算法在撰写新消息时同样处理每个变量节点的邻居,这也限制了其探索能力。为了解决这些问题,我们在消息通讯框架内无缝集成了BP,封闭式复发单元(GRU)和图形注意网络(GAT),以推理动态权重和阻尼因子,以组成新的BP消息。我们的模型,深切的信念传播(DABP),将因子图和每个迭代中的BP消息作为输入中的因子图和BP消息,并通过Grus和Gats渗透最佳权重和阻尼因子,然后是多头注意力层。此外,与现有的基于神经的BP变体不同,我们提出了一种新颖的DABP的自我监督学习算法,其解决方案成本不需要昂贵的培训标签,并且还可以通过有效的在线学习避免常见的分发问题。广泛的实验表明,我们的模型明显胜过最先进的基线。

Belief Propagation (BP) is an important message-passing algorithm for various reasoning tasks over graphical models, including solving the Constraint Optimization Problems (COPs). It has been shown that BP can achieve state-of-the-art performance on various benchmarks by mixing old and new messages before sending the new one, i.e., damping. However, existing methods of tuning a static damping factor for BP not only are laborious but also harm their performance. Moreover, existing BP algorithms treat each variable node's neighbors equally when composing a new message, which also limits their exploration ability. To address these issues, we seamlessly integrate BP, Gated Recurrent Units (GRUs), and Graph Attention Networks (GATs) within the message-passing framework to reason about dynamic weights and damping factors for composing new BP messages. Our model, Deep Attentive Belief Propagation (DABP), takes the factor graph and the BP messages in each iteration as the input and infers the optimal weights and damping factors through GRUs and GATs, followed by a multi-head attention layer. Furthermore, unlike existing neural-based BP variants, we propose a novel self-supervised learning algorithm for DABP with a smoothed solution cost, which does not require expensive training labels and also avoids the common out-of-distribution issue through efficient online learning. Extensive experiments show that our model significantly outperforms state-of-the-art baselines.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源