论文标题
通过子零件对齐来回答的强大问题
Robust Question Answering Through Sub-part Alignment
论文作者
论文摘要
当前的文本问题回答模型在内域测试集上实现了强大的性能,但通常通过在数据中拟合表面级别的模式来实现,因此它们无法推广到分布式设置。为了使更强大,更可理解的质量检查系统,我们将回答作为对齐问题进行建模。我们将问题和上下文分解为基于现成的语义表示(在此,语义角色),并将问题与上下文的子图相结合以找到答案的较小单元。我们将模型作为结构化的SVM制定,并通过BERT计算对齐得分,尽管使用梁搜索大概推断,但我们仍可以端到端训练。我们明确地使用对齐方式使我们能够探索一组约束,我们可以禁止在跨域设置中产生的某些类型的不良模型行为。此外,通过调查不同潜在答案的分数差异,我们可以寻求了解输入的特定方面导致模型在不依赖事后解释技术的情况下选择答案。我们在小队V1.1上训练模型,并在几个对抗和室外数据集上进行测试。结果表明,我们的模型比标准BERT QA模型更强大,并且从一致性分数得出的约束使我们能够有效地权衡覆盖范围和准确性。
Current textual question answering models achieve strong performance on in-domain test sets, but often do so by fitting surface-level patterns in the data, so they fail to generalize to out-of-distribution settings. To make a more robust and understandable QA system, we model question answering as an alignment problem. We decompose both the question and context into smaller units based on off-the-shelf semantic representations (here, semantic roles), and align the question to a subgraph of the context in order to find the answer. We formulate our model as a structured SVM, with alignment scores computed via BERT, and we can train end-to-end despite using beam search for approximate inference. Our explicit use of alignments allows us to explore a set of constraints with which we can prohibit certain types of bad model behavior arising in cross-domain settings. Furthermore, by investigating differences in scores across different potential answers, we can seek to understand what particular aspects of the input lead the model to choose the answer without relying on post-hoc explanation techniques. We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets. The results show that our model is more robust cross-domain than the standard BERT QA model, and constraints derived from alignment scores allow us to effectively trade off coverage and accuracy.