论文标题
维斯伯特:变压器的隐藏状态可视化
VisBERT: Hidden-State Visualizations for Transformers
论文作者
论文摘要
解释性和可解释性是两个重要的概念,缺乏这些概念可以并且应该阻碍表现良好的神经网络在现实世界中的应用。同时,它们很难将实现最先进的大型黑盒模型纳入大量的NLP任务。来自变形金刚(BERT)的双向编码器表示形式就是这样的黑框模型。它已成为解决许多不同NLP任务的主食架构,并启发了许多相关的变压器模型。了解这些模型如何得出结论对于它们的改进和应用至关重要。我们通过展示Visbert来贡献这一挑战,Visbert是一种可视化BERT中的上下文令牌表示的工具,以解决(多跳)问题回答的任务。我们没有分析注意力权重,而是专注于BERT模型中每个编码器块产生的隐藏状态。这样,我们可以观察到在模型的整个层中如何转换语义表示。维斯伯特(Visbert)使用户能够了解该模型的内部状态并探索其推理步骤或潜在的缺点。该工具使我们能够在伯特的转换中识别与传统NLP管道相似的不同阶段,并在预测失败期间提供见解。
Explainability and interpretability are two important concepts, the absence of which can and should impede the application of well-performing neural networks to real-world problems. At the same time, they are difficult to incorporate into the large, black-box models that achieve state-of-the-art results in a multitude of NLP tasks. Bidirectional Encoder Representations from Transformers (BERT) is one such black-box model. It has become a staple architecture to solve many different NLP tasks and has inspired a number of related Transformer models. Understanding how these models draw conclusions is crucial for both their improvement and application. We contribute to this challenge by presenting VisBERT, a tool for visualizing the contextual token representations within BERT for the task of (multi-hop) Question Answering. Instead of analyzing attention weights, we focus on the hidden states resulting from each encoder block within the BERT model. This way we can observe how the semantic representations are transformed throughout the layers of the model. VisBERT enables users to get insights about the model's internal state and to explore its inference steps or potential shortcomings. The tool allows us to identify distinct phases in BERT's transformations that are similar to a traditional NLP pipeline and offer insights during failed predictions.