论文标题
可解释的3D卷积神经网络通过学习时间转换
Explainable 3D Convolutional Neural Networks by Learning Temporal Transformations
论文作者
论文摘要
在本文中,我们将时间分级的3D卷积(3TCONV)介绍为常规3D卷积(3DCONV)的可解释替代方案。在3TCONV中,通过学习2D滤波器和一组时间变换参数获得3D卷积滤波器,从而导致稀疏的滤波器在时间维度中依次依次依赖2D切片。我们证明3TCONV学习了可以直接解释的时间转变。时间参数可以与各种现有的2D可视化方法结合使用。我们还表明,通过分析层和模型级别上的转换参数统计信息可以实现有关模型学习的知识。最后,我们隐含地证明,在流行的Convnets中,2DCONV可以用3TCONV替换,并且可以将权重转移以产生预验证的3TCONV。预估计的3TCONVNET能够利用已被证明可以在图像分类基准上带来出色成果的功能,从而在传统2DConvnets上利用了十多年的工作。
In this paper we introduce the temporally factorized 3D convolution (3TConv) as an interpretable alternative to the regular 3D convolution (3DConv). In a 3TConv the 3D convolutional filter is obtained by learning a 2D filter and a set of temporal transformation parameters, resulting in a sparse filter where the 2D slices are sequentially dependent on each other in the temporal dimension. We demonstrate that 3TConv learns temporal transformations that afford a direct interpretation. The temporal parameters can be used in combination with various existing 2D visualization methods. We also show that insight about what the model learns can be achieved by analyzing the transformation parameter statistics on a layer and model level. Finally, we implicitly demonstrate that, in popular ConvNets, the 2DConv can be replaced with a 3TConv and that the weights can be transferred to yield pretrained 3TConvs. pretrained 3TConvnets leverage more than a decade of work on traditional 2DConvNets by being able to make use of features that have been proven to deliver excellent results on image classification benchmarks.