论文标题
边缘感知图表示学习和面部解析的推理
Edge-aware Graph Representation Learning and Reasoning for Face Parsing
论文作者
论文摘要
面部解析会侵入每个面部成分的像素标签,最近引起了很多关注。先前的方法表明了它们在面部解析方面的效率,但是忽略了不同面部区域之间的相关性。相关性是关于面部外观,姿势,表达等的关键线索,应考虑到面部解析。为此,我们建议通过学习图表表示图形表示,并在区域之间建模和推理区域关系,并利用区域之间的边缘信息进行优化的抽象。具体来说,我们将面部图像编码在一个全局图表示上,其中一个具有相似特征的像素(“区域”)的集合投影到每个顶点。我们的模型通过在图表上跨顶点传播信息来了解和原因。此外,我们将边缘信息合并为将像素的特征汇总到顶点,该特征强调边缘周围的特征,以沿边缘进行细分分割。最终学习的图表表示回到像素网格进行解析。实验表明,我们的模型在广泛使用的Helen数据集上优于最先进的方法,并且在大尺寸的Celebamask-HQ和LAPA数据集上也表现出卓越的性能。该代码可在https://github.com/tegusi/eagrnet上找到。
Face parsing infers a pixel-wise label to each facial component, which has drawn much attention recently. Previous methods have shown their efficiency in face parsing, which however overlook the correlation among different face regions. The correlation is a critical clue about the facial appearance, pose, expression etc., and should be taken into account for face parsing. To this end, we propose to model and reason the region-wise relations by learning graph representations, and leverage the edge information between regions for optimized abstraction. Specifically, we encode a facial image onto a global graph representation where a collection of pixels ("regions") with similar features are projected to each vertex. Our model learns and reasons over relations between the regions by propagating information across vertices on the graph. Furthermore, we incorporate the edge information to aggregate the pixel-wise features onto vertices, which emphasizes on the features around edges for fine segmentation along edges. The finally learned graph representation is projected back to pixel grids for parsing. Experiments demonstrate that our model outperforms state-of-the-art methods on the widely used Helen dataset, and also exhibits the superior performance on the large-scale CelebAMask-HQ and LaPa dataset. The code is available at https://github.com/tegusi/EAGRNet.