论文标题

人图像生成的两流外观转移网络

Two-Stream Appearance Transfer Network for Person Image Generation

论文作者

Shen, Chengkang, Wang, Peiyan, Tang, Wei

论文摘要

姿势引导的人形象产生意味着生成以输入人形象和所需姿势为条件的照片真实的人形象。此任务需要根据目标姿势对源图像进行空间操纵。但是,用于图像生成和翻译广泛用于图像的生成对抗网络(GAN)依赖于空间局部和翻译模棱两可的操作员,即卷积,池化和不冷冻,无法处理大图像变形。本文介绍了一个新型的两流外观传递网络(2S-ATN),以应对这一挑战。它是由源流和目标流组成的多阶段体系结构。每个阶段都有一个外观传递模块和几个两流特征融合模块。前者发现两流特征图之间的密集对应关系,然后将外观信息从源流传输到目标流。后者交换了两个流之间的局部信息,并补充了非本地外观转移。定量和定性结果表明,所提出的2S-ATN可以有效处理大型空间变形和遮挡,同时保留外观细节。在两个广泛使用的基准上,它的表现优于先前的艺术状态。

Pose guided person image generation means to generate a photo-realistic person image conditioned on an input person image and a desired pose. This task requires spatial manipulation of the source image according to the target pose. However, the generative adversarial networks (GANs) widely used for image generation and translation rely on spatially local and translation equivariant operators, i.e., convolution, pooling and unpooling, which cannot handle large image deformation. This paper introduces a novel two-stream appearance transfer network (2s-ATN) to address this challenge. It is a multi-stage architecture consisting of a source stream and a target stream. Each stage features an appearance transfer module and several two-stream feature fusion modules. The former finds the dense correspondence between the two-stream feature maps and then transfers the appearance information from the source stream to the target stream. The latter exchange local information between the two streams and supplement the non-local appearance transfer. Both quantitative and qualitative results indicate the proposed 2s-ATN can effectively handle large spatial deformation and occlusion while retaining the appearance details. It outperforms prior states of the art on two widely used benchmarks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源