论文标题
过度参数改善了样式逆转
Overparameterization Improves StyleGAN Inversion
论文作者
论文摘要
诸如stylegan之类的深层生成模型具有语义图像编辑的希望:通过其内容修改图像,而不是像素值。不幸的是,使用任意图像需要颠倒stylegan生成器,到目前为止,这仍然具有挑战性。现有的反转方法获得了有希望而又不完美的结果,必须在重建质量和下游编辑性之间进行权衡。为了提高质量,这些方法必须采用各种技术,这些技术在训练后扩展了潜在空间。退后一步,我们观察到这些方法本质上都以一种或另一种方式提出了增加自由参数的数量。这表明反转可能很困难,因为它是不足的。在这项工作中,我们在训练之前直接和巨大地解决了潜在空间的过度参数,并简单地更改了原始的StyleGan体系结构。我们的过度参数增加了可用的自由度,这反过来促进了反转。我们表明,这使我们能够获得近乎完美的图像重建,而无需编码器或在训练后更改潜在空间。我们的方法还保留了编辑性,我们通过在图像之间实际插值来证明这一点。
Deep generative models like StyleGAN hold the promise of semantic image editing: modifying images by their content, rather than their pixel values. Unfortunately, working with arbitrary images requires inverting the StyleGAN generator, which has remained challenging so far. Existing inversion approaches obtain promising yet imperfect results, having to trade-off between reconstruction quality and downstream editability. To improve quality, these approaches must resort to various techniques that extend the model latent space after training. Taking a step back, we observe that these methods essentially all propose, in one way or another, to increase the number of free parameters. This suggests that inversion might be difficult because it is underconstrained. In this work, we address this directly and dramatically overparameterize the latent space, before training, with simple changes to the original StyleGAN architecture. Our overparameterization increases the available degrees of freedom, which in turn facilitates inversion. We show that this allows us to obtain near-perfect image reconstruction without the need for encoders nor for altering the latent space after training. Our approach also retains editability, which we demonstrate by realistically interpolating between images.