论文标题
与条件可逆神经网络的网络到网络翻译
Network-to-Network Translation with Conditional Invertible Neural Networks
论文作者
论文摘要
鉴于现代机器学习模型的计算成本不断增加,我们需要找到新的方法来重复使用此类专家模型,从而利用已投资于其创建的资源。最近的工作表明,这些庞大的模型的力量是由他们所学的表现所捕获的。因此,我们寻求一个可以在不同现有表示形式之间关联的模型,并建议通过条件可逆的网络解决此任务。该网络通过(i)通过(i)在不同域之间提供一般传递来证明其能力,(ii)通过允许在其他域中进行修改来启用受控内容综合,以及(iii)通过将它们转化为可解释的域(例如图像)来促进现有表示形式的诊断。我们的域转移网络可以在固定表示之间转换,而无需学习或对它们进行填补。这使用户可以利用经过广泛的计算资源培训的文献中的各种现有领域特定的专家模型。关于各种条件图像综合任务,竞争图像修改结果的实验以及图像到图像和文本对图像生成的实验证明了我们方法的通用适用性。例如,我们在Bert和Biggan,最先进的文本和图像模型之间转换,以提供文本到图像生成,两位专家都无法独自执行。
Given the ever-increasing computational costs of modern machine learning models, we need to find new ways to reuse such expert models and thus tap into the resources that have been invested in their creation. Recent work suggests that the power of these massive models is captured by the representations they learn. Therefore, we seek a model that can relate between different existing representations and propose to solve this task with a conditionally invertible network. This network demonstrates its capability by (i) providing generic transfer between diverse domains, (ii) enabling controlled content synthesis by allowing modification in other domains, and (iii) facilitating diagnosis of existing representations by translating them into interpretable domains such as images. Our domain transfer network can translate between fixed representations without having to learn or finetune them. This allows users to utilize various existing domain-specific expert models from the literature that had been trained with extensive computational resources. Experiments on diverse conditional image synthesis tasks, competitive image modification results and experiments on image-to-image and text-to-image generation demonstrate the generic applicability of our approach. For example, we translate between BERT and BigGAN, state-of-the-art text and image models to provide text-to-image generation, which neither of both experts can perform on their own.