论文标题
重新思考基于物理视觉中图像恢复图像恢复的生成方法:从信息的角度来看的理论分析
Rethinking Generative Methods for Image Restoration in Physics-based Vision: A Theoretical Analysis from the Perspective of Information
论文作者
论文摘要
与基于手工制作的构图模型相比,端到端生成方法被认为是基于物理视觉的图像恢复的更有希望的解决方案。但是,现有的生成方法仍然有足够的空间来改善定量性能。更重要的是,由于解释性弱,这些方法被认为是黑匣子,而且很少有一种理论试图解释其机制和学习过程。在这项研究中,我们尝试使用信息理论重新解释这些生成方法,以换取图像恢复任务。与常规理解不同,我们分析了这些方法的信息流,并确定了三个信息来源(提取的高级信息,保留的低级信息以及来自源输入中缺乏的外部信息)分别参与并在生成恢复结果方面进行了优化。我们通过扩展信息瓶颈原则进一步得出了他们的学习行为,优化目标和相应的信息边界。基于这个理论框架,我们发现许多现有的生成方法往往是专为传统生成任务设计的一般模型的直接应用,这些模型可能会遭受包括过度投资的抽象过程,固有的细节损失以及消失的梯度或培训失衡的问题。我们用直观和理论解释分析了这些问题,并分别用经验证据证明了这些问题。最终,我们提出了一般解决方案或想法来解决上述问题,并在三个不同图像恢复任务的六个数据集上提高了这些方法,并验证了这些方法。
End-to-end generative methods are considered a more promising solution for image restoration in physics-based vision compared with the traditional deconstructive methods based on handcrafted composition models. However, existing generative methods still have plenty of room for improvement in quantitative performance. More crucially, these methods are considered black boxes due to weak interpretability and there is rarely a theory trying to explain their mechanism and learning process. In this study, we try to re-interpret these generative methods for image restoration tasks using information theory. Different from conventional understanding, we analyzed the information flow of these methods and identified three sources of information (extracted high-level information, retained low-level information, and external information that is absent from the source inputs) are involved and optimized respectively in generating the restoration results. We further derived their learning behaviors, optimization objectives, and the corresponding information boundaries by extending the information bottleneck principle. Based on this theoretic framework, we found that many existing generative methods tend to be direct applications of the general models designed for conventional generation tasks, which may suffer from problems including over-invested abstraction processes, inherent details loss, and vanishing gradients or imbalance in training. We analyzed these issues with both intuitive and theoretical explanations and proved them with empirical evidence respectively. Ultimately, we proposed general solutions or ideas to address the above issue and validated these approaches with performance boosts on six datasets of three different image restoration tasks.