论文标题
具有收敛性和鲁棒性的基于记忆效率的基于模型的深度学习
Memory-efficient model-based deep learning with convergence and robustness guarantees
论文作者
论文摘要
计算成像通过压缩传感算法进行了革命,这些算法提供了保证的唯一性,收敛性和稳定性。将成像物理学与学习的正规化先验相结合的基于模型的深度学习方法已成为图像恢复的更强大的替代方法。本文的主要重点是引入具有与CS方法相似的理论保证的基于内存有效的算法。所提出的迭代算法在涉及分数函数的梯度下降与共轭梯度算法之间交替,以鼓励数据一致性。分数函数被建模为单调卷积神经网络。我们的分析表明,单调约束是必要的,足以在任意反问题中强制固定点的唯一性。此外,它还保证了汇聚到固定点,这对输入扰动是可靠的。我们介绍了所提出的MOL框架的两个实现,这些实现在单调性质的施加方式方面有所不同。第一种方法实施严格的单调约束,而第二种方法依赖于近似值。从严格意义上讲,保证对于第二种方法无效。但是,我们的实证研究表明,两种方法的融合和鲁棒性都是可比性的,而近似近似实施的差异则提供了更好的性能。所提出的深度平衡公式比展开的方法明显高得多,这使我们能够将其应用于当前展开的算法无法处理的3D或2D+时间问题。
Computational imaging has been revolutionized by compressed sensing algorithms, which offer guaranteed uniqueness, convergence, and stability properties. Model-based deep learning methods that combine imaging physics with learned regularization priors have emerged as more powerful alternatives for image recovery. The main focus of this paper is to introduce a memory efficient model-based algorithm with similar theoretical guarantees as CS methods. The proposed iterative algorithm alternates between a gradient descent involving the score function and a conjugate gradient algorithm to encourage data consistency. The score function is modeled as a monotone convolutional neural network. Our analysis shows that the monotone constraint is necessary and sufficient to enforce the uniqueness of the fixed point in arbitrary inverse problems. In addition, it also guarantees the convergence to a fixed point, which is robust to input perturbations. We introduce two implementations of the proposed MOL framework, which differ in the way the monotone property is imposed. The first approach enforces a strict monotone constraint, while the second one relies on an approximation. The guarantees are not valid for the second approach in the strict sense. However, our empirical studies show that the convergence and robustness of both approaches are comparable, while the less constrained approximate implementation offers better performance. The proposed deep equilibrium formulation is significantly more memory efficient than unrolled methods, which allows us to apply it to 3D or 2D+time problems that current unrolled algorithms cannot handle.