论文标题
基于自动加权层表示基于3-D视频编码的视图综合造成失真估计
Auto-Weighted Layer Representation Based View Synthesis Distortion Estimation for 3-D Video Coding
论文作者
论文摘要
最近,已经研究了各种视图综合失真估计模型,以更好地用于3D视频编码。但是,它们几乎无法在不同级别的深度变化,纹理变性和视图合成扭曲(VSD)之间进行定量建模,这对于费率降低优化和速率分配至关重要。在本文中,开发了基于自动加权层表示综合估计模型。首先,根据深度变化及其相关纹理变性的水平来定义子VSD(S-VSD)。之后,一组理论派生表明,可以将VSD大约分解到S-VSD中,乘以其相关权重。为了获得S-VSD,开发了S-VSD的基于层的表示形式,其中所有具有相同深度变化级别的像素都用层表示,以在层级别启用有效的S-VSD计算。同时,学会了一个非线性映射函数,以准确地表示VSD和S-VSD之间的关系,在VSD估计过程中自动提供S-VSD的权重。为了学习此类功能,构建了VSD及其相关的S-VSD的数据集。实验结果表明,一旦可用的S-VSD可用,就可以通过非线性映射函数学到的权重准确地估算VSD。所提出的方法在准确性和效率方面都优于相关的最新方法。该方法的数据集和源代码将在https://github.com/jianjin008/上找到。
Recently, various view synthesis distortion estimation models have been studied to better serve for 3-D video coding. However, they can hardly model the relationship quantitatively among different levels of depth changes, texture degeneration, and the view synthesis distortion (VSD), which is crucial for rate-distortion optimization and rate allocation. In this paper, an auto-weighted layer representation based view synthesis distortion estimation model is developed. Firstly, the sub-VSD (S-VSD) is defined according to the level of depth changes and their associated texture degeneration. After that, a set of theoretical derivations demonstrate that the VSD can be approximately decomposed into the S-VSDs multiplied by their associated weights. To obtain the S-VSDs, a layer-based representation of S-VSD is developed, where all the pixels with the same level of depth changes are represented with a layer to enable efficient S-VSD calculation at the layer level. Meanwhile, a nonlinear mapping function is learnt to accurately represent the relationship between the VSD and S-VSDs, automatically providing weights for S-VSDs during the VSD estimation. To learn such function, a dataset of VSD and its associated S-VSDs are built. Experimental results show that the VSD can be accurately estimated with the weights learnt by the nonlinear mapping function once its associated S-VSDs are available. The proposed method outperforms the relevant state-of-the-art methods in both accuracy and efficiency. The dataset and source code of the proposed method will be available at https://github.com/jianjin008/.