论文标题

闭塞意识到从视频中无监督的光流学习

Occlusion Aware Unsupervised Learning of Optical Flow From Video

论文作者

Li, Jianfeng, Zhao, Junqiao, Feng, Tiantian, Ye, Chen, Xiong, Lu

论文摘要

在本文中,我们提出了一种无监督的学习方法,用于估计视频帧之间的光流,尤其是解决遮挡问题。遮挡是由对象的运动或相机的移动引起的,该对象的移动定义为当某些像素在一个视频框架中可见时,而在相邻的框架中不可见。由于遮挡区域中框架之间缺乏像素对应关系,错误的光度损失计算会误导光流训练过程。在视频序列中,我们发现向前($ t \ rightarrow t+1 $)和向后($ t \ rightarrow t-1 $)框架对框架的闭塞通常是互补的。也就是说,在后续帧中被遮挡的像素通常不会在上一个帧中遮住,反之亦然。因此,通过使用这种互补性,提出了新的加权损失来解决遮挡问题。此外,我们在多个方向上计算梯度以提供更丰富的监督信息。与基线和Kitti 2012和2015基准的一些监督方法相比,我们的方法达到了竞争性的光流精度。该源代码已在https://github.com/jianfenglihg/unopticalflow.git上发布。

In this paper, we proposed an unsupervised learning method for estimating the optical flow between video frames, especially to solve the occlusion problem. Occlusion is caused by the movement of an object or the movement of the camera, defined as when certain pixels are visible in one video frame but not in adjacent frames. Due to the lack of pixel correspondence between frames in the occluded area, incorrect photometric loss calculation can mislead the optical flow training process. In the video sequence, we found that the occlusion in the forward ($t\rightarrow t+1$) and backward ($t\rightarrow t-1$) frame pairs are usually complementary. That is, pixels that are occluded in subsequent frames are often not occluded in the previous frame and vice versa. Therefore, by using this complementarity, a new weighted loss is proposed to solve the occlusion problem. In addition, we calculate gradients in multiple directions to provide richer supervision information. Our method achieves competitive optical flow accuracy compared to the baseline and some supervised methods on KITTI 2012 and 2015 benchmarks. This source code has been released at https://github.com/jianfenglihg/UnOpticalFlow.git.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源