论文标题
深海视频中无脊椎动物物种的上下文驱动检测
Context-Driven Detection of Invertebrate Species in Deep-Sea Video
论文作者
论文摘要
每年,水下远程操作的车辆(ROV)收集了数千个未开发的海洋栖息地的视频,揭示了有关地球上生物多样性的大量信息。但是,充分利用这些信息仍然是一个挑战,因为适当的注释和分析需要训练有素的科学家时间,这既有限又昂贵。为此,我们提供了一个用于水下基板和无脊椎动物分析(DUSIA)的数据集,该数据集是一个基准套件和生长的大规模数据集,用于训练,验证和测试方法,用于暂时地定位四个水下底物以及暂时和空间定位的59个下水下污染物植物物种。目前,杜西亚(Dusia)包括在加利福尼亚海峡群岛附近的海底横跨海底的预计划的横断面,在30 fps中捕获的25个视频中,包括10个小时的镜头。每个视频都包括注释,表明视频中底物的开始和结束时间,除了感兴趣的物种计数。如图1所示,有些框架用精确的无脊椎动物盒位置进行注释。据我们所知,杜西亚是深海探索的第一个同类数据集,其中包括来自移动的相机的视频,其中包括底物注释和无脊椎动物物种,在阳光下没有渗透到阳光下的重要深度。此外,我们介绍了新颖的上下文驱动对象检测器(CDD),在其中我们使用显式底物分类来影响对象检测网络,以同时预测受该基板影响的底物和物种类别。我们还提出了一种改善部分注释边界框框架培训的方法。最后,我们提供了一种基线方法来自动化无脊椎动物感兴趣的物种。
Each year, underwater remotely operated vehicles (ROVs) collect thousands of hours of video of unexplored ocean habitats revealing a plethora of information regarding biodiversity on Earth. However, fully utilizing this information remains a challenge as proper annotations and analysis require trained scientists time, which is both limited and costly. To this end, we present a Dataset for Underwater Substrate and Invertebrate Analysis (DUSIA), a benchmark suite and growing large-scale dataset to train, validate, and test methods for temporally localizing four underwater substrates as well as temporally and spatially localizing 59 underwater invertebrate species. DUSIA currently includes over ten hours of footage across 25 videos captured in 1080p at 30 fps by an ROV following pre planned transects across the ocean floor near the Channel Islands of California. Each video includes annotations indicating the start and end times of substrates across the video in addition to counts of species of interest. Some frames are annotated with precise bounding box locations for invertebrate species of interest, as seen in Figure 1. To our knowledge, DUSIA is the first dataset of its kind for deep sea exploration, with video from a moving camera, that includes substrate annotations and invertebrate species that are present at significant depths where sunlight does not penetrate. Additionally, we present the novel context-driven object detector (CDD) where we use explicit substrate classification to influence an object detection network to simultaneously predict a substrate and species class influenced by that substrate. We also present a method for improving training on partially annotated bounding box frames. Finally, we offer a baseline method for automating the counting of invertebrate species of interest.