论文标题
CAID:在医学成像中进行自我监督学习的上下文意识实例歧视
CAiD: Context-Aware Instance Discrimination for Self-supervised Learning in Medical Imaging
论文作者
论文摘要
最近,从未标记的摄影图像学习视觉表示方面取得了显着成功的实例歧视方法。但是,鉴于摄影图像和医学图像之间的明显差异,基于实例的目标的功效(重点是学习图像中最具歧视性的全局特征(即自行车中的车轮),在医学成像中仍然未知。我们的初步分析表明,就解剖缩影而言,医学图像的全球相似性很高,用于捕获一系列不同特征的实例歧视方法,从而对其在下游任务上的医学上产生负面影响。为了减轻这一限制,我们开发了一个简单而有效的自我监督框架,称为上下文感知实例歧视(CAID)。 CAID的目的是通过提供从无标记的医学图像的各种本地环境中编码的更细致,更具歧视性的信息来改善实例歧视学习。我们进行了系统的分析,以从三方面的角度研究学习特征的实用性:(i)概括性和可传递性,(ii)嵌入空间中的可分离性以及(iii)可重复使用性。我们的广泛实验表明,CAID(1)丰富了从现有实例歧视方法中学到的表示。 (2)通过从单个内侧图像中捕获更精细的上下文信息来提供更多的判别特征; (3)与标准实例判别方法相比,低/中级特征的可重复性。作为开放科学,所有代码和预培训模型均可在我们的GitHub页面上提供:https://github.com/jlianglab/caid。
Recently, self-supervised instance discrimination methods have achieved significant success in learning visual representations from unlabeled photographic images. However, given the marked differences between photographic and medical images, the efficacy of instance-based objectives, focusing on learning the most discriminative global features in the image (i.e., wheels in bicycle), remains unknown in medical imaging. Our preliminary analysis showed that high global similarity of medical images in terms of anatomy hampers instance discrimination methods for capturing a set of distinct features, negatively impacting their performance on medical downstream tasks. To alleviate this limitation, we have developed a simple yet effective self-supervised framework, called Context-Aware instance Discrimination (CAiD). CAiD aims to improve instance discrimination learning by providing finer and more discriminative information encoded from a diverse local context of unlabeled medical images. We conduct a systematic analysis to investigate the utility of the learned features from a three-pronged perspective: (i) generalizability and transferability, (ii) separability in the embedding space, and (iii) reusability. Our extensive experiments demonstrate that CAiD (1) enriches representations learned from existing instance discrimination methods; (2) delivers more discriminative features by adequately capturing finer contextual information from individual medial images; and (3) improves reusability of low/mid-level features compared to standard instance discriminative methods. As open science, all codes and pre-trained models are available on our GitHub page: https://github.com/JLiangLab/CAiD.