论文标题
超声心动图视频的自我监督对比度学习可以使标签有效的心脏病诊断
Self-supervised contrastive learning of echocardiogram videos enables label-efficient cardiac disease diagnosis
论文作者
论文摘要
自我监督学习的进步(SSL)表明,在医学成像数据上进行自我监督的预处理可以为下游监督分类和细分提供强大的初始化。鉴于难以获得医学图像识别任务的专家标签,因此通常需要使用标准转移学习的标签效率提高标签效率,因此通常需要使用这种“内域” SSL初始化。但是,大多数用于SSL医学成像数据的努力并不适合基于视频的医学成像方式。考虑到这一进展,我们开发了一种自制的对比学习方法,Echoclr迎合了超声心动图视频,其目的是学习强有力的表现,以在下游心脏病诊断上有效进行微调。 echoclr杠杆(i)同一患者的不同视频是对比度学习的正对,以及(ii)重新排序借口任务以执行时间连贯性。当对标记的数据的一小部分数据进行微调(少于51次检查)时,对其他转移学习和SSL方法的左心室肥大(LVH)和主动脉狭窄(AS)的分类性能显着改善了分类性能,并且在内部和外部测试集中进行了SSL方法。例如,当对10%的可用培训数据(519项研究)进行微调时,与0.61 AUROC(95%CI:[0.57,0.64])相比,LVH分类的EchoclR预测模型在LVH分类上达到了0.72 AUROC(95%CI:[0.69,0.75])。同样,使用1%的Echoclr预审计,与0.61 AUROC(95%CI:[0.58,0.65])相比,Echoclr预审进达到0.82 AUROC(95%CI:[0.79,0.84])。 Echoclr在学习医学视频表示的能力方面是独一无二的,并证明SSL可以从小标签的数据集中启用标签有效的疾病分类。
Advances in self-supervised learning (SSL) have shown that self-supervised pretraining on medical imaging data can provide a strong initialization for downstream supervised classification and segmentation. Given the difficulty of obtaining expert labels for medical image recognition tasks, such an "in-domain" SSL initialization is often desirable due to its improved label efficiency over standard transfer learning. However, most efforts toward SSL of medical imaging data are not adapted to video-based medical imaging modalities. With this progress in mind, we developed a self-supervised contrastive learning approach, EchoCLR, catered to echocardiogram videos with the goal of learning strong representations for efficient fine-tuning on downstream cardiac disease diagnosis. EchoCLR leverages (i) distinct videos of the same patient as positive pairs for contrastive learning and (ii) a frame re-ordering pretext task to enforce temporal coherence. When fine-tuned on small portions of labeled data (as few as 51 exams), EchoCLR pretraining significantly improved classification performance for left ventricular hypertrophy (LVH) and aortic stenosis (AS) over other transfer learning and SSL approaches across internal and external test sets. For example, when fine-tuning on 10% of available training data (519 studies), an EchoCLR-pretrained model achieved 0.72 AUROC (95% CI: [0.69, 0.75]) on LVH classification, compared to 0.61 AUROC (95% CI: [0.57, 0.64]) with a standard transfer learning approach. Similarly, using 1% of available training data (53 studies), EchoCLR pretraining achieved 0.82 AUROC (95% CI: [0.79, 0.84]) on severe AS classification, compared to 0.61 AUROC (95% CI: [0.58, 0.65]) with transfer learning. EchoCLR is unique in its ability to learn representations of medical videos and demonstrates that SSL can enable label-efficient disease classification from small, labeled datasets.