论文标题
使用多阶段对抗引导注意训练对OCT扫描的超分辨率和分割
Superresolution and Segmentation of OCT scans using Multi-Stage adversarial Guided Attention Training
论文作者
论文摘要
光学相干断层扫描(OCT)是未侵入性且易于吸引的生物标志物(视网膜层的厚度(可在OCT扫描中可检测到的)),以诊断阿尔茨海默氏病(AD)。这项工作旨在自动细分OCT图像。但是,由于各种问题,例如斑点噪声,小目标区域和不利的成像条件,这是一项具有挑战性的任务。在我们以前的工作中,我们提出了多阶段和多歧视性生成对抗网络(Multisdgan),以在高分辨率分段标签中翻译OCT扫描。在这项调查中,我们旨在评估和比较渠道和空间关注的各种组合与多gan架构的各种组合,以通过捕获丰富的上下文关系以提高细分性能来提取更强大的特征图。此外,我们开发并评估了一个引导的Mutli阶段注意力框架,在该框架中,我们通过在特定设计的二进制掩码和生成的注意图之间强迫L-1损失来结合引导的注意机制。我们的消融研究在五倍的交叉验证(5-CV)中对WVU-OCT数据集的结果结果表明,带有序列注意模块的拟议的多型人提供了最有竞争力的性能,并通过二进制蒙版指导空间注意力图,进一步提高了我们建议的网络的性能。将基线模型与增加指导性注意事项进行比较,我们的结果表明,骰子系数和SSIM的相对改善分别为21.44%和19.45%。
Optical coherence tomography (OCT) is one of the non-invasive and easy-to-acquire biomarkers (the thickness of the retinal layers, which is detectable within OCT scans) being investigated to diagnose Alzheimer's disease (AD). This work aims to segment the OCT images automatically; however, it is a challenging task due to various issues such as the speckle noise, small target region, and unfavorable imaging conditions. In our previous work, we have proposed the multi-stage & multi-discriminatory generative adversarial network (MultiSDGAN) to translate OCT scans in high-resolution segmentation labels. In this investigation, we aim to evaluate and compare various combinations of channel and spatial attention to the MultiSDGAN architecture to extract more powerful feature maps by capturing rich contextual relationships to improve segmentation performance. Moreover, we developed and evaluated a guided mutli-stage attention framework where we incorporated a guided attention mechanism by forcing an L-1 loss between a specifically designed binary mask and the generated attention maps. Our ablation study results on the WVU-OCT data-set in five-fold cross-validation (5-CV) suggest that the proposed MultiSDGAN with a serial attention module provides the most competitive performance, and guiding the spatial attention feature maps by binary masks further improves the performance in our proposed network. Comparing the baseline model with adding the guided-attention, our results demonstrated relative improvements of 21.44% and 19.45% on the Dice coefficient and SSIM, respectively.