Dual Adaptive Pyramid Network for Cross-Stain Histopathology Image Segmentation

1. Author

Xianxu Hou1;2*, Jingxin Liu1;2;3, Bolei Xu1;2, Bozhi Liu1;2, Xin Chen4, Mohammad Ilyas5, Ian Ellis5, Jon Garibaldi4, and Guoping Qiu1;2;4
1 College of Information Engineering, Shenzhen University, Shenzhen, China
2 Guangdong Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen, China
3 Histo Pathology Diagnostic Center, Shanghai, China
4 School of Computer Science, University of Nottingham, Nottingham, United Kingdom
5 School of Medicine, University of Nottingham, Nottingham, United Kingdom
* Equal contribution

2. Abstract

Supervised semantic segmentation normally assumes the test data being in a similar data domain as the training data. However, in practice, the domain mismatch between the training and unseen data could lead to a significant performance drop.
Obtaining accurate pixelwise label for images in different domains is tedious and labor intensive, especially for histopathology images.
We tackle the domain adaptation problem on two levels:

  1. the image-level considers the differences of image color and style;
  2. the feature-level addresses the spatial inconsistency between two domains.

在这里插入图片描述

3. Introduction

Although excellent performance has been achieved on benchmark dataset, deep segmentation models have poor generalization capability to unseen datasets due to the domain shift between the training and test data. [Adversarial Discriminative Domain Adaptation]
The model trained on one (source) dataset would not generalize well when applied to the other (target)
dataset.
Although fine-tuning the model with labelled target data could possibly alleviate the impact of domain shift, manually annotating is a time-consuming, expensive and subjective process in medical area.
Therefore, it is of great interest to develop algorithms to adapt segmentation models from a source domain to a visually different target domain without requiring additional labels in the target domain.
The main insight behind these methods is trying to align visual appearance or feature distribution between the source and target domains.
The image-level adaptation considers the overall difference between source and target domain like image color and style, while feature-level adaptation addresses the spatial inconsistency of the two domains.

4. Method

在这里插入图片描述

4.1 Model Overview

It contains a semantic segmentation network G G G and two adversarial learning modules D i m g D_{img} Dimg and D f e a t D_{feat} Dfeat.

4.2 Segmentation Network

A dilated ResNet-18 is used as backbone to encode the input images.
In order to achieve larger receptive field of our model, we apply a Pyramid Pooling Module (PPM) from PSPNet on the last layer of the backbone network.
The different levels of features are then upsampled and concatenated as the pyramid pooling global feature.
Furthermore, we adopt skip connections from U-Net and a pyramid feature fusion architecture to achieve final segmentation.
The segmentation task is learned by minimizing both standard cross-entropy loss and Dice coefficient for images from the source domain:
L s e g = E x s ∼ X S [ − y s log ⁡ ( y ~ s ) ] + α E x s ∼ X S [ − 2 y s y ~ s y s + y ~ s ] \mathcal{L}_{s e g}=\mathbb{E}_{x_{s} \sim X_{S}}\left[-y_{s} \log \left(\widetilde{y}_{s}\right)\right]+\alpha \mathbb{E}_{x_{s} \sim X_{S}}\left[-\frac{2 y_{s} \widetilde{y}_{s}}{y_{s}+\widetilde{y}_{s}}\right] Lseg=ExsXS[yslog(y s)]+αExsXS[ys+y s2ysy s]

4.3 Domain Adaptation

4.3.1 Image-level Adaptation

In this work, image-level representation refers to the PPM outputs of the segmentation network G G G.
Image-level adaptation helps to reduce the shift by the global image difference such as image color and image style between the source and target domains.
To eliminate the domain distribution mismatch, we employ a discriminator D i m g D_{img} Dimg to distinguish PPM features between source images and target images.

In particular, we employ PatchGAN, a fully convolutional neural operating on image patches, from which we can get a two-dimensional feature map as the discriminator outputs.

4.3.2 Feature-level Adaptation

The feature-level representation refers to the fused feature maps before feeding into the final segmentation classifier.
Aligning the feature-level representations helps to reduce the segmentation differences in both global layout and local context.

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值