UNet进行病理图像分割

数据集链接:https://pan.baidu.com/s/1IBe_P0AyHgZC39NqzOxZhA?pwd=nztc
提取码:nztc

  • UNet模型
import torch
import torch.nn as nn

class conv_block(nn.Module):
    def __init__(self, ch_in, ch_out):
        super(conv_block, self).__init__()
        self.conv = nn.Sequential(
            nn.Conv2d(ch_in, ch_out, kernel_size=3, stride=1, padding=1, bias=True),
            nn.BatchNorm2d(ch_out),
            nn.ReLU(inplace=True),
            nn.Conv2d(ch_out, ch_out, kernel_size=3, stride=1, padding=1, bias=True),
            nn.BatchNorm2d(ch_out),
            nn.ReLU(inplace=True)
        )
    def forward(self, x):
        x = self.conv(x)
        return x

class up_conv(nn.Module):
    def __init__(self, ch_in, ch_out):
        super(up_conv, self).__init__()
        self.up = nn.Sequential(
            nn.Upsample(scale_factor=2),
            nn.Conv2d(ch_in, ch_out, kernel_size=3, stride=1, padding=1, bias=True),
            nn.BatchNorm2d(ch_out),
            nn.ReLU(inplace=True)
        )
    def forward(self, x):
        x = self.up(x)
        return x
class UNet(nn.Module):
    def __init__(self, img_ch=3, output_ch=1):
        super(UNet, self).__init__()
        self.Maxpool = nn.MaxPool2d(kernel_size=2, stride=2)
        self.Conv1 = conv_block(ch_in=img_ch, ch_out=64)
        self.Conv2 = conv_block(ch_in=64, ch_out=128)
        self.Conv3 = conv_block(ch_in=128, ch_out=256)
        self.Conv4 = conv_block(ch_in=256, ch_out=512)
        self.Conv5 = conv_block(ch_in=512, ch_out=1024)
        self.Up5 = up_conv(ch_in=1024, ch_out=512)
        self.Up_conv5 = conv_block(ch_in=1024, ch_out=512)
        self.Up4 = up_conv(ch_in=512, ch_out=256)
        self.Up_conv4 = conv_block(ch_in=512, ch_out=256)
        self.Up3 = up_conv(ch_in=256, ch_out=128)
        self.Up_conv3 = conv_block(ch_in=256, ch_out=128)
        self.Up2 = up_conv(ch_in=128, ch_out=64)
        self.Up_conv2 = conv_block(ch_in=128, ch_out=64)
        self.Conv_1x1 = nn.Conv2d(64, output_ch, kernel_size=1, stride=1, padding=0)

    def forward(self, x):
        # encoding path
        x1 = self.Conv1(x)
        x2 = self.Maxpool(x1)
        x2 = self.Conv2(x2)
        x3 = self.Maxpool(x2)
        x3 = self.Conv3(x3)
        x4 = self.Maxpool(x3)
        x4 = self.Conv4(x4)
        x5 = self.Maxpool(x4)
        x5 = self.Conv5(x5)
        # decoding + concat path
        d5 = self.Up5(x5)
        d5 = torch.cat((x4, d5), dim=1)
        d5 = self.Up_conv5(d5)
        d4 = self.Up4(d5)
        d4 = torch.cat((x3, d4), dim=1)
        d4 = self.Up_conv4(d4)
        d3 = self.Up3(d4)
        d3 = torch.cat((x2, d3), dim=1)
        d3 = self.Up_conv3(d3)
        d2 = self.Up2(d3)
        d2 = torch.cat((x1, d2), dim=1)
        d2 = self.Up_conv2(d2)
        d1 = self.Conv_1x1(d2)
        output = torch.sigmoid(d1)  # 在最后加上Sigmoid激活函数
        return output
  • 数据加载
import os
from PIL import Image
from torch.utils.data import Dataset
from torchvision import transforms

class SegmentationDataset(Dataset):
    def __init__(self, image_dir, mask_dir, output_size=(256, 256)):
        self.image_dir = image_dir
        self.mask_dir = mask_dir
        self.image_list = os.listdir(image_dir)
        self.output_size = output_size
        # 定义图像和掩码的变换
        self.image_transform = transforms.Compose([
            transforms.Resize(self.output_size),
            transforms.ToTensor()
        ])
        self.mask_transform = transforms.Compose([
            transforms.Resize(self.output_size),
            transforms.ToTensor()
        ])

    def __len__(self):
        return len(self.image_list)

    def __getitem__(self, idx):
        image_name = self.image_list[idx]
        image_path = os.path.join(self.image_dir, image_name)
        mask_path = os.path.join(self.mask_dir, image_name)
        image = Image.open(image_path).convert("RGB")  # 确保是RGB
        mask = Image.open(mask_path).convert("L")  # 确保是灰度图像
        image = self.image_transform(image)
        mask = self.mask_transform(mask)
        return image, mask
  • 训练和测试。训练函数中保存的最好模型后缀最大(因为loss小才保存当前这个epoch的模型,我训练的最好模型是第171轮产生的),测试代码包含计算模型性能指标的代码和保存结果图片的代码。
import os
import numpy as np
import torch
import torch.optim as optim
from sklearn.metrics import confusion_matrix
from torch import nn
from torch.utils.data import DataLoader
import matplotlib.pyplot as plt
from tqdm import tqdm
from UNet import UNet
from DataLoader2 import SegmentationDataset

# IoU计算
def compute_iou(pred_mask, true_mask):
    smooth = 1e-6  # 避免分母为0
    pred_mask = (pred_mask > 0.5).float()
    true_mask = (true_mask > 0.5).float()

    intersection = (pred_mask * true_mask).sum()
    union = pred_mask.sum() + true_mask.sum() - intersection

    return (intersection + smooth) / (union + smooth)

# Dice系数计算
def compute_dice(pred_mask, true_mask):
    smooth = 1e-6  # 避免分母为0
    pred_mask = (pred_mask > 0.5).float()
    true_mask = (true_mask > 0.5).float()

    intersection = (pred_mask * true_mask).sum()

    return (2. * intersection + smooth) / (pred_mask.sum() + true_mask.sum() + smooth)

# 精度、召回率和F1分数计算
def compute_precision_recall_f1(pred_mask, true_mask):
    pred_mask = (pred_mask > 0.5).numpy().astype(int)
    true_mask = (true_mask > 0.5).numpy().astype(int)

    # 将mask平展为一维数组
    pred_mask_flat = pred_mask.flatten()
    true_mask_flat = true_mask.flatten()

    conf_matrix = confusion_matrix(true_mask_flat, pred_mask_flat)
    tn, fp, fn, tp = conf_matrix.ravel()

    precision = tp / (tp + fp)
    recall = tp / (tp + fn)
    f1_score = 2 * (precision * recall) / (precision + recall)

    return precision, recall, f1_score


# 训练函数
def train():
    model = UNet()
    dataset = SegmentationDataset('./dataset_exp2/train/image', './dataset_exp2/train/label')
    dataloader = DataLoader(batch_size=16, shuffle=True, dataset=dataset)
    # 训练参数
    num_epochs = 200
    learning_rate = 1e-4
    # 损失函数和优化器
    criterion = nn.BCELoss()
    optimizer = optim.Adam(model.parameters(), lr=learning_rate)
    # 设备
    device = torch.device('cuda:3' if torch.cuda.is_available() else 'cpu')
    model = model.to(device)
    model.train()
    best_loss = float('inf')
    for epoch in range(num_epochs):
        epoch_loss = 0
        for images, labels in dataloader:
            images = images.to(device)
            labels = labels.to(device)
            outputs = model(images)

            loss = criterion(outputs, labels)
            optimizer.zero_grad()
            loss.backward()
            optimizer.step()
            epoch_loss += loss.item()

        if epoch_loss < best_loss:
            best_loss = epoch_loss
            torch.save(model.state_dict(), f'./save_model_UNet/res_{epoch + 1}.pth')
        print(f'Epoch {epoch + 1}/{num_epochs}, Loss: {epoch_loss / len(dataloader)}')


def test():
    model = UNet()
    # 确保模型在CPU上
    model.load_state_dict(torch.load('./save_model_UNet/res_171.pth'))
    save_dir = './test_results_UNet'
    model.eval()
    dataset = SegmentationDataset('./dataset_exp2/test/image', './dataset_exp2/test/label')
    dataloader = DataLoader(batch_size=1, shuffle=False, dataset=dataset)
    iou_list = []
    dice_list = []
    precision_list = []
    recall_list = []
    f1_list = []
    plt.ion()
    with torch.no_grad():
        for idx, (images, labels) in tqdm(enumerate(dataloader)):
            pre = model(images)
            img_pre = torch.squeeze(pre)
            img_true = torch.squeeze(labels)
            iou = compute_iou(img_pre, img_true)
            dice = compute_dice(img_pre, img_true)
            precision, recall, f1_score = compute_precision_recall_f1(img_pre, img_true)
            img_pre = img_pre.numpy()
            img_true = img_true.numpy()
            img_x = torch.squeeze(images).numpy().transpose(1, 2, 0)
            img_x = (img_x * 255).astype(np.uint8)  # 恢复到0-255的范围
            # 保存结果
            plt.figure(figsize=(12, 4))
            plt.subplot(1, 3, 1)
            plt.title('Input Image')
            plt.imshow(img_x)
            plt.axis('off')

            plt.subplot(1, 3, 2)
            plt.title('True Mask')
            plt.imshow(img_true, cmap='gray')
            plt.axis('off')

            plt.subplot(1, 3, 3)
            plt.title('UNet Predicted Mask')
            plt.imshow(img_pre, cmap='gray')
            plt.axis('off')

            plt.savefig(os.path.join(save_dir, f'result_{idx + 1}.png'))
            plt.close()  # 关闭当前figure,避免内存占用过多

            iou_list.append(iou.item())
            dice_list.append(dice.item())
            precision_list.append(precision)
            recall_list.append(recall)
            f1_list.append(f1_score)

        plt.ioff()  # 关闭交互模式
        print(f'Results saved in {save_dir}')
        print(f'Average IoU: {np.mean(iou_list)}')
        print(f'Average Dice Coefficient: {np.mean(dice_list)}')
        print(f'Average Precision: {np.mean(precision_list)}')
        print(f'Average Recall: {np.mean(recall_list)}')
        print(f'Average F1 Score: {np.mean(f1_list)}')

if __name__ == '__main__':
    print('++++++++++++++++train++++++++++++++++')
    train()
    print('++++++++++++++++test++++++++++++++++')
    test()

测试效果:
在这里插入图片描述
在这里插入图片描述

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 1
    评论
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

魔云连洲

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值