⽤方法 to() 可以将 Tensor 在CPU和GPU(需要硬件支持)之间相互移动

# 以下代码只有在PyTorch GPU版本上才会执⾏行行
if torch.cuda.is_available():
device = torch.device("cuda") # GPU
y = torch.ones_like(x, device=device) # 直接创建一个在GPU上的Tensor
x = x.to(device) # 等价于 .to("cuda")
z = x + y
print(z)
print(z.to("cpu", torch.double)) # to()还可以同时更改数据类型

 

  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
将下列生成器改造成能够匹配edge-connect中的InpaintingModel的预训练模型键值的结构:class Generator(nn.Module): def init(self): super(Generator, self).init() self.encoder = nn.Sequential( nn.Conv2d(3, 64, 3, stride=2, padding=1), nn.BatchNorm2d(64), nn.LeakyReLU(0.2), nn.Conv2d(64, 128, 3, stride=2, padding=1), nn.BatchNorm2d(128), nn.LeakyReLU(0.2), nn.Conv2d(128, 256, 3, stride=2, padding=1), nn.BatchNorm2d(256), nn.LeakyReLU(0.2), nn.Conv2d(256, 512, 3, stride=2, padding=1), nn.BatchNorm2d(512), nn.LeakyReLU(0.2), nn.Conv2d(512, 4000, 1), nn.BatchNorm2d(4000), nn.LeakyReLU(0.2) ) self.decoder = nn.Sequential( nn.ConvTranspose2d(4000, 512, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(512), nn.LeakyReLU(0.2), nn.ConvTranspose2d(512, 256, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(256), nn.LeakyReLU(0.2), nn.ConvTranspose2d(256, 128, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(128), nn.LeakyReLU(0.2), nn.ConvTranspose2d(128, 64, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(64), nn.LeakyReLU(0.2), nn.ConvTranspose2d(64, 3, 3, stride=1, padding=1), nn.Tanh() ) def forward(self, x): x = self.encoder(x) x = self.decoder(x) return x 另外修复部分代码定义为:mask = cv.inRange(img, (0, 0, 0), (1, 1, 1)) # 转换为张量 image_tensor = transforms.ToTensor()(img) mask_tensor = transforms.ToTensor()(mask) # 扩展维度 image_tensor = image_tensor.unsqueeze(0) mask_tensor = mask_tensor.unsqueeze(0) generator = Generator() load_edgeconnect_weights(generator, 'E:/fin/models/gen.pth') image_tensor = image_tensor.cuda() mask_tensor = mask_tensor.cuda() generator = generator.cuda() with torch.no_grad(): output_tensor = generator(image_tensor, mask_tensor)
05-11
以下是将生成器改造成能够匹配edge-connect中的InpaintingModel的预训练模型键值的结构的代码: import torch import torch.nn as nn import cv2 as cv from torchvision import transforms class Generator(nn.Module): def __init__(self): super(Generator, self).__init__() self.encoder = nn.Sequential( nn.Conv2d(3, 64, 3, stride=2, padding=1), nn.BatchNorm2d(64), nn.LeakyReLU(0.2), nn.Conv2d(64, 128, 3, stride=2, padding=1), nn.BatchNorm2d(128), nn.LeakyReLU(0.2), nn.Conv2d(128, 256, 3, stride=2, padding=1), nn.BatchNorm2d(256), nn.LeakyReLU(0.2), nn.Conv2d(256, 512, 3, stride=2, padding=1), nn.BatchNorm2d(512), nn.LeakyReLU(0.2), nn.Conv2d(512, 4000, 1), nn.BatchNorm2d(4000), nn.LeakyReLU(0.2) ) self.decoder = nn.Sequential( nn.ConvTranspose2d(4000, 512, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(512), nn.LeakyReLU(0.2), nn.ConvTranspose2d(512, 256, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(256), nn.LeakyReLU(0.2), nn.ConvTranspose2d(256, 128, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(128), nn.LeakyReLU(0.2), nn.ConvTranspose2d(128, 64, 3, stride=2, padding=1, output_padding=1), nn.BatchNorm2d(64), nn.LeakyReLU(0.2), nn.ConvTranspose2d(64, 3, 3, stride=1, padding=1), nn.Tanh() ) def forward(self, x, mask): x = x * (1 - mask) x = self.encoder(x) x = self.decoder(x) x = x * (1 - mask) + x * mask return x def load_edgeconnect_weights(model, weight_path): state_dict = torch.load(weight_path) new_state_dict = {} for key, value in state_dict.items(): if 'netG.' in key: new_key = key.replace('netG.', '') new_state_dict[new_key] = value model.load_state_dict(new_state_dict) # 读取图像和遮罩 img = cv.imread('example.jpg')[:, :, ::-1] / 255.0 mask = cv.inRange(img, (0, 0, 0), (1, 1, 1)) # 转换为张量 image_tensor = transforms.ToTensor()(img) mask_tensor = transforms.ToTensor()(mask) # 扩展维度 image_tensor = image_tensor.unsqueeze(0) mask_tensor = mask_tensor.unsqueeze(0) # 加载预训练模型权重 generator = Generator() load_edgeconnect_weights(generator, 'gen.pth') # 将张量和模型移动GPU上 image_tensor = image_tensor.cuda() mask_tensor = mask_tensor.cuda() generator = generator.cuda() # 使用预训练模型生成修复后的图像 with torch.no_grad(): output_tensor = generator(image_tensor, mask_tensor) # 将张量转换为图像 output = output_tensor.detach().cpu().squeeze().numpy().transpose(1, 2, 0) output = cv.cvtColor(output, cv.COLOR_RGB2BGR) cv.imwrite('output.jpg', output * 255)

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

Clark-dj

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值