VAE实战


# coding: utf-8

# In[3]:VAE.py


import torch
from torch import nn

class VAE(nn.Module):
    
    
    def __init__(self):
        
        super (VAE,self).__init__()
        # [b, 784] => [b, 20]
        # u: [b, 10]
        # sigma: [b, 10]
        self.encoder=nn.Sequential(
        nn.Linear(784,256),
        nn.ReLU(),
        nn.Linear(256,64),
        nn.ReLU(),
        nn.Linear(64,20),
        nn.ReLU()
        )
        self.decoder = nn.Sequential(
        nn.Linear(10,64),
        nn.ReLU(),
        nn.Linear(64,256),
        nn.ReLU(),
        nn.Linear(256,784))
    def forward(self,x):
        """

        :param x: [b, 1, 28, 28]
        :return:
        """
        batchsz = x.size(0)
        #flatten
        x = x.view(batchsz,784)
        #encoder
        # [b, 20], including mean and sigma 中间隐藏20个值包含均值和方差
        h_=self.encoder(x)
        # [b, 20] => [b, 10] and [b, 10]
        mu,sigma =h_.chunk(2,dim=1)  #在第一维拆成两部分
        
        h= mu +sigma*torch.randn_like(sigma)
        
        #decoder
        
        x_hat =self.decoder(h)
        
        #reshape
        
        x_hat =x_hat.view(batchsz,1,28,28)
        
        #求每个像素的kld 散度 散度越小两个分布越接近
        
        kld = (0.5 *torch.sum(torch.pow(mu,2)+torch.pow(sigma,2)-torch.log(1e-8+torch.pow(sigma,2))-1))/batchsz*28*28
        
        return x_hat,kld
        


#VAEmain 函数

import torch
from torch.utils.data import  DataLoader
from torch import nn,optim
from torchvision import transforms,datasets
from VAEmain import  VAE

import visdom


def main():
    mnist_train = datasets.MNIST('../data',True,transform=transforms.Compose([transforms.ToTensor()]),download=False)
    mnist_train=DataLoader(mnist_train,batch_size=32,shuffle=True)
    mnist_test = datasets.MNIST('../data',False,transform=transforms.Compose([transforms.ToTensor()]),download=False)
    mnist_test=DataLoader(mnist_test,batch_size=32,shuffle=True)
    print(1)
    x, _ = iter(mnist_train).next()
    print("x:",x.shape)
    
    device =torch.device('cuda')
    
    model =VAE().to(device)
    criteon = nn.MSELoss()
    optimizer =optim.Adam(model.parameters(),lr=1e-2)
    
    viz=visdom.Visdom()
    for epoch in range(1000):
        
        for batchid,(x,_) in enumerate(mnist_train):
            x=x.to(device)
            x_hat,kld=model(x)
            loss=criteon(x_hat,x)
            if kld is not None:
                elbo=-loss-1.0*kld
                loss=-elbo
            optimizer.zero_grad()
            loss.backward()
            optimizer.step()
        print(epoch,'loss:',loss.item(),'kld',kld.item())
        
        x,_ =iter(mnist_test).next()
        x=x.to(device)
        with torch.no_grad():#测试的时候没有梯度信息
            x_hat,kld =model(x)
            viz.images(x,nrow=8,win='x',opts=dict(title='x'))
            viz.images(x_hat,nrow=8,win='x_hat',opts=dict(title='x_hat')) 
    
if __name__=='__main__':
    main()
        
        


  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值