标签:lib return gray forms var 实例 训练 ISE 加载
去噪自编码器模拟人类视觉机制能够自动忍受图像的噪声来识别图片。自编码器的目标是要学习一个近似的恒等函数,使得输出近似等于输入。去噪自编码器采用随机的部分带噪输入来解决恒等函数问题,自编码器能够获得输入的良好表征,该表征使得自编码器能进行去噪或恢复。
下面是代码:
#加载库和配置参数 #去噪自编码器 import torch import torch.nn as nn import torch.utils as utils from torch.autograd import Variable import torchvision.datasets as dset import torchvision.transforms as transforms import numpy as np import matplotlib.pyplot as plt #配置参数 torch.manual_seed(1) n_epoch=200 batch_size=100 learning_rate=0.002 #下载图片库训练集 mnist_train=dset.MNIST("./",train=True,transform=transforms.ToTensor(),target_transform=None,download=True) train_loader=torch.utils.data.DataLoader(dataset=mnist_train,batch_size=batch_size,shuffle=True) #Encoder和Decoder模型设置 class Encoder(nn.Module): def __init__(self): super(Encoder,self).__init__() self.layer1=nn.Sequential( nn.Conv2d(1,32,3,padding=1),#batch*32*28*28 nn.ReLU(), nn.BatchNorm2d(32), nn.Conv2d(32,32,3,padding=1),#batch*32*28*28 nn.ReLU(), nn.BatchNorm2d(32), nn.Conv2d(32,64,3,padding=1),#batch*64*28*28 nn.ReLU(), nn.BatchNorm2d(64), nn.Conv2d(64,64,3,padding=1),#batch*64*28*28 nn.ReLU(), nn.BatchNorm2d(64), nn.MaxPool2d(2,2)#batch*64*14*14 ) self.layer2=nn.Sequential( nn.Conv2d(64,128,3,padding=1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.Conv2d(128,128,3,padding=1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.MaxPool2d(2,2), nn.Conv2d(128,256,3,padding=1),#batch*256*14*14 nn.ReLU() ) def forward(self,x): out=self.layer1(x) out=self.layer2(out) out=out.view(batch_size,-1) return out encoder=Encoder().cuda() #decoder设置 class Decoder(nn.Module): def __init__(self): super(Decoder,self).__init__() self.layer1=nn.Sequential( nn.ConvTranspose2d(256,128,3,2,1,1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.ConvTranspose2d(128,128,3,1,1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.ConvTranspose2d(128,64,3,1,1),#batch*64*14*14 nn.ReLU(), nn.BatchNorm2d(64), nn.ConvTranspose2d(64,64,3,1,1),#batch*64*14*14 nn.ReLU(), nn.BatchNorm2d(64) ) self.layer2=nn.Sequential( nn.ConvTranspose2d(64,32,3,1,1),#batch*32*14*14 nn.ReLU(), nn.BatchNorm2d(32), nn.ConvTranspose2d(32,32,3,1,1),#batch*32*14*14 nn.ReLU(), nn.BatchNorm2d(32), nn.ConvTranspose2d(32,1,3,2,1,1),#batch*1*28*28 nn.ReLU() ) def forward(self,x): out=x.view(batch_size,256,7,7) out=self.layer1(out) out=self.layer2(out) return out decoder=Decoder().cuda() ###Loss 函数和优化器 parameters=list(encoder.parameters())+list(decoder.parameters()) loss_func=nn.MSELoss() optimizer=torch.optim.Adam(parameters,lr=learning_rate) ###自编码器训练 #添加噪声 noise=torch.rand(batch_size,1,28,28) for I in range(n_epoch): for image,label in train_loader: image_n=torch.mul(image+0.25,0.1*noise) image=Variable(image).cuda() image_n=Variable(image_n).cuda() optimizer.zero_grad() output=encoder(image_n) output=decoder(output) loss=loss_func(output,image) loss.backward() optimizer.step() break print(‘epoch[{}/{}],loss:{:.4f}‘.format(I+1,n_epoch,loss.item())) ####带噪图片和去噪图片对比 img=image[0].cpu() input_img=image_n[0].cpu() output_img=output[0].cpu() origin=img.data.numpy() inp=input_img.data.numpy() out=output_img.data.numpy() plt.figure(‘denoising autoencoder‘) plt.subplot(131) plt.imshow(origin[0],cmap=‘gray‘) plt.subplot(132) plt.imshow(inp[0],cmap=‘gray‘) plt.subplot(133) plt.imshow(out[0],cmap=‘gray‘) plt.show() print(label[0])
标签:lib return gray forms var 实例 训练 ISE 加载
原文地址:https://www.cnblogs.com/candyRen/p/12113091.html