简单的图像神经风格迁移(Pytorch)

import torch
import torch.nn as nn
import torchvision.transforms as transforms
import torch.nn.functional as F
from PIL import Image
import matplotlib
import matplotlib.pyplot as plt
%matplotlib inline
import copy
from torchvision.models import vgg19, VGG19_Weights

Checking for GPU Availaibility

device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
torch.set_default_device(device)
device
device(type='cuda')
content_image_directory = 'content_image.jpg'
style_image_directory = 'style_image.jpg'

Initial DataPreprocessing and basic functions

image_size = 512 if torch.cuda.is_available() else 128
transform = transforms.Compose([
                  transforms.Resize(image_size),
                  transforms.ToTensor()])
def image_loader(image_directory):
  image = Image.open(image_directory)
  image = transform(image).unsqueeze(0)
  return image.to(device, torch.float)
content_img = image_loader(content_image_directory)
style_img = image_loader(style_image_directory)
assert content_img.size() == style_img.size()
reform = transforms.ToPILImage()


def image_unloader(tensor_input,title=None):
  image = tensor_input.cpu().clone()
  image = image.squeeze(0)
  image = reform(image)
  plt.imshow(image)
  plt.title(title)
  plt.show()
  plt.close()
plt.figure()
image_unloader(content_img, title = 'Content Image')
plt.figure()
image_unloader(style_img, title = 'Style Image')

Loss Functions for Custom Model

class Content_Loss(nn.Module):
  def __init__(self, target):
    super(Content_Loss, self).__init__()
    self.target = target.detach()
  def forward(self, input):
    self.loss = F.mse_loss(input, self.target)
    return input
def gram_matrix(matrix):
  a,b,c,d = matrix.size()
  features = matrix.view(a*b, c*d)
  G = torch.mm(features, features.t())
  return G.div(a*b*c*d)
class Style_Loss(nn.Module):
  def __init__(self, target):
    super(Style_Loss,self).__init__()
    self.target = gram_matrix(target).detach()
  def forward(self, input):
    G_input = gram_matrix(input)
    self.loss = F.mse_loss(G_input, self.target)
    return input

Pretrained VGG19

vgg = vgg19(weights=VGG19_Weights.DEFAULT).features.eval()
vgg
Sequential(
  (0): Conv2d(3, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (1): ReLU(inplace=True)
  (2): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (3): ReLU(inplace=True)
  (4): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
  (5): Conv2d(64, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (6): ReLU(inplace=True)
  (7): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (8): ReLU(inplace=True)
  (9): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
  (10): Conv2d(128, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (11): ReLU(inplace=True)
  (12): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (13): ReLU(inplace=True)
  (14): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (15): ReLU(inplace=True)
  (16): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (17): ReLU(inplace=True)
  (18): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
  (19): Conv2d(256, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (20): ReLU(inplace=True)
  (21): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (22): ReLU(inplace=True)
  (23): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (24): ReLU(inplace=True)
  (25): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (26): ReLU(inplace=True)
  (27): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
  (28): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (29): ReLU(inplace=True)
  (30): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (31): ReLU(inplace=True)
  (32): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (33): ReLU(inplace=True)
  (34): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (35): ReLU(inplace=True)
  (36): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
)vgg_normalization_mean = torch.tensor([0.485, 0.456, 0.406])

vgg_normalization_std = torch.tensor([0.229, 0.224, 0.225])
class Normalization(nn.Module):

  def __init__(self, mean, std):

    super(Normalization, self).__init__()

    self.mean = torch.tensor(mean).view(-1, 1, 1)

    self.std = torch.tensor(std).view(-1, 1, 1)




  def forward(self, img):

    return (img-self.mean)/self.stdContent Layers and Style Layers
content_layers = ['conv_4']
style_layers = ['conv_1', 'conv_2', 'conv_3', 'conv_4', 'conv_5']
for i in vgg.children():
  print(i)
  break
Conv2d(3, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))

Function for getting style and content losses

def style_and_content_losses(pre_model = vgg, normalization_mean=vgg_normalization_mean,
                             normalization_std=vgg_normalization_std,
                                content_layers=content_layers, style_layers=style_layers,
                                content_img=content_img, style_img=style_img):
  normalization = Normalization(vgg_normalization_mean, vgg_normalization_std)


  content_losses = []
  style_losses = []


  model = nn.Sequential(normalization)


  i = 0
  for layer in vgg.children():
    if isinstance(layer, nn.Conv2d):
      i+=1
      name = 'conv_{}'.format(i)
    elif isinstance(layer, nn.ReLU):
      name = 'ReLU_{}'.format(i)
      layer = nn.ReLU(inplace=False)
    elif isinstance(layer, nn.MaxPool2d):
      name = 'pool_{}'.format(i)
    elif isinstance(layer, nn.BatchNorm2d):
      name = 'batch_norm_{}'.format(i)
    else:
      raise RuntimeError('Unrecongnized_Layer:{}'.format(layer.__class__.__name__))
    model.add_module(name,layer)


    if name in content_layers:
      target = model(content_img).detach()
      content_loss = Content_Loss(target)
      model.add_module('content_loss_{}'.format(i), content_loss)
      content_losses.append(content_loss)
    if name in style_layers:
      target_feature = model(style_img).detach()
      style_loss = Style_Loss(target=target_feature)
      model.add_module('style_loss_{}'.format(i), style_loss)
      style_losses.append(style_loss)


  for i in range(len(model)-1, -1, -1):
    if isinstance(model[i], Content_Loss) or isinstance(model[i], Style_Loss):
      break


  model = model[:(i+1)]
  return model, style_losses, content_losses

Input Image same as content Image

input_img = content_img.clone()
plt.figure()
image_unloader(input_img, title = 'Input_Image')
def get_optimizer(input_img):
  optimizer = torch.optim.LBFGS([input_img])
  return optimizer

Final Function for Neural Transfer

def style_transfer(pre_model=vgg, normalization_mean=vgg_normalization_mean,
                   normalization_std=vgg_normalization_std,
                   content_img=content_img, style_img=style_img,
                   input_img = input_img, num_epochs=400,
                   style_weight=1000000, content_weight=1):
  print('Building the style transfer model')
  model, style_losses, content_losses = style_and_content_losses()
  input_img.requires_grad_(True)
  model.eval()
  model.requires_grad_(False)
  optimizer= get_optimizer(input_img)


  print('Optimizing...')
  run = [0]
  while run[0]<=num_epochs:
    def closure():
      with torch.no_grad():
        input_img.clamp_(0,1)
      optimizer.zero_grad()
      model(input_img)
      style_score=0
      content_score = 0
      for s_loss in style_losses:
        style_score += s_loss.loss
      for c_loss in content_losses:
        content_score += c_loss.loss


      style_score *= style_weight
      content_score *= content_weight


      loss = style_score + content_score
      loss.backward()


      run[0] += 1
      if run[0] % 50 == 0:
        print('run {}:'.format(run))
        print('Style_loss : {:4f} & Content_loss : {:4f}'.format(
            style_score.item(), content_score.item()
        ))
        print()
      return style_score + content_score
    optimizer.step(closure)


  with torch.no_grad():
        input_img.clamp_(0, 1)


  return input_img

Getting Output

output = style_transfer()
Optimizing...
run [50]:
Style_loss : 4.008237 & Content_loss : 4.146012

run [100]:
Style_loss : 1.135739 & Content_loss : 3.031179

run [150]:
Style_loss : 0.714863 & Content_loss : 2.651623

run [200]:
Style_loss : 0.481216 & Content_loss : 2.490812

run [250]:
Style_loss : 0.349222 & Content_loss : 2.403970

run [300]:
Style_loss : 0.264658 & Content_loss : 2.349341

run [350]:
Style_loss : 0.214068 & Content_loss : 2.314269

run [400]:
Style_loss : 0.184173 & Content_loss : 2.288790

The Output

plt.figure()
image_unloader(output, title='Output Image')


plt.ioff()
plt.show()

Initialization with a white noise image

input_img2 = torch.randn(content_img.data.size())
plt.figure()
image_unloader(input_img2, title = 'white noise image')
output2 = style_transfer(input_img=input_img2,num_epochs=700)
Building the style transfer model
Optimizing...
run [50]:
Style_loss : 83.288612 & Content_loss : 11.821552

run [100]:
Style_loss : 23.298397 & Content_loss : 9.524027

run [150]:
Style_loss : 5.048073 & Content_loss : 7.467306

run [200]:
Style_loss : 1.898000 & Content_loss : 5.806565

run [250]:
Style_loss : 1.291034 & Content_loss : 4.793440

run [300]:
Style_loss : 0.958348 & Content_loss : 4.172914

run [350]:
Style_loss : 0.734883 & Content_loss : 3.720185

run [400]:
Style_loss : 0.569834 & Content_loss : 3.423393

run [450]:
Style_loss : 0.448026 & Content_loss : 3.190432

run [500]:
Style_loss : 0.360451 & Content_loss : 3.018604

run [550]:
Style_loss : 0.303826 & Content_loss : 2.879138

run [600]:
Style_loss : 0.262109 & Content_loss : 2.767931

run [650]:
Style_loss : 0.232545 & Content_loss : 2.680881

run [700]:
Style_loss : 0.209969 & Content_loss : 2.609402
plt.figure()
image_unloader(output2, title='Output2 Image')
知乎学术咨询:https://www.zhihu.com/consult/people/792359672131756032?isMe=1

担任《Mechanical System and Signal Processing》等审稿专家,擅长领域:现代信号处理,机器学习,深度学习,数字孪生,时间序列分析,设备缺陷检测、设备异常检测、设备智能故障诊断与健康管理PHM等。

相关推荐
一点媛艺8 分钟前
Kotlin函数由易到难
开发语言·python·kotlin
qzhqbb40 分钟前
基于统计方法的语言模型
人工智能·语言模型·easyui
冷眼看人间恩怨1 小时前
【话题讨论】AI大模型重塑软件开发:定义、应用、优势与挑战
人工智能·ai编程·软件开发
2401_883041081 小时前
新锐品牌电商代运营公司都有哪些?
大数据·人工智能
魔道不误砍柴功1 小时前
Java 中如何巧妙应用 Function 让方法复用性更强
java·开发语言·python
pianmian11 小时前
python数据结构基础(7)
数据结构·算法
_.Switch2 小时前
高级Python自动化运维:容器安全与网络策略的深度解析
运维·网络·python·安全·自动化·devops
AI极客菌2 小时前
Controlnet作者新作IC-light V2:基于FLUX训练,支持处理风格化图像,细节远高于SD1.5。
人工智能·计算机视觉·ai作画·stable diffusion·aigc·flux·人工智能作画
阿_旭2 小时前
一文读懂| 自注意力与交叉注意力机制在计算机视觉中作用与基本原理
人工智能·深度学习·计算机视觉·cross-attention·self-attention
王哈哈^_^2 小时前
【数据集】【YOLO】【目标检测】交通事故识别数据集 8939 张,YOLO道路事故目标检测实战训练教程!
前端·人工智能·深度学习·yolo·目标检测·计算机视觉·pyqt