matplotlib和tensorboard可视化(学习笔记)

本文记录了pytorch训练可视化的内容,分别用matplotlib和tensorboard实现,内容比较简单,适合入门,博客记录在此,仅备今后之用。

参考博客:

1.matplotlib

安装matplotlib包

pip install matplotlib

引入模块

import matplotlib.pyplot as plt

用plot绘图函数

tran_ls.append(running_loss / 500)
plt.plot(range(len(tran_ls)) ,tran_ls,color="red")
plt.xlabel("epoch")
plt.ylabel("loss")
plt.show()

喜闻乐见的代码:

import torch
import torchvision
import torch.optim as optim
import torchvision.transforms as transforms
from torch.utils.tensorboard import SummaryWriter
import torch.nn as nn
import torch.nn.functional as F
import matplotlib.pyplot as plt

class LeNet(nn.Module):
    def __init__(self):
        super(LeNet, self).__init__()
        self.conv1 = nn.Conv2d(3, 16, 5)
        self.pool1 = nn.MaxPool2d(2, 2)
        self.conv2 = nn.Conv2d(16, 32, 5)
        self.pool2 = nn.MaxPool2d(2, 2)
        self.fc1 = nn.Linear(32*5*5, 120)
        self.fc2 = nn.Linear(120, 84)
        self.fc3 = nn.Linear(84, 10)

    def forward(self, x):
        x = F.relu(self.conv1(x))
        x = self.pool1(x)
        x = F.relu(self.conv2(x))
        x = self.pool2(x)
        x = x.view(-1, 32*5*5)
        x = F.relu(self.fc1(x))
        x = F.relu(self.fc2(x))
        x = self.fc3(x)
        return x

def main():
    transform = transforms.Compose(
        [transforms.ToTensor(),
         transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))])

    train_set = torchvision.datasets.CIFAR10(root='./data', train=True,
                                             download=True, transform=transform)
    train_loader = torch.utils.data.DataLoader(train_set, batch_size=36,
                                               shuffle=True, num_workers=0)

    val_set = torchvision.datasets.CIFAR10(root='./data', train=False,
                                           download=True, transform=transform)
    val_loader = torch.utils.data.DataLoader(val_set, batch_size=5000,
                                             shuffle=False, num_workers=0)
    val_data_iter = iter(val_loader)
    val_image, val_label = val_data_iter.next()

    net = LeNet()

    loss_function = nn.CrossEntropyLoss()
    optimizer = optim.Adam(net.parameters(), lr=0.001)
    tran_ls = []
    for epoch in range(5):
        running_loss = 0.0
        for step, data in enumerate(train_loader, start=0):

            inputs, labels = data

            optimizer.zero_grad()

            outputs = net(inputs)
            loss = loss_function(outputs, labels)
            loss.backward()
            optimizer.step()

            running_loss += loss.item()
            if step % 500 == 499:
                with torch.no_grad():
                    outputs = net(val_image)
                    predict_y = torch.max(outputs, dim=1)[1]
                    accuracy = torch.eq(predict_y, val_label).sum().item() / val_label.size(0)

                    print('[%d, %5d] train_loss: %.3f  test_accuracy: %.3f' %
                          (epoch + 1, step + 1, running_loss / 500, accuracy))
                    tran_ls.append(running_loss / 500)
                    running_loss = 0.0
    plt.plot(range(len(tran_ls)) ,tran_ls,color="red")
    plt.xlabel("epoch")
    plt.ylabel("loss")
    plt.show()
    print('Finished Training')
    save_path = './Lenet.pth'
    torch.save(net.state_dict(), save_path)

if __name__ == '__main__':
    main()

2.tensorboard

安装tensorboard包

pip install tensorboard

引入模块

from torch.utils.tensorboard import SummaryWriter

实例化SummaryWriter对象

writer = SummaryWriter("runs/cifar10_experiment")

命令行启动tensorboard网页,打开网址进入tensorboard主页面

tensorboard --logdir=runs

模型可视化,tensorboard可以通过传入一个样本,写入graph到日志中,在网页上看到模型的结构


trainloader=train_loader
dataiter = iter(trainloader)

images, labels = dataiter.next()
writer.add_graph(net, images)

刷新页面即可看到模型

喜闻乐见的代码环节

import torch
import torchvision
import torch.optim as optim
import torchvision.transforms as transforms

from torch.utils.tensorboard import SummaryWriter

import torch.nn as nn
import torch.nn.functional as F

writer = SummaryWriter("runs/cifar10_experiment")

class LeNet(nn.Module):
    def __init__(self):
        super(LeNet, self).__init__()
        self.conv1 = nn.Conv2d(3, 16, 5)
        self.pool1 = nn.MaxPool2d(2, 2)
        self.conv2 = nn.Conv2d(16, 32, 5)
        self.pool2 = nn.MaxPool2d(2, 2)
        self.fc1 = nn.Linear(32*5*5, 120)
        self.fc2 = nn.Linear(120, 84)
        self.fc3 = nn.Linear(84, 10)

    def forward(self, x):
        x = F.relu(self.conv1(x))
        x = self.pool1(x)
        x = F.relu(self.conv2(x))
        x = self.pool2(x)
        x = x.view(-1, 32*5*5)
        x = F.relu(self.fc1(x))
        x = F.relu(self.fc2(x))
        x = self.fc3(x)
        return x

def main():
    transform = transforms.Compose(
        [transforms.ToTensor(),
         transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))])

    train_set = torchvision.datasets.CIFAR10(root='./data', train=True,
                                             download=True, transform=transform)
    train_loader = torch.utils.data.DataLoader(train_set, batch_size=36,
                                               shuffle=True, num_workers=0)

    val_set = torchvision.datasets.CIFAR10(root='./data', train=False,
                                           download=True, transform=transform)
    val_loader = torch.utils.data.DataLoader(val_set, batch_size=5000,
                                             shuffle=False, num_workers=0)
    val_data_iter = iter(val_loader)
    val_image, val_label = val_data_iter.next()

    net = LeNet()

    trainloader=train_loader
    dataiter = iter(trainloader)

    images, labels = dataiter.next()
    writer.add_graph(net, images)
    loss_function = nn.CrossEntropyLoss()
    optimizer = optim.Adam(net.parameters(), lr=0.001)

    for epoch in range(20):
        running_loss = 0.0
        for step, data in enumerate(train_loader, start=0):

            inputs, labels = data

            optimizer.zero_grad()

            outputs = net(inputs)
            loss = loss_function(outputs, labels)
            loss.backward()
            optimizer.step()

            running_loss += loss.item()
            if step % 500 == 499:
                with torch.no_grad():
                    outputs = net(val_image)
                    predict_y = torch.max(outputs, dim=1)[1]
                    accuracy = torch.eq(predict_y, val_label).sum().item() / val_label.size(0)

                    print('[%d, %5d] train_loss: %.3f  test_accuracy: %.3f' %
                          (epoch + 1, step + 1, running_loss / 500, accuracy))
                    writer.add_scalar('training loss',
                                      running_loss / 500,
                                      epoch )
                    running_loss = 0.0
    print('Finished Training')
    save_path = './Lenet.pth'
    torch.save(net.state_dict(), save_path)

if __name__ == '__main__':
    main()

Original: https://blog.csdn.net/lxh248866/article/details/120440505
Author: ee-redbull
Title: matplotlib和tensorboard可视化(学习笔记)

原创文章受到原创版权保护。转载请注明出处:https://www.johngo689.com/769262/

转载文章受原作者版权保护。转载请注明原作者出处!

(0)

大家都在看

亲爱的 Coder【最近整理,可免费获取】👉 最新必读书单  | 👏 面试题下载  | 🌎 免费的AI知识星球