【学习笔记】Pytorch-两层BiLSTM情感计算Demo代码解读

BiLSTM for Sentiment Computing Demo

  • 模型: 两层、双向LSTM
  • 数据集:IMDB
  • 环境:
  • Python3.7
  • torch==1.10.0
  • torchtext==0.11.0
  • spacy==2.2.4
  • 相关代码参考自:https://www.bilibili.com/video/BV1Rv411y7oE?p=75
import torch
import torch.nn as nn
import torch.optim as optim
from torchtext.legacy import data, datasets

在神经网络中,参数默认是进行随机初始化的。如果不设置的话每次训练时的初始化都是随机的,导致结果不确定。如果设置初始化,则 每次初始化都是固定的。

print('GPU:', torch.cuda.is_available())

torch.cuda.manual_seed(123)

输出:GPU: True

1.加载数据


TEXT = data.Field(tokenize='spacy')

LABEL = data.LabelField(dtype=torch.float)

train_data,test_data = datasets.IMDB.splits(TEXT, LABEL)

查看数据集样本个数和某个文本数据的内容

print('len of train data:', len(train_data))
print('len of test data:', len(test_data))
print(train_data.examples[15].text)
print(train_data.examples[15].label)

output:
len of train data: 25000
len of test data: 25000
[‘A’, ‘bus’, ‘full’, ‘of’, ‘passengers’, ‘is’, ‘stuck’, ‘during’, ‘a’, ‘snow’, ‘storm’, ‘.’, ‘The’, ‘police’, ‘have’, ‘closed’, ‘the’, ‘bridge’, ‘–’, ‘saying’, ‘it’, “‘s”, ‘unsafe’, ‘and’, ‘they’, ‘are’, ‘stuck’, ‘in’, ‘a’, ‘little’, ‘café’, ‘until’, ‘the’, ‘road’, ‘has’, ‘been’, ‘cleared’, ‘.’, ‘However’, ‘,’, ‘after’, ‘a’, ‘while’, ‘,’, ‘their’, ‘boredom’, ‘is’, ‘turned’, ‘to’, ‘concern’, ‘,’, ‘as’, ‘it’, ‘seems’, ‘that’, ‘one’, ‘of’, ‘the’, ‘passengers’, ‘was’, ‘NOT’, ‘originally’, ‘on’, ‘the’, ‘bus’, ‘and’, ‘may’, ‘just’, ‘be’, ‘an’, ‘alien’, ‘!’, ‘!’, ‘This’, ‘leads’, ‘to’, ‘a’, ‘conclusion’, ‘that’, ‘is’, ‘ironic’, ‘but’, ‘also’, ‘rather’, ‘funny’, ‘in’, ‘a’, ‘low’, ‘-‘, ‘brow’, ‘way.

pos

利用训练集构建词表,vocabulary把每个单词一一映射到一个数字。使用10k个单词来构建单词表(用max_size这个参数可以设定),所有其他的单词都用来表示。映射到数字后使用vectors创建词向量。

TEXT.build_vocab(train_data, max_size=10000, vectors='glove.6B.100d')

LABEL.build_vocab(train_data)

print(TEXT.vocab.stoi)
print(LABEL.vocab.stoi)

batchsz = 32

device = torch.device('cuda')

创建迭代器,相当于torch的DataLoader。每个iterator中各有两部分:词(.text)和标签(.label),其中text全部转换成数字了。

BucketIterator会把长度差不多的句子放到同一个batch中,确保每个batch中不出现太多的padding。这里因为pad比较少,所以把也当做了模型的输入进行训练。如果有GPU,还可以指定每个iteration返回的tensor都在GPU上。

train_iterator, test_iterator = data.BucketIterator.splits(
    (train_data, test_data),
    batch_size = batchsz,
    device=device
)

2.定义模型

举个例子,我们定义一个num_layers=3的双向LSTM。
h_n第一个维度的大小就等于 6 (23),
h_n[0]表示
第一层前向传播最后一个time step的输出,h_n[1]表示 第一层后向传播*最后一个time step的输出;
h_n[2]表示第二层前向传播最后一个time step的输出,h_n[3]表示第二层后向传播最后一个time step的输出;
h_n[4]和h_n[5]分别表示第三层前向和后向传播时最后一个time step的输出。

class RNN(nn.Module):

    def __init__(self, vocab_size, embedding_dim, hidden_dim):

        super(RNN, self).__init__()

        self.embedding = nn.Embedding(vocab_size, embedding_dim)

        self.rnn = nn.LSTM(embedding_dim, hidden_dim, num_layers=2,
                           bidirectional=True, dropout=0.5)

        self.fc = nn.Linear(hidden_dim*2, 1)
        self.dropout = nn.Dropout(0.5)

    def forward(self, x):
"""
        x: [seq_len, b] vs [b, 3, 28, 28]
"""

        embedding = self.dropout(self.embedding(x))

        output, (hidden, cell) = self.rnn(embedding)

        hidden = torch.cat([hidden[-2], hidden[-1]], dim=1)

        hidden = self.dropout(hidden)
        out = self.fc(hidden)

        return out

3.初始化模型


rnn = RNN(len(TEXT.vocab), 100, 256)

pretrained_embedding = TEXT.vocab.vectors
print('pretrained_embedding:', pretrained_embedding.shape)
rnn.embedding.weight.data.copy_(pretrained_embedding)
print('embedding layer inited.')

optimizer = optim.Adam(rnn.parameters(), lr=1e-3)

criterion = nn.BCEWithLogitsLoss().to(device)
rnn.to(device)

输出

pretrained_embedding: torch.Size([10002, 100])
embedding layer inited.

RNN(
  (embedding): Embedding(10002, 100)
  (rnn): LSTM(100, 256, num_layers=2, dropout=0.5, bidirectional=True)
  (fc): Linear(in_features=512, out_features=1, bias=True)
  (dropout): Dropout(p=0.5, inplace=False)
)

4.定义评估指标函数

import numpy as np

def binary_acc(preds, y):
"""
    get accuracy
"""

    preds = torch.round(torch.sigmoid(preds))
    correct = torch.eq(preds, y).float()
    acc = correct.sum() / len(correct)
    return acc

5.定义训练和测试函数


def train(rnn, iterator, optimizer, criterion):

    avg_acc = []
    rnn.train()

    for i, batch in enumerate(iterator):

        pred = rnn(batch.text).squeeze(1)

        loss = criterion(pred, batch.label)
        acc = binary_acc(pred, batch.label).item()
        avg_acc.append(acc)

        optimizer.zero_grad()
        loss.backward()
        optimizer.step()

        if i%10 == 0:
            print(i, acc)

    avg_acc = np.array(avg_acc).mean()
    print('avg acc:', avg_acc)

def eval(rnn, iterator, criterion):

    avg_acc = []

    rnn.eval()

    with torch.no_grad():
        for batch in iterator:

            pred = rnn(batch.text).squeeze(1)

            loss = criterion(pred, batch.label)

            acc = binary_acc(pred, batch.label).item()
            avg_acc.append(acc)

    avg_acc = np.array(avg_acc).mean()

    print('>>test:', avg_acc)

6.训练

for epoch in range(10):
    train(rnn, train_iterator, optimizer, criterion)
    eval(rnn, test_iterator, criterion)

Original: https://blog.csdn.net/m0_47779101/article/details/123854237
Author: JoonleonWong
Title: 【学习笔记】Pytorch-两层BiLSTM情感计算Demo代码解读

原创文章受到原创版权保护。转载请注明出处:https://www.johngo689.com/528220/

转载文章受原作者版权保护。转载请注明原作者出处!

(0)

大家都在看

亲爱的 Coder【最近整理,可免费获取】👉 最新必读书单  | 👏 面试题下载  | 🌎 免费的AI知识星球