pytorch代码实践--情感分类

我下次再写

导入包

import torch
from torchtext import data
from torchtext import datasets
import random
import torch
import torch.nn as nn
import torch.nn.functional as F
import spacy


SEED = 1234

torch.manual_seed(SEED)
torch.cuda.manual_seed(SEED)
torch.backends.cudnn.deterministic = True

# 安装spaCy
# pip install -U spacy
# python -m spacy download en以前       python -m spacy download en_core_web_sm现在
# en = spacy.load('en_core_web_sm')
TEXT = data.Field(tokenize='spacy', tokenizer_language='en_core_web_sm')
LABEL = data.LabelField(dtype=torch.float)

train_data, test_data = datasets.IMDB.splits(TEXT, LABEL)
print(f'Number of training examples: {len(train_data)}')
print(f'Number of testing examples: {len(test_data)}')
print(vars(train_data.examples[0]))

# train_data, valid_data = train_data.split(random_state=random.seed(SEED),split_ratio=0.7)
train_data, valid_data = train_data.split(random_state=random.seed(SEED))
# TEXT.build_vocab(train_data, max_size=25000)
# LABEL.build_vocab(train_data)
TEXT.build_vocab(train_data, max_size=25000, vectors="glove.6B.100d", unk_init=torch.Tensor.normal_)
LABEL.build_vocab(train_data)
print(TEXT.vocab.itos[:10])
print(LABEL.vocab.stoi)
BATCH_SIZE = 64

device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')

train_iterator, valid_iterator, test_iterator = data.BucketIterator.splits(
    (train_data, valid_data, test_data),
    batch_size=BATCH_SIZE,
    device=device)

# seq_len * batch_size
batch = next(iter(valid_iterator))


 定义模型weight average

class WordAVGModel(nn.Module):
    def __init__(self, vocab_size, embedding_size, output_size, pad_idx):
        super(WordAVGModel, self).__init__()
        self.embed = nn.Embedding(vocab_size, embedding_size, padding_idx=pad_idx)
        self.linear = nn.Linear(embedding_size, output_size)

    def forward(self, text):
        embedded = self.embed(text)  # [seq_len, batch_size, embedding_size]
        #         embedded = embedded.transpose(1,0) # [batch_size, seq_len, embedding_size]
        embedded = embedded.permute(1, 0, 2)  # [batch_size, seq_len, embedding_size]
        pooled = F.avg_pool2d(embedded, (embedded.shape[1], 1)).squeeze()  # [batch_size, embedding_size]
        return self.linear(pooled)

VOCAB_SIZE = len(TEXT.vocab)
EMBEDDING_SIZE = 100
OUTPUT_SIZE = 1
PAD_IDX = TEXT.vocab.stoi[TEXT.pad_token]

model = WordAVGModel(vocab_size=VOCAB_SIZE,
                     embedding_size=EMBEDDING_SIZE,
                     output_size=OUTPUT_SIZE,
                     pad_idx=PAD_IDX)


 定义模型 CNN

class CNN(nn.Module):
    def __init__(self, vocab_size, embedding_size, output_size, pad_idx, num_filters, filter_sizes, dropout):
        super(CNN, self).__init__()
        self.embed = nn.Embedding(vocab_size, embedding_size, padding_idx=pad_idx)
        self.convs = nn.ModuleList([
            nn.Conv2d(in_channels=1, out_channels=num_filters,
                      kernel_size=(fs, embedding_size))
            for fs in filter_sizes
        ])  # 3个CNN
        #         self.conv = nn.Conv2d(in_channels=1, out_channels=num_filters, kernel_size=(filter_size, embedding_size))
        self.linear = nn.Linear(num_filters * len(filter_sizes), output_size)
        self.dropout = nn.Dropout(dropout)

    def forward(self, text):
        text = text.permute(1, 0)  # [batch_size, seq_len]
        embedded = self.embed(text)  # [batch_size, seq_len, embedding_size]
        embedded = embedded.unsqueeze(1)  # # [batch_size, 1, seq_len, embedding_size]
        #         conved = F.relu(self.conv(embedded)) # [batch_size, num_filters, seq_len-filter_size+1, 1]
        #         conved = conved.squeeze(3) # [batch_size, num_filters, seq_len-filter_size+1]
        conved = [F.relu(conv(embedded)).squeeze(3) for conv in self.convs]
        # max over time pooling
        #         pooled = F.max_pool1d(conved, conved.shape[2]) # [batch_size, num_filters, 1]
        #         pooled = pooled.squeeze(2)
        pooled = [F.max_pool1d(conv, conv.shape[2]).squeeze(2) for conv in conved]
        pooled = torch.cat(pooled, dim=1)  # batch_size, 3*num_filters
        pooled = self.dropout(pooled)

        return self.linear(pooled)
#     Conv1d? 1x1 Conv2d?

# model = CNN(vocab_size=VOCAB_SIZE,
#            embedding_size=EMBEDDING_SIZE,
#            output_size=OUTPUT_SIZE,
#            pad_idx=PAD_IDX,
#            num_filters=100,
#            filter_sizes=[3,4,5],
#            dropout=0.5)

 定义模型RNN

class RNNModel(nn.Module):
    def __init__(self, vocab_size, embedding_size, output_size, pad_idx, hidden_size, dropout):
        super(RNNModel, self).__init__()
        self.embed = nn.Embedding(vocab_size, embedding_size, padding_idx=pad_idx)
        self.lstm = nn.LSTM(embedding_size, hidden_size, bidirectional=True, num_layers=2)
        self.linear = nn.Linear(hidden_size * 2, output_size)
        self.dropout = nn.Dropout(dropout)

    def forward(self, text):
        embedded = self.embed(text)  # [seq_len, batch_size, embedding_size]
        embedded = self.dropout(embedded)
        output, (hidden, cell) = self.lstm(embedded)

        # hidden: 2 * batch_size * hidden_size
        hidden = torch.cat([hidden[-1], hidden[-2]], dim=1)
        hidden = self.dropout(hidden.squeeze())
        return self.linear(hidden)

# model = RNNModel(vocab_size=VOCAB_SIZE,
#                  embedding_size=EMBEDDING_SIZE,
#                  output_size=OUTPUT_SIZE,
#                  pad_idx=PAD_IDX,
#                  hidden_size=100,
#                  dropout=0.5)



训练

# model
def count_parameters(model):
    return sum(p.numel() for p in model.parameters() if p.requires_grad)    #numel():帮你数一共有多少个参数

count_parameters(model)
pretrained_embedding = TEXT.vocab.vectors
model.embed.weight.data.copy_(pretrained_embedding)         # 这里实现了将预训练的词向量迁移过来使用

UNK_IDX = TEXT.vocab.stoi[TEXT.unk_token]
model.embed.weight.data[PAD_IDX] = torch.zeros(EMBEDDING_SIZE)
model.embed.weight.data[UNK_IDX] = torch.zeros(EMBEDDING_SIZE)

optimizer = torch.optim.Adam(model.parameters())
crit = nn.BCEWithLogitsLoss()       #loss function

model = model.to(device)
crit = crit.to(device)

def binary_accuracy(preds, y):
    rounded_preds = torch.round(torch.sigmoid(preds))
    correct = (rounded_preds == y).float()
    acc = correct.sum() / len(correct)
    return acc


def train(model, iterator, optimizer, crit):
    epoch_loss, epoch_acc = 0., 0.
    model.train()
    total_len = 0.
    for batch in iterator:
        preds = model(batch.text).squeeze()  # [batch_size]
        #         print(preds.shape, batch.label.shape)
        loss = crit(preds, batch.label)
        acc = binary_accuracy(preds, batch.label)

        # sgd
        optimizer.zero_grad()
        loss.backward()
        optimizer.step()

        epoch_loss += loss.item() * len(batch.label)
        epoch_acc += acc.item() * len(batch.label)
        total_len += len(batch.label)

    return epoch_loss / total_len, epoch_acc / total_len


def evaluate(model, iterator, crit):
    epoch_loss, epoch_acc = 0., 0.
    model.eval()
    total_len = 0.
    for batch in iterator:
        preds = model(batch.text).squeeze()
        loss = crit(preds, batch.label)
        acc = binary_accuracy(preds, batch.label)

        epoch_loss += loss.item() * len(batch.label)
        epoch_acc += acc.item() * len(batch.label)
        total_len += len(batch.label)
    model.train()

    return epoch_loss / total_len, epoch_acc / total_len


N_EPOCHS = 10
best_valid_acc = 0.
for epoch in range(N_EPOCHS):
    train_loss, train_acc = train(model, train_iterator, optimizer, crit)
    valid_loss, valid_acc = evaluate(model, valid_iterator, crit)

    if valid_acc > best_valid_acc:
        best_valid_acc = valid_acc
        torch.save(model.state_dict(), "wordavg-model.pth")

    print("Epoch", epoch, "Train Loss", train_loss, "Train Acc", train_acc)
    print("Epoch", epoch, "Valid Loss", valid_loss, "Valid Acc", valid_acc)



 测试

# model.load_state_dict(torch.load("wordavg-model.pth"))
# import spacy
# nlp = spacy.load("en")
#
# def predict_sentiment(sentence):
#     tokenized = [tok.text for tok in nlp.tokenizer(sentence)]
#     indexed = [TEXT.vocab.stoi[t] for t in tokenized]
#     tensor = torch.LongTensor(indexed).to(device) # seq_len
#     tensor = tensor.unsqueeze(1) # seq_len * batch_size(1)
#     pred = torch.sigmoid(model(tensor))
#     return pred.item()
#
# predict_sentiment("This film is horrible!")

 

  • 1
    点赞
  • 2
    收藏
    觉得还不错? 一键收藏
  • 1
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值