forked from wiseodd/controlled-text-generation
-
Notifications
You must be signed in to change notification settings - Fork 0
/
test.py
103 lines (74 loc) · 2.47 KB
/
test.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
import math
import os
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.autograd as autograd
import torch.optim as optim
import numpy as np
from torch.autograd import Variable
from ctextgen.dataset import SST_Dataset
from ctextgen.model import RNN_VAE
import argparse
import random
import time
parser = argparse.ArgumentParser(
description='Conditional Text Generation'
)
parser.add_argument('--gpu', default=False, action='store_true',
help='whether to run in the GPU')
parser.add_argument('--model', default='ctextgen', metavar='',
help='choose the model: {`vae`, `ctextgen`}, (default: `ctextgen`)')
args = parser.parse_args()
mb_size = 32
z_dim = 20
h_dim = 64
lr = 1e-3
lr_decay_every = 1000000
n_iter = 20000
log_interval = 1000
z_dim = h_dim
c_dim = 2
dataset = SST_Dataset()
torch.manual_seed(int(time.time()))
model = RNN_VAE(
dataset.n_vocab, h_dim, z_dim, c_dim, p_word_dropout=0.3,
pretrained_embeddings=dataset.get_vocab_vectors(), freeze_embeddings=True,
gpu=args.gpu
)
if args.gpu:
model.load_state_dict(torch.load('models/{}.bin'.format(args.model)))
else:
model.load_state_dict(torch.load('models/{}.bin'.format(args.model), map_location=lambda storage, loc: storage))
# Samples latent and conditional codes randomly from prior
z = model.sample_z_prior(1)
c = model.sample_c_prior(1)
# Generate positive sample given z
c[0, 0], c[0, 1] = 1, 0
_, c_idx = torch.max(c, dim=1)
sample_idxs = model.sample_sentence(z, c, temp=0.1)
print('\nSentiment: {}'.format(dataset.idx2label(int(c_idx))))
print('Generated: {}'.format(dataset.idxs2sentence(sample_idxs)))
# Generate negative sample from the same z
c[0, 0], c[0, 1] = 0, 1
_, c_idx = torch.max(c, dim=1)
sample_idxs = model.sample_sentence(z, c, temp=0.8)
print('\nSentiment: {}'.format(dataset.idx2label(int(c_idx))))
print('Generated: {}'.format(dataset.idxs2sentence(sample_idxs)))
print()
# Interpolation
c = model.sample_c_prior(1)
z1 = model.sample_z_prior(1).view(1, 1, z_dim)
z1 = z1.cuda() if args.gpu else z1
z2 = model.sample_z_prior(1).view(1, 1, z_dim)
z2 = z2.cuda() if args.gpu else z2
# Interpolation coefficients
alphas = np.linspace(0, 1, 5)
print('Interpolation of z:')
print('-------------------')
for alpha in alphas:
z = float(1-alpha)*z1 + float(alpha)*z2
sample_idxs = model.sample_sentence(z, c, temp=0.1)
sample_sent = dataset.idxs2sentence(sample_idxs)
print("{}".format(sample_sent))
print()