-
Notifications
You must be signed in to change notification settings - Fork 11
/
Copy pathtransfer_learning.py
72 lines (62 loc) · 2.59 KB
/
transfer_learning.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
#!/usr/bin/env python
import torch
from torch.utils.data import DataLoader
import pickle
from rdkit import Chem
from tqdm import tqdm
import data_struct as ds
from data_struct import MolData, Vocabulary
from data_struct import Variable, decrease_learning_rate
from model import RNN
import sys
def Transfer(restore_from = None):
"""Trains the Prior RNN"""
voc = Vocabulary(init_from_file="./Voc")
moldata = MolData("tl_filtered.smi", voc)
data = DataLoader(moldata, batch_size=32, shuffle=True, drop_last=True,
collate_fn=MolData.collate_fn)
Prior = RNN(voc)
# Can restore from a saved RNN
if restore_from:
Prior.rnn.load_state_dict(torch.load(restore_from))
optimizer = torch.optim.Adam(Prior.rnn.parameters(), lr = 0.001)
for epoch in range(1, 101):
for step, batch in tqdm(enumerate(data), total=len(data)):
# Sample from DataLoader
seqs = batch.long()
# Calculate loss
log_p, _ = Prior.likelihood(seqs)
loss = - log_p.mean()
# Calculate gradients and take a step
optimizer.zero_grad()
loss.backward()
optimizer.step()
# Every 2 epoch we decrease learning rate and print some information
if epoch % 2 == 0 and step == 1:
#decrease_learning_rate(optimizer, decrease_by=0.03)
decrease_learning_rate(optimizer, decrease_by=0.03)
if epoch % 10 == 0 and step == 1:
tqdm.write("*" * 50)
tqdm.write("Epoch {:3d} step {:3d} loss: {:5.2f}\n".format(epoch, step, loss.data[0]))
seqs, likelihood, _ = Prior.sample(100)
valid = 0
f = open('tran_output.smi', 'a')
for i, seq in enumerate(seqs.cpu().numpy()):
smile = voc.decode(seq)
if Chem.MolFromSmiles(smile):
valid += 1
f.write(smile + "\n")
if i < 10:
tqdm.write(smile)
f.close()
tqdm.write("\n{:>4.1f}% valid SMILES".format(100 * valid / len(seqs)))
tqdm.write("*" * 50 + "\n")
# Save the Prior
torch.save(Prior.rnn.state_dict(), "data/100_epochs_transfer.ckpt")
if __name__ == "__main__":
smiles_file = sys.argv[1]
voc_file = './Voc'
print("Reading smiles...")
smiles_list = ds.canonicalize_smiles_from_file(smiles_file)
ds.filter_file_on_chars(smiles_list,voc_file)
Transfer('data/biogenic.ckpt')