Skip to content

Commit

Permalink
fix tests
Browse files Browse the repository at this point in the history
  • Loading branch information
albertbou92 committed May 17, 2024
1 parent 7a04a47 commit abc1b1b
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions scripts/pretrain/config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,18 +3,18 @@ experiment_name: acegen
agent_name: pretrain_gru
logger_backend: null # csv, wandb, tensorboard, or null
seed: 101
recompute_dataset: True # if False and dataset_log_dir contains a dataset, it will be used
dataset_log_dir: /tmp/pretrain
model_log_dir: /tmp/pretrain
log_frequency: 500

# Dataset configuration
train_dataset_path: null
tokenizer: SMILESTokenizer # SMILESTokenizer, SMILESTokenizer2, DeepSMILESTokenizer, SELFIESTokenizer, AISTokenizer, SAFETokenizer, SmiZipTokenizer
tokenizer: SMILESTokenizerChEMBL # SMILESTokenizerChEMBL, SMILESTokenizerEnamine, DeepSMILESTokenizer, SELFIESTokenizer, AISTokenizer, SAFETokenizer, SmiZipTokenizer
recompute_dataset: True # if False and dataset_log_dir contains a dataset, it will be used
dataset_log_dir: /tmp/pretrain # if recomputing dataset, save it here

# Model configuration
model: gru # gru, lstm, or gpt2
custom_model_factory: null # Path to a custom model factory (e.g. my_module.create_model)
model_log_dir: /tmp/pretrain # save model here

# Training configuration
lr: 0.0001
Expand Down

0 comments on commit abc1b1b

Please sign in to comment.