Skip to content

Releases: ThilinaRajapakse/simpletransformers

Bug Fixes

18 Jul 19:41
Compare
Choose a tag to compare

Fixed

  • Bug fixed in Seq2Seq tasks.
  • Bug fixed in NERModel where the classification report was missing in checkpoints.
  • Bug fixed in ELECTRA.
  • Bug fixed in Seq2Seq generic encoder-decoder model.
  • Bug fixed in Seq2Seq tasks.
  • Bug fixed in regression prediction.
  • Bug fixed in loading multiclass classification models when num_labels aren't specified.

Custom Parameter Groups (Freezing layers)

05 Jul 22:15
Compare
Choose a tag to compare

Added

  • Added support for custom parameter groups. docs

Fixed

  • ELECTRA pretraining no longer replaces 10% of masked inputs with random tokens. @dev-sngwn

MobileBERT, Chinese text handling, and Transformers 3.0.1 compatibility

05 Jul 22:00
Compare
Choose a tag to compare

Added

  • Added better support for Chinese text in Language Modeling. @taranais
  • Added mobilebert for classification, NER, QA, and Seq2Seq. @flozi00

Fixed

  • Fixed incorrect logic when using early_stopping_metric_minimize. @djstrong
  • Fixed issue with cache directory being created even when no_cache is set. @henninglebbaeus

Changed

  • Running loss is now shown next to the tqdm bar (with the tqdm bar description)
  • Removed tokenizers and transformers version pins (added earlier to avoid compatibility issues)

interact_single() for Conv AI and bug fixes

25 Jun 18:32
Compare
Choose a tag to compare

Added

  • Added interact_single() method to ConvAIModel. This accepts a message and conversation history (and an optional personality). @Amit80007

Fixed

  • Fixed bug in multi modal classification @tekkon
  • Fixed bug in Multi-Modal classification when using evaluate_during_training.

Changed

Hyperparameter optimization and Dataclass args

23 Jun 14:57
Compare
Choose a tag to compare

Added

  • All Simple Transformers models can now be used with W&B sweeps. (docs)
  • eval_model() now logs to wandb in the following models (can be turned off with wandb_log=False)
    • ClassificationModel
    • NERModel
  • Model args can now be specified through the relevant dataclass. (E.g. ClassificationArgs, T5Args, etc.) (docs)
  • All model args are now saved and loaded automatically with Simple Transformers models.
  • Multiprocessed decoding support added to Seq2Seq models
  • Custom label support for Classification tasks (except multilabel). (docs)
  • NER labels can be given as NERArgs.labels_list (persists through args saving)

Changed

  • Changed NERModel.eval_model() to return usable model_outputs
  • Improved tqdm descriptions in progress bars
  • ConvAIModel arg no_sample renamed to do_sample for consistency

Resolving compatibility issues introduced in transformers v2.11.0

04 Jun 10:52
Compare
Choose a tag to compare

Fixed

  • Fixed compatibility issues with breaking changes in transformers==2.11.0. @fcggamou

Lazy Loading Classification Datasets

31 May 20:52
Compare
Choose a tag to compare

Added

  • Added Lazy Loading support for classification tasks (except multi-label). (Docs)

Longformer for Language Modeling

30 May 18:10
Compare
Choose a tag to compare

Added

  • Added Longformer model support for Language Modeling.

XLM-RoBERTa support for Question Answering

26 May 19:20
Compare
Choose a tag to compare

Added

  • Added XLM-RoBERTa support for question answering tasks.
  • Added save_optimizer_and_scheduler (default 1) to global_args which controls whether optimizer and scheduler is saved along with the model. Disabling significantly reduces the disk space used by saved models.

Fixed

  • Bug in XLM tokenizer when preprocessing QA datasets.
  • QuestionAnsweringModel.predict(n_best_size=n) now correctly returns n answers per question (along with n probabilities).

BREAKING CHANGE

  • QuestionAnsweringModel.predict() now returns two lists (a list of dicts with question ids mapped to answers and a list of dicts with question ids mapped to the answer probabilities). @alexucb

Fixed major issues with ELECTRA language model training from scratch

24 May 17:38
Compare
Choose a tag to compare

Fixed

  • Fixed issues with training ELECTRA language models from scratch. @aced125 @Laksh1997
  • Fixed bug in save_discriminator() method.

Changed

  • The parallel process count is now limited to 61 by default on Windows systems. @leungi