Skip to content

V 0.4 beta 2

Compare
Choose a tag to compare
@robvanderg robvanderg released this 22 Sep 12:22
· 26 commits to 0.4-beta since this release
  • Much lower memory by having a maximum number of tokens per batch
  • support lower torch versions (and updated requirements.txt accordingly)
  • fixed output predictions
  • added multiseq and multiclas task type (and multi-accuracy)
  • log the losses of each task
  • Support also language models that have no start/end token. Tested with: [facebook/nllb-200-distilled-600M', 'google/mt5-base', 't5-base', 'google/byt5-base', 'Helsinki-NLP/opus-mt-mul-en', 'google/canine-s', 'google/canine-c', 'facebook/xglm-564M', 'facebook/xglm-564M', 'facebook/mgenre-wiki', 'setu4993/LaBSE', 'bigscience/bloom-560m', 'facebook/mbart-large-50', "microsoft/mdeberta-v3-base", "studio-ousia/mluke-large", "google/rembert", "cardiffnlp/twitter-xlm-roberta-base", "xlm-roberta-large", "bert-base-multilingual-cased", "xlm-roberta-base", 'distilbert-base-multilingual-cased', 'microsoft/infoxlm-large', 'bert-base-multilingual-uncased', 'Peltarion/xlm-roberta-longformer-base-4096', 'Peltarion/xlm-roberta-longformer-base-4096', 'studio-ousia/mluke-base', 'xlm-mlm-100-1280