V 0.4 beta 2
robvanderg
released this
22 Sep 12:22
·
26 commits
to 0.4-beta
since this release
- Much lower memory by having a maximum number of tokens per batch
- support lower torch versions (and updated requirements.txt accordingly)
- fixed output predictions
- added multiseq and multiclas task type (and multi-accuracy)
- log the losses of each task
- Support also language models that have no start/end token. Tested with: [facebook/nllb-200-distilled-600M', 'google/mt5-base', 't5-base', 'google/byt5-base', 'Helsinki-NLP/opus-mt-mul-en', 'google/canine-s', 'google/canine-c', 'facebook/xglm-564M', 'facebook/xglm-564M', 'facebook/mgenre-wiki', 'setu4993/LaBSE', 'bigscience/bloom-560m', 'facebook/mbart-large-50', "microsoft/mdeberta-v3-base", "studio-ousia/mluke-large", "google/rembert", "cardiffnlp/twitter-xlm-roberta-base", "xlm-roberta-large", "bert-base-multilingual-cased", "xlm-roberta-base", 'distilbert-base-multilingual-cased', 'microsoft/infoxlm-large', 'bert-base-multilingual-uncased', 'Peltarion/xlm-roberta-longformer-base-4096', 'Peltarion/xlm-roberta-longformer-base-4096', 'studio-ousia/mluke-base', 'xlm-mlm-100-1280