A Data-to-Text Generation Library built on top of Huggingface Transformers and Datasets
pip install -r requirements.txt
- BART
- T5
- Roberta2Roberta shared
- Bert2Bert shared
- Bert2GPT
<TRIPLE>...</TRIPLE> <TRIPLE>...</TRIPLE> ..... <TRIPLE>...</TRIPLE>
- permutations of strategy 0 with permutations controlled by
max_permutations
- hierarchical linearization of triple
- hierarchical linearization of triple with permutations
bash train.sh
or
python pipeline/main.py \
--data_dir ../data/dart/data/v1.1.1/ \
--dataset DART \
--linearize True \
--linearize_strategy 0 \
--run_id test \
--model BART \
--batch_size 16 \
--use_mixed_precision True \
--num_epochs 1 \
--overwrite True