Skip to content

mounicam/neural_splitter

Repository files navigation

Introduction

This directory consists of the code for BERT-initialized Transformer and also instructions to run our Transformer-based sentence splitter.

Requirements and Installation

The code is based on fairseq toolkit. It requires PyTorch version >= 1.0.0 and Python version >= 3.5. For training new models, you'll also need an NVIDIA GPU and NCCL

If you do not have PyTorch, please follow the instructions here to install: https://github.com/pytorch/pytorch#installation.

And then install fairseq from the source code shared in the repository.

cd neural_splitter
pip install --editable .

Please note that the shared version of fairseq is different from the latest version and BERT-initalized Transformer is implemented specifically for this version.

Pre-trained models

  1. You need to preprocess the data using the preprocess.sh script. First, the script performs BERT Tokenization and then creates a binarized fairseq dataset.
sh preprocess.sh <raw data directory> <tokenized data directory>  <binarized data directory>

# The script assumes that the raw data has the following format
# <directory>/
#   |-- train.src
#   |-- train.dst
#   |-- test.src
#   |-- test.dst
#   |-- valid.src
#   |-- valid.dst
# .src files contain complex sentences and .dst files contain simple sentences.
  1. Download the Transformer checkpoint. You can perform generation using the following command.
sh generate.sh <binarized data directory> <checkpoint> <output file name> <GPU device id> <split>

<checkpoint> refers to the path of the checkpoint
<split> takes one of the following values: train, valid, test

Training

  1. Follow the step 1 described in the above section for Transformer.

  2. Download the BERT-base checkpoint from here and unzip the folder. Then, train using the following command:

CUDA_VISIBLE_DEVICES=<GPU1,GPU2...> python3  train.py <binarized data path from previous step> --save-dir <checkpoint directory>  \
    --lr 0.0001 --optimizer adam  -a bert_rand --max-update 100000 --user-dir my_model --batch-size 32  \
    --lr-scheduler inverse_sqrt --warmup-updates 40000 --max-source-positions 512 --max-target-positions 512 \
    --bert-path cased_L-12_H-768_A-12/bert_model.ckpt

All the model parameters are specified in my_model/__init__.py file.

  1. Follow the step 2 described in the previous section. You can generate using the best checkpoint according to the cross-entropy loss, i.e. checkpoint_best.pt in the specified checkpoint directory. Alternatively, you can also choose the best checkpoint according to the SARI score on the validation dataset.

Citation

If you use any of these resources, please cite fairseq and our paper:

@inproceedings{jiang2020neural,
  title={Neural CRF Model for Sentence Alignment in Text Simplification},
  author={Jiang, Chao and Maddela, Mounica and Lan, Wuwei and Zhong, Yang and Xu, Wei},
  booktitle={Proceedings of the Association for Computational Linguistics (ACL)},
  year={2020}
}
@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}

About

No description, website, or topics provided.

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages