-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changed is_pipe_parallel setting to fix pipeline-parallel inference #866
Conversation
/gpt-neox/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1671, in load_state_dict |
I have read the CLA Document and I hereby sign the CLA |
@crazyofapple -- You're seeing an error because you're trying to load a sequential checkpoint that you saved before the PR (with self.pipe_parallel_size >= 2, leading to a sequential model/ckpt), then tried to load it with the PR (with self.pipe_parallel_size >= 1), which tries to convert the checkpoint to a GPT2ModelPipe and fails. If you need to load those model weights intact, you'll have to leave this commit out. Otherwise, delete that old checkpoint and update to this commit. |
* add flash_attn_kvpacked * fix formatting * accept changes from main & resolve conflicts * Error Signed-off-by: Dashiell Stander <dstander@protonmail.com> * errors Signed-off-by: Dashiell Stander <dstander@protonmail.com> * feat(ci): add pip caching to CI * Set training attribute appropriately Signed-off-by: Dashiell Stander <dstander@protonmail.com> * Split up FlashAttention methods Signed-off-by: Dashiell Stander <dstander@protonmail.com> * Comment out clear_cache Signed-off-by: Dashiell Stander <dstander@protonmail.com> * Just remove clear_cache Signed-off-by: Dashiell Stander <dstander@protonmail.com> * Fix pre-commit formatting Signed-off-by: Dashiell Stander <dstander@protonmail.com> * Changed is_pipe_parallel setting to fix pipeline-parallel inference (#866) * Changed is_pipe_parallel setting to fix pipeline-parallel inference * Update NeoXArgs docs automatically * Update NeoXArgs docs automatically --------- Co-authored-by: github-actions <github-actions@github.com> Co-authored-by: Quentin Anthony <qganthony@yahoo.com> * feat: improve typing * Added DeeperSpeed to requirements.txt * Update NeoXArgs docs automatically * Update NeoXArgs docs automatically * Update train.py update train.py 1. black formatter. 2. remove unnecessary import 3. add more arguments * Update utils.py Black formatting Add logic required to expand "~" * Update train.py removed num_proc temporarily disabled emoji added continuing subword prefix option ( does not work well with Bytelevel) * Update utils.py improve reader error handling * Update train.py add whitespace related handling. add whitespace argument expose reconstruct pre_tokenizer_list add more whitespace to check tokenizer invertibility * Update train.py * Update utils.py remove unnecessary print * Update train.py set dropout default to None import path related code. Change normalizer change buffer_tokens change whitespace reservation handling * Update train.py Clear whitespace_reservation TODO add single_whitespace argument (might be necessary for invertibility) * Create .gitignore add gitignore file to ignore artifacts * Update train.py add directory parsing error checks add more metrics (tokenizer reconstructions, unicode fallback portion) * Update preprocess.py path handling changes black formatting * Update train.py change from GPT2TokenizerFast to PreTrainedTokenizerFast class * Update train.py enhanced test string * Update utils.py add logic to handle jsonl, txt input add logic to handle folder with jsonl,txt or arrow dataset * Update train.py add byte_fallback option expose (incompatible with current transformer wrapper) change dataset_loading with new util.py add dataset shuffling option * Update utils.py fix error in loading sequence * Update train.py fix whitespace preservation logic * Update train.py simplify data loading logic. remove unnecessary special tokens * Update train.py remove emoji related code * Update train.py add whitespace processing regex r"\s{16,}" * update tokenizer add whitespace pretokenizer (only processes looong whitespaces) * Update train.py * Update train.py add camel case regex * Update train.py separate camel_case regex * Update train.py * Update train.py --------- Signed-off-by: Dashiell Stander <dstander@protonmail.com> Co-authored-by: Satpal Singh Rathore <satpal.code@gmail.com> Co-authored-by: Dashiell Stander <dstander@protonmail.com> Co-authored-by: Saurav Maheshkar <sauravvmaheshkar@gmail.com> Co-authored-by: Stella Biderman <stellabiderman@gmail.com> Co-authored-by: Curt Tigges <ct@curttigges.com> Co-authored-by: github-actions <github-actions@github.com> Co-authored-by: Quentin Anthony <qganthony@yahoo.com>
Fix for #854