Fix documentation for converting SFT/DPO weights back to HF Llama #1318
+1
−1
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Quick documentation fix.
In the documentation post-training/README.md, the base weights were converted from HF Llama:
gpt-neox/post-training/README.md
Line 6 in 59a5236
When converting fine-tuned weights from GPTNeoX format back to HF Llama, the
--architecture llama
flag seems to be required (otherwise the default value"neox"
would be selected.) While specified for RM conversion, this flag is not included in the command for converting GPTNeoX SFT/DPO weights back to HF in this documentation.https://github.com/EleutherAI/gpt-neox/blob/59a5236ddaf721890e3d6ef98fb8ca66c2266ce0/post-training/README.md?plain=1#L53-56
gpt-neox/tools/ckpts/convert_neox_to_hf.py
Lines 469 to 478 in 59a5236
Fixes #1317