-
Notifications
You must be signed in to change notification settings - Fork 442
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Expose packed: False, set log_peak_memory_stats: True, set compile: False #1872
Changes from all commits
3978d15
00d09d0
bb7648f
22c84ab
8acfa75
9b47501
e60eebd
dacc3d2
a311232
b0b4b14
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -26,6 +26,7 @@ tokenizer: | |
|
||
# Dataset | ||
dataset: | ||
packed: False # Set to true for great speed ups | ||
_component_: torchtune.datasets.alpaca_dataset | ||
seed: null | ||
shuffle: True | ||
|
@@ -57,7 +58,7 @@ loss: | |
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss | ||
max_steps_per_epoch: null | ||
gradient_accumulation_steps: 1 | ||
|
||
compile: False | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Should we be explicit that this will There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. agreed. I think we can add a comment, similar to packed
wdyt? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Add some comment there? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Ok, will add |
||
|
||
# Training env | ||
device: cuda | ||
|
@@ -78,3 +79,4 @@ metric_logger: | |
log_dir: ${output_dir} | ||
output_dir: /tmp/alpaca-llama3-finetune | ||
log_every_n_steps: null | ||
log_peak_memory_stats: True |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Huge nit: can we move this below the
_component_
declaration?That way it reads more as an option for the specific builder.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, though like this first time.. But I followed original declaration from issue:
Will be fixed