Skip to content

Commit

Permalink
Fixes
Browse files Browse the repository at this point in the history
Signed-off-by: Hemil Desai <hemild@nvidia.com>
  • Loading branch information
hemildesai committed Aug 22, 2024
1 parent c7464b4 commit 0e05a35
Showing 1 changed file with 5 additions and 2 deletions.
7 changes: 5 additions & 2 deletions examples/llm/run/llama_pretraining.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@
from typing import Any, Optional

import nemo_run as run

from nemo.collections import llm


Expand Down Expand Up @@ -30,7 +31,7 @@ def get_parser():
parser.add_argument(
"--local",
action="store_true",
help="Do a dryrun and exit",
help="Run locally using run.LocalExecutor",
default=False,
)
return parser
Expand Down Expand Up @@ -139,7 +140,8 @@ def main():
ckpt_dir=f"/{exp_name}/checkpoints",
)

# dataloader =
# TODO: Overwrite the dataloader in the recipe.
# dataloader = set_your_custom_dataloader
# pretrain.data = dataloader

pretrain.trainer.val_check_interval = 400
Expand All @@ -153,6 +155,7 @@ def main():
if args.local:
executor = local_executor_torchrun(nodes=pretrain.trainer.num_nodes, devices=pretrain.trainer.devices)
else:
# TODO: Set your custom parameters for the Slurm Executor.
executor = slurm_executor(
user="",
host="",
Expand Down

0 comments on commit 0e05a35

Please sign in to comment.