Skip to content

Commit

Permalink
update (mosaicml#1097)
Browse files Browse the repository at this point in the history
  • Loading branch information
dakinggg authored Apr 5, 2024
1 parent 94301cd commit 60a1ab4
Show file tree
Hide file tree
Showing 10 changed files with 30 additions and 30 deletions.
6 changes: 3 additions & 3 deletions mcli/mcli-1b-eval.yaml
Original file line number Diff line number Diff line change
@@ -1,15 +1,15 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu]
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

command: |
cd llm-foundry/scripts/
composer eval/eval.py /mnt/config/parameters.yaml
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: mpt-1b-eval

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-1b-max-seq-len-8k.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu-flash2]
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

# We are fetching, converting, and training on the 'val' split
Expand All @@ -17,7 +17,7 @@ command: |
--out_root ./my-copy-c4 --splits train_small val_small \
--concat_tokens 8192 --tokenizer EleutherAI/gpt-neox-20b --eos_text '<|endoftext|>'
composer train/train.py /mnt/config/parameters.yaml
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: mpt-1b-ctx-8k-gpus-8

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-1b.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu-flash2]
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

# We are fetching, converting, and training on the 'val' split
Expand All @@ -21,7 +21,7 @@ command: |
eval_loader.dataset.split=val_small \
max_duration=100ba \
eval_interval=0
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: mpt-1b-gpus-8

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-benchmark-mpt.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,14 +6,14 @@ compute:
# cluster: TODO # Name of the cluster to use for this run
# gpu_type: a100_80gb # Type of GPU to use. We use a100_80gb in our experiments

image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest

integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: ".[gpu-flash2]"
pip_install: .[gpu]

command: |
cd llm-foundry/scripts/inference/benchmarking
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-convert-composer-to-hf.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .
pip_install: .
ssh_clone: false # Should be true if using a private repo

command: |
Expand All @@ -13,7 +13,7 @@ command: |
--hf_output_path s3://bucket/folder/hf/ \
--output_precision bf16 \
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: convert-composer-hf

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-hf-eval.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e ".[gpu-flash2]"
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

command: |
Expand All @@ -16,7 +16,7 @@ gpu_num: 8
# gpu_type:
# cluster: # replace with your cluster here!

image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest

# The below is injected as a YAML file: /mnt/config/parameters.yaml
parameters:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-hf-generate.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu-flash2]
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

command: |
Expand Down Expand Up @@ -35,7 +35,7 @@ command: |
"Here's a quick recipe for baking chocolate chip cookies: Start by" \
"The best 5 cities to visit in Europe are"
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: hf-generate

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-llama2-finetune.yaml
Original file line number Diff line number Diff line change
@@ -1,15 +1,15 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu-flash2]
pip_install: .[gpu]
ssh_clone: false # Should be true if using a private repo

command: |
cd llm-foundry/scripts
composer train/train.py /mnt/config/parameters.yaml
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
name: llama2-finetune

compute:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-openai-eval.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
integrations:
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: -e ".[gpu-flash2,openai]"
pip_install: .[gpu,openai]
ssh_clone: false # Should be true if using a private repo

command: |
Expand All @@ -16,7 +16,7 @@ gpu_num: #
gpu_type: #
cluster: # replace with your cluster here!

image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest

# The below is injected as a YAML file: /mnt/config/parameters.yaml
parameters:
Expand Down
6 changes: 3 additions & 3 deletions mcli/mcli-pretokenize-oci-upload.yaml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
name: c4-2k-pre-tokenized
image: mosaicml/llm-foundry:2.1.0_cu121_flash2-latest
image: mosaicml/llm-foundry:2.2.1_cu121_flash2-latest
compute:
gpus: 8 # Number of GPUs to use

Expand All @@ -14,9 +14,9 @@ integrations:
- oci-cli==3.23.2
- integration_type: git_repo
git_repo: mosaicml/llm-foundry
git_branch: v0.6.0
git_branch: v0.7.0
# git_commit: # OR use your commit hash
pip_install: "."
pip_install: .
ssh_clone: false # Should be true if using a private repo

command: |
Expand Down

0 comments on commit 60a1ab4

Please sign in to comment.