diff --git a/mcli/mcli-1b-eval.yaml b/mcli/mcli-1b-eval.yaml index b4d49ab2e9..b9fafe385b 100644 --- a/mcli/mcli-1b-eval.yaml +++ b/mcli/mcli-1b-eval.yaml @@ -1,8 +1,8 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - # git_branch: # Specify your git branch - git_commit: 186dd19888a8c8874584f9e78619f3fb0348309f # TODO: repin after next release + git_branch: v0.3.0 + # git_commit: # OR use your commit hash pip_install: -e .[gpu] ssh_clone: false # Should be true if using a private repo @@ -33,7 +33,7 @@ parameters: model_max_length: ${max_seq_len} model: name: mpt_causal_lm - init_device: meta + init_device: mixed d_model: 2048 n_heads: 16 # Modified 24->16 so that d_head == 128 to satisfy FlashAttention n_layers: 24 diff --git a/mcli/mcli-1b-max-seq-len-8k.yaml b/mcli/mcli-1b-max-seq-len-8k.yaml index f2108f964f..24af39234c 100644 --- a/mcli/mcli-1b-max-seq-len-8k.yaml +++ b/mcli/mcli-1b-max-seq-len-8k.yaml @@ -1,7 +1,7 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 # git_commit: # OR use your commit hash pip_install: -e .[gpu] ssh_clone: false # Should be true if using a private repo diff --git a/mcli/mcli-1b.yaml b/mcli/mcli-1b.yaml index bf2b32ee7e..e9cbbb531b 100644 --- a/mcli/mcli-1b.yaml +++ b/mcli/mcli-1b.yaml @@ -1,7 +1,7 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 # git_commit: # OR use your commit hash pip_install: -e .[gpu] ssh_clone: false # Should be true if using a private repo diff --git a/mcli/mcli-benchmark-mpt.yaml b/mcli/mcli-benchmark-mpt.yaml index 508d7ebcac..1d3b972027 100644 --- a/mcli/mcli-benchmark-mpt.yaml +++ b/mcli/mcli-benchmark-mpt.yaml @@ -11,7 +11,8 @@ image: mosaicml/pytorch:1.13.1_cu117-python3.10-ubuntu20.04 integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 + # git_commit: # OR use your commit hash pip_install: '.[gpu]' command: | diff --git a/mcli/mcli-convert-composer-to-hf.yaml b/mcli/mcli-convert-composer-to-hf.yaml index c401a07fbb..0ddc90f572 100644 --- a/mcli/mcli-convert-composer-to-hf.yaml +++ b/mcli/mcli-convert-composer-to-hf.yaml @@ -1,16 +1,16 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 # git_commit: # OR use your commit hash pip_install: -e . ssh_clone: false # Should be true if using a private repo command: | - cd llm-foundry/llmfoundry/inference + cd llm-foundry/scripts/inference python convert_composer_to_hf.py \ - --composer_path s3://bucket/folder/checkpoint-path.pt \ - --hf_output_path s3://bucket/folder/hf/ \ + --composer_path s3://mosaicml-internal-checkpoints-shared/daniel/i5-adamw-5e-7-09-099-00-005dur-00-cos-1e-6-64/checkpoints/ep3-ba2781-rank0.pt \ + --hf_output_path s3://mosaicml-internal-checkpoints-shared/daniel/hf-convert-030-test \ --output_precision bf16 \ image: mosaicml/pytorch:1.13.1_cu117-python3.10-ubuntu20.04 diff --git a/mcli/mcli-hf-eval.yaml b/mcli/mcli-hf-eval.yaml index 5cf4cf98da..accff7d5c0 100644 --- a/mcli/mcli-hf-eval.yaml +++ b/mcli/mcli-hf-eval.yaml @@ -1,7 +1,7 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 # git_commit: # OR use your commit hash pip_install: -e ".[gpu]" ssh_clone: false # Should be true if using a private repo @@ -11,7 +11,7 @@ command: | composer eval/eval.py /mnt/config/parameters.yaml # Mosaic Cloud will use run_name (with a unique suffix) to populate the env var $RUN_NAME -run_name: all-eval +run_name: mpt-eval gpu_num: 8 # gpu_type: # cluster: # replace with your cluster here! diff --git a/mcli/mcli-hf-generate.yaml b/mcli/mcli-hf-generate.yaml index 8861be54d8..eeb71c1ce6 100644 --- a/mcli/mcli-hf-generate.yaml +++ b/mcli/mcli-hf-generate.yaml @@ -1,7 +1,7 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_branch: v0.2.0 + git_branch: v0.3.0 # git_commit: # OR use your commit hash pip_install: -e .[gpu] ssh_clone: false # Should be true if using a private repo diff --git a/mcli/mcli-llama2-finetune.yaml b/mcli/mcli-llama2-finetune.yaml index cdba4d41cf..89c9c0cd9c 100644 --- a/mcli/mcli-llama2-finetune.yaml +++ b/mcli/mcli-llama2-finetune.yaml @@ -1,7 +1,7 @@ integrations: - integration_type: git_repo git_repo: mosaicml/llm-foundry - git_commit: 5ec4016b40652557d57a1d4949ad13a65251184b # TODO: repin this after next release + git_branch: v0.3.0 pip_install: -e .[gpu] ssh_clone: false # Should be true if using a private repo