Skip to content

Commit

Permalink
ddp
Browse files Browse the repository at this point in the history
  • Loading branch information
LauraGPT committed Jan 25, 2024
1 parent bfd08c2 commit 531d2d7
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 2 deletions.
4 changes: 2 additions & 2 deletions scripts/conf/asr_vicuna_lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ train_config:
weight_decay: 0.0
gamma: 0.85
seed: 42
use_fp16: False
use_fp16: false
mixed_precision: true
val_batch_size: 1

Expand Down Expand Up @@ -91,7 +91,7 @@ dataset_config:

fsdp_config:
mixed_precision: true
use_fp16: II"${train_config.use_fp16}"
use_fp16: false
# sharding_strategy: "FULL_SHARD" #ShardingStrategy = ShardingStrategy.FULL_SHARD
sharding_strategy: "NO_SHARD" #ShardingStrategy.NO_SHARD #MZY: set NO_SHARD to use DDP mode in FSDP
checkpoint_type: "StateDictType.SHARDED_STATE_DICT" # alternatively can use SHARDED_STATE_DICT save one file per rank, and can resize the world-size.
Expand Down
1 change: 1 addition & 0 deletions src/llama_recipes/pipeline/finetune.py
Original file line number Diff line number Diff line change
Expand Up @@ -81,6 +81,7 @@ def main(kwargs: DictConfig):
kwargs.model_config, \
kwargs.log_config, \
kwargs.dataset_config
fsdp_config.use_fp16 = train_config.use_fp16
del kwargs.train_config
del kwargs.fsdp_config
del kwargs.model_config
Expand Down

0 comments on commit 531d2d7

Please sign in to comment.