You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I am using command accelerate launch --config_file accelerate_config.yaml train.py --anchor_model_dir google/gemma-7b --aug_model_dir google/gemma-7b --num_heads 2 --num_connections 2 --learning_rate 3e-4 --batch_size 2 --output_dir './tmp' to train the composed model I am getting below error. Could you please help me to resolve this issue as soon as possible?
Also even after trying with FSDP setup, it is not executing in multiple gpus.
The text was updated successfully, but these errors were encountered:
I am using the same yaml file you have given in the repo. Also I am using 4 gpus. Somehow I managed fix this issue by modifying the parameters. Although I would appreciate if authors could mention the parameters specifically responsible for gpus allocations etc in their repo. In that way, one can easily execute the code without much hassle.
Also, I am now facing problem in loading the saved model in './tmp' folder after finetuning. Can you please explicitly provide the code for loading the model from that saved folder '/tmp'?
The generic model loading code is not working and throwing some errors.
Hi,
When I am using command
accelerate launch --config_file accelerate_config.yaml train.py --anchor_model_dir google/gemma-7b --aug_model_dir google/gemma-7b --num_heads 2 --num_connections 2 --learning_rate 3e-4 --batch_size 2 --output_dir './tmp'
to train the composed model I am getting below error. Could you please help me to resolve this issue as soon as possible?Also even after trying with FSDP setup, it is not executing in multiple gpus.
The text was updated successfully, but these errors were encountered: