You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am following instructions here for EVA-02 pretraining. It is recommended to download this EVA-CLIP model, which seems to be the first EVA-CLIP version, based on its ~2.2G size (i.e. EVA-01-CLIP, also available from here).
Would pre-training seamlessly work when directly using an EVA-02-CLIP model (which also seems the one used in the EVA-02 paper), or even one of the larger EVA-CLIP-8B or EVA-CLIP-18B?
In that case, how the --teacher_type and --clip_model parameters should be set in the pre-training script for those models?
Thanks you very much!
The text was updated successfully, but these errors were encountered:
Hello, thanks for releasing this great work!
I am following instructions here for EVA-02 pretraining. It is recommended to download this EVA-CLIP model, which seems to be the first EVA-CLIP version, based on its ~2.2G size (i.e. EVA-01-CLIP, also available from here).
Would pre-training seamlessly work when directly using an EVA-02-CLIP model (which also seems the one used in the EVA-02 paper), or even one of the larger EVA-CLIP-8B or EVA-CLIP-18B?
In that case, how the
--teacher_type
and--clip_model
parameters should be set in the pre-training script for those models?Thanks you very much!
The text was updated successfully, but these errors were encountered: