Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix T5 tokenizer loading #3544

Merged
merged 1 commit into from
Oct 11, 2024
Merged

fix T5 tokenizer loading #3544

merged 1 commit into from
Oct 11, 2024

Conversation

helpmefindaname
Copy link
Collaborator

Closes #3543

The T5 tokenizer fast is required to load from the slow tokenizer when add_prefix_space is True. See the code
I am not sure why exactly this is implemented that way, as the traces do not show a specific reasoning, but with a small exception on our side, we can easily support all T5 tokenizers.

@helpmefindaname helpmefindaname added bug Something isn't working labels Sep 13, 2024
@DhruvSondhi
Copy link

Hello,

Yes, this indeed fixes the problem. I was able to load the best model from the set of models trained. Hoping that this could be merged into the master branch as soon as possible. Thanks, @helpmefindaname!

@alanakbik alanakbik merged commit 2993108 into master Oct 11, 2024
1 check passed
@alanakbik alanakbik deleted the fix-t5-tokenizer branch October 11, 2024 11:07
@alanakbik
Copy link
Collaborator

@helpmefindaname thanks for fixing this!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[Bug]: Cannot load pre-trained models after fine-tuning (Transformers)
3 participants