You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Better serialization for all models and tokenizers (BERT, GPT, GPT-2 and Transformer-XL) with best practices for saving/loading in readme and examples.
Relaxing network connection requirements (fallback on the last downloaded model in the cache when we can't reach AWS to check eTag)
Breaking changes:
warmup_linear method in OpenAIAdam and BertAdam is now replaced by flexible schedule classes for linear, cosine and multi-cycles schedules.
Bug fixes and improvements to the library modules:
add a flag in BertTokenizer to skip basic tokenization (@john-hewitt)
Allow tokenization of sequences > 512 (@CatalinVoss)
clean up and extend learning rate schedules in BertAdam and OpenAIAdam (@lukovnikov)