-
Notifications
You must be signed in to change notification settings - Fork 420
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[PyTorch][references] fix pretrained with different vocabs #874
Conversation
Codecov Report
@@ Coverage Diff @@
## main #874 +/- ##
==========================================
- Coverage 94.84% 94.84% -0.01%
==========================================
Files 133 133
Lines 5200 5215 +15
==========================================
+ Hits 4932 4946 +14
- Misses 268 269 +1
Flags with carried forward coverage won't be shown. Click here to find out more.
Continue to review full report at Codecov.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the fix!
@charlesmindee TF needs the same but will do in another PR :) i think in TF its enough to modify the checkpoint loading |
* backup * fix-first * update * flake * update * revert ugly new head idea * add reco test
* backup * fix-first * update * flake * update * revert ugly new head idea * add reco test
This PR fixes the mismatching linear layers at each model if we pass a different vocab as french and the
pretrained
flag to the training scriptsAny feedback is welcome 🤗