-
Notifications
You must be signed in to change notification settings - Fork 6.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RoBERTa model #908
Comments
Yes, it’s the last layer before lm_head. I will add an option to expose specific layers, but for now you can copy this: https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta.py#L106 If you add return_all_hiddens=True, then the second element of the returned tuple will contain all of the inner states. |
Exposed new functionality in #909:
|
Hi @myleott thanks for your kind help and extending the interface ❤️ I could now integrate the RoBERTa into an upcoming version of Flair 🤗 |
Summary: (1) Enable to print the iterative refinement history for all NAT models by setting --retain-iter-history during decoding; (2) Fix a small bug in the decoding process in Levenshtein Transformer. Pull Request resolved: fairinternal/fairseq-py#908 Differential Revision: D18493234 Pulled By: MultiPath fbshipit-source-id: 9e7702adcea49f39d3c10b5349b5a9ae66399a24
…search#908) Summary: (1) Enable to print the iterative refinement history for all NAT models by setting --retain-iter-history during decoding; (2) Fix a small bug in the decoding process in Levenshtein Transformer. Pull Request resolved: fairinternal/fairseq-py#908 Differential Revision: D18493234 Pulled By: MultiPath fbshipit-source-id: 9e7702adcea49f39d3c10b5349b5a9ae66399a24
…search#908) Summary: (1) Enable to print the iterative refinement history for all NAT models by setting --retain-iter-history during decoding; (2) Fix a small bug in the decoding process in Levenshtein Transformer. Pull Request resolved: fairinternal/fairseq-py#908 Differential Revision: D18493234 Pulled By: MultiPath fbshipit-source-id: 9e7702adcea49f39d3c10b5349b5a9ae66399a24
Hi,
thanks for releasing the RoBERTa model ❤️
I've one question regarding to the output features:
Are these features the output of the last layer (layer nr. 24) of the Transformer model? Is it currently possible to select a specific layer?
Thanks in advance,
Stefan
The text was updated successfully, but these errors were encountered: