Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- In certain cases, rather than fine-tuning the entire pre-trained model
- end-to-end, it can be beneficial to obtained pre-trained contextual
- embeddings, which are fixed contextual representations of each input
- token generated from the hidden layers of the pre-trained model. This
- should also mitigate most of the out-of-memory issues.
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement