Advertisement
Guest User

Untitled

a guest
Jun 17th, 2019
63
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.34 KB | None | 0 0
  1. In certain cases, rather than fine-tuning the entire pre-trained model
  2. end-to-end, it can be beneficial to obtained pre-trained contextual
  3. embeddings, which are fixed contextual representations of each input
  4. token generated from the hidden layers of the pre-trained model. This
  5. should also mitigate most of the out-of-memory issues.
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement