636 9 months ago

GPT-2 is a transformers model pretrained on a very large corpus of English data in a self-supervised fashion.

124m 355m 774m 1.5b

16 models