Large Language Models, GPT-3: Language Models are Few-Shot Learners
Efficiently scaling GPT from large to titanic magnitudes within the meta-learning frameworkIntroductionGPT is a family of language models that has been recently gaining a lot of popularity. The attention of the Data Science community was rapidly captured by the release of GPT-3 in 2020. After the appearance of GPT-2, almost nobody could even assume that nearly in a year there would appear a titanic version of GPT containing 175B of parameters! This is by two orders of magnitude more, compared to its predecessor.The…