What makes GPT?

The key GPT (Generative Pretrained Transformer) concepts start with Transformer Architecture, Large Scale Language Models, and Fine-tuning. Transformer Architecture The Transformer architecture is a type of model architecture used in natural language processing tasks. It was introduced in the paper “Attention is All You Need” by Vaswani et al (https://arxiv.org/abs/1706.03762). The key innovation of theContinue reading “What makes GPT?”