A GPT in 60 Lines of NumPy
In this post, we'll implement a GPT from scratch in just 60 lines of numpy. We'll then load the trained GPT-2 model weights released by OpenAI into our implementation and generate some text. This post assumes familiarity with Python, NumPy, and some basic experience training neural networks. Code for this blog post can be found…