How did we arrive at GPT-3?

GPT-3 is what is popularly known today, But its interesting to learn that there was a long history to reach this stage:

The introduction of Back Propagation model in 1960s,

Introduction of Recurrent Neural Network in 1986 allowing feedback

Introduction of Long Short Term Memory in 1998 to resolve the vanishing gradient problem,

Introduction of GPU based processors by Nvidia in 2000’s,

Introduction of GPU to Deep Learning in 2009 by Andrew NG,

Introduction of Attention based LSTM RNN in 2014 to remember large sequences   

Introduction of Transformer model solely based on Attention in 2017, and then later

the rise of transformer models like BERT, Megatron and GPTs from 2018…

all of these advancements have propelled AI to the next stage to where we are today!!!