๐ญ๐ฒ๐ฟ๐ผ-๐บ๐ฎ๐๐ต ๐ถ๐ป๐๐ฟ๐ผ ๐๐ผ ๐๐ ๐ต๐ถ๐๐๐ผ๐ฟ๐: ๐ณ๐ฟ๐ผ๐บ ๐๐ต๐ฒ ๐ญ๐ต๐ฑ๐ฌ๐ ๐๐ผ ๐๐ผ๐ฑ๐ฎ๐'๐ ๐๐๐ ๐ ๐
I wanted to structure my thinking about LLMs by going through their history since the 50s. This history is captivating, with the opposition between Connexionists (Rosenblatt, LeCun) and Symbolists, the first victories of "deep" neural networks, the revolution of Attention...
So I might have gone a bit too far! ๐
๐ I've made a long post summarizing the main stages of building LLMs: neural networks, optimization, backpropagation, attention layers...
โ And I've made sure to keep it 100% horrible-latex-math-free: the technical stuff is conveyed in graphs only, so it should be accessible to really anyone, even your grandfather (I'm sending it to mine right now).
Read it here in english ๐ https://aymeric-roucher.github.io/brief-history-of-ai/
Pour le post en franรงais ๐ https://aymeric-roucher.github.io/breve-histoire-de-l-ia/
I wanted to structure my thinking about LLMs by going through their history since the 50s. This history is captivating, with the opposition between Connexionists (Rosenblatt, LeCun) and Symbolists, the first victories of "deep" neural networks, the revolution of Attention...
So I might have gone a bit too far! ๐
๐ I've made a long post summarizing the main stages of building LLMs: neural networks, optimization, backpropagation, attention layers...
โ And I've made sure to keep it 100% horrible-latex-math-free: the technical stuff is conveyed in graphs only, so it should be accessible to really anyone, even your grandfather (I'm sending it to mine right now).
Read it here in english ๐ https://aymeric-roucher.github.io/brief-history-of-ai/
Pour le post en franรงais ๐ https://aymeric-roucher.github.io/breve-histoire-de-l-ia/