Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I’ve completely avoided GPT and LLMs. This looks like it would generate some level of fluidity in text output, but not be able to parse and answer a question.

Is there any simplistic blog posts / training courses which go through how they work, or expose a toy engine in python or similar that? All the training I’ve seen so far seems oriented at how to use the platforms rather than how they actually work.



Jay Alammar has my favorite sequence of tutorials from basic neural network math to GPT2.

Particularly [0], [1], and [2]

[0] http://jalammar.github.io/illustrated-transformer/

[1] http://jalammar.github.io/illustrated-gpt2/

[2] https://jalammar.github.io/visualizing-neural-machine-transl...


Strap in, this is by far the best resource: https://www.youtube.com/watch?v=kCc8FmEb1nY




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: