Hacker News
- The Illustrated GPT-2: Visualizing Transformer Language Models (2019) https://jalammar.github.io/illustrated-gpt2/ 5 comments
- Visualizing Transformer Language Models: Illustrated GPT-2 https://jalammar.github.io/illustrated-gpt2/ 3 comments
Linking pages
- How GPT3 Works - Visualizations and Animations – Jay Alammar – Visualizing machine learning one concept at a time. https://jalammar.github.io/how-gpt3-works-visualizations-animations/ 109 comments
- GitHub - amitness/learning: A log of things I'm learning https://github.com/amitness/learning 17 comments
- The Turing Bot | The Topics I Would Choose If I Ever Did A PhD in AI/ML https://turing-bot.com/posts/masters-degree-new-learn 12 comments
- GitHub - mlabonne/llm-course: Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks. https://github.com/mlabonne/llm-course 10 comments
- The Annotated GPT-2 | Committed towards better future https://amaarora.github.io/2020/02/18/annotatedGPT2.html 2 comments
- How well does ChatGPT speak Japanese? https://www.passaglia.jp/gpt-japanese/ 2 comments
- Can a Robot Make You Laugh? — Teaching an AI to Tell Jokes | by Lorenzo Ampil | Towards Data Science https://towardsdatascience.com/can-a-robot-make-you-laugh-teaching-an-ai-to-tell-jokes-815f1e1e689c?sk=342342be4cbf0064f8f0e7cb7ec0b6bc&source=friends_link 1 comment
- Getting Artificial Neural Networks Closer to Animal Brains https://maraoz.com/2020/07/12/brains-vs-anns/ 1 comment
- Building a Slogan Generator with GPT-2 | Jonathan Bgn https://jonathanbgn.com/gpt2/2020/01/20/slogan-generator.html 0 comments
- “Sports Twitter” Conversation Analysis with Autoregressive Classification and Zero-shot Transfer Testing https://credwood.substack.com/p/sportsbot 0 comments
- The Illustrated Wav2vec | Jonathan Bgn https://jonathanbgn.com/2021/06/29/illustrated-wav2vec.html 0 comments
- Aman's AI Journal • Primers • Transformers https://aman.ai/primers/ai/transformers/ 0 comments
- All AI learning is tacit learning - by Rohit https://www.strangeloopcanon.com/p/all-ai-learning-is-tacit-learning 0 comments
- GPT-3 Is Amazing—And Overhyped https://www.forbes.com/sites/robtoews/2020/07/19/gpt-3-is-amazingand-overhyped/#b977db71b1cb 0 comments
- Google’s REALM — A Knowledge-base Augmented Language Model | by swapp19902 | Level Up Coding https://levelup.gitconnected.com/googles-realm-a-knowledge-base-augmented-language-model-bc1a9c9b3d09?sk=4315038c74989641638f3c9ea0505537&source=friends_link 0 comments
- Doctor GPT-3 - by Leon Lin - Avoid Boring People https://avoidboringpeople.substack.com/p/doctor-gpt-3 0 comments
- GPT-3: Language Models are Few-Shot Learners | by Grigory Sapunov | Intento https://blog.inten.to/gpt-3-language-models-are-few-shot-learners-a13d1ae8b1f9 0 comments
- NLP Year in Review — 2019. NLP highlights for the year 2019. | by elvis | DAIR.AI | Medium https://medium.com/dair-ai/nlp-year-in-review-2019-fb8d523bcb19 0 comments
- Dissecting Batching Effects in GPT Inference https://le.qun.ch/en/blog/2023/05/13/transformer-batching/ 0 comments
- Transformer Memory Arithmetic: Understanding all the Bytes in nanoGPT https://erees.dev/transformer-memory/ 0 comments
Linked pages
- https://deepmind.com/blog/alphastar-mastering-real-time-strategy-game-starcraft-ii/ 508 comments
- The Unreasonable Effectiveness of Recurrent Neural Networks https://karpathy.github.io/2015/05/21/rnn-effectiveness/ 434 comments
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- Better Language Models and Their Implications https://openai.com/blog/better-language-models/ 99 comments
- Hugging Face – The AI community building the future. https://huggingface.co/ 57 comments
- The Illustrated Transformer – Jay Alammar – Visualizing machine learning one concept at a time. https://jalammar.github.io/illustrated-transformer/ 36 comments
- http://youtube.com/watch?v=ipzr9bhei_o 26 comments
- Creative Commons — Attribution-NonCommercial-ShareAlike 4.0 International — CC BY-NC-SA 4.0 https://creativecommons.org/licenses/by-nc-sa/4.0/ 4 comments
- GitHub - openai/gpt-2: Code for the paper "Language Models are Unsupervised Multitask Learners" https://github.com/openai/gpt-2 2 comments
- [1801.10198] Generating Wikipedia by Summarizing Long Sequences https://arxiv.org/abs/1801.10198 0 comments
- Music Transformer: Generating Music with Long-Term Structure https://magenta.tensorflow.org/music-transformer 0 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/pytorch-transformers 0 comments