Hacker News
- MinGPT: Minimal PyTorch re-implementation of GPT https://github.com/karpathy/minGPT 24 comments
- Karpathy's MinGPT https://github.com/karpathy/minGPT 102 comments
Linking pages
- GitHub - karpathy/nanoGPT: The simplest, fastest repository for training/finetuning medium-sized GPTs. https://github.com/karpathy/nanoGPT 366 comments
- Will scaling work? - by Dwarkesh Patel - Dwarkesh Podcast https://www.dwarkeshpatel.com/p/will-scaling-work 286 comments
- GitHub - kakaobrain/minDALL-E: PyTorch implementation of a 1.3B text-to-image generation model trained on 14 million image-text pairs https://github.com/kakaobrain/minDALL-E 22 comments
- GitHub - joennlae/tensorli: Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines). https://github.com/joennlae/tensorli 19 comments
- GPT-2 Neural Network Poetry · Gwern.net https://www.gwern.net/GPT-2 13 comments
- GitHub - certik/fastGPT: Fast GPT-2 inference written in Fortran https://github.com/certik/fastGPT 13 comments
- GitHub - google/maxtext: A simple, performant and scalable Jax LLM! https://github.com/google/maxtext 9 comments
- GitHub - taishi-i/awesome-ChatGPT-repositories: A curated list of resources dedicated to open source GitHub repositories related to ChatGPT https://github.com/taishi-i/awesome-ChatGPT-repositories 5 comments
- MAD, China, and the Semiconductor Showdown (Part 2) https://eastwind.substack.com/p/mad-china-and-the-semiconductor-showdown-2ba 2 comments
- GitHub - trekhleb/homemade-gpt-js: A minimal TensorFlow.js re-implementation of Karpathy's minGPT (Generative Pre-trained Transformer) https://github.com/trekhleb/homemade-gpt-js 2 comments
- minGPT in Julia using Flux! | Can Candan https://cancandan.github.io/julia/flux/machine-learning/2022/03/30/mingpt-julia.html 1 comment
- GitHub - ritchieng/the-incredible-pytorch: The Incredible PyTorch: a curated list of tutorials, papers, projects, communities and more relating to PyTorch. https://github.com/ritchieng/the-incredible-pytorch 0 comments
- Ahead of AI #2: Transformers, Fast and Slow | Revue https://newsletter.sebastianraschka.com/issues/ahead-of-ai-2-transformers-fast-and-slow-1402662 0 comments
- Accessible Multi-Billion Parameter Model Training with PyTorch Lightning + DeepSpeed | by PyTorch Lightning team | PyTorch Lightning Developer Blog https://medium.com/pytorch-lightning/accessible-multi-billion-parameter-model-training-with-pytorch-lightning-deepspeed-c9333ac3bb59 0 comments
- GitHub - hbchen-one/Transformer-Models-from-Scratch: implementing various transformer models for various tasks https://github.com/hbchen-one/Transformer-Models-from-Scratch 0 comments
- GitHub - febinsathar/goodreads: goodreads https://github.com/febinsathar/goodreads 0 comments
- Minix and nanoGPT https://matt-rickard.com/nanogpt 0 comments
- GitHub - jaymody/picoGPT: An unnecessarily tiny implementation of GPT-2 in NumPy. https://github.com/jaymody/picoGPT 0 comments
- GitHub - jiep/offensive-ai-compilation: A curated list of useful resources that cover Offensive AI. https://github.com/jiep/offensive-ai-compilation 0 comments
- GitHub - ZSvedic/educational-transformer: Easy-to-follow, educational implementation of the transformer model in PyTorch. https://github.com/ZSvedic/educational-transformer 0 comments
Linked pages
- GitHub - karpathy/nanoGPT: The simplest, fastest repository for training/finetuning medium-sized GPTs. https://github.com/karpathy/nanoGPT 366 comments
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- GitHub - openai/gpt-2: Code for the paper "Language Models are Unsupervised Multitask Learners" https://github.com/openai/gpt-2 2 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - karpathy/minGPT: A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
See how to search.