Linking pages
- GitHub - dive-into-machine-learning/dive-into-machine-learning: Free ways to dive into machine learning with Python and Jupyter Notebook. Notebooks, courses, and other links. (First posted in 2016.) https://github.com/hangtwenty/dive-into-machine-learning 39 comments
- 100+ Best GitHub Repositories For Machine Learning https://www.theinsaneapp.com/2021/09/best-github-repository-for-machine-learning.html 3 comments
- Image Segmentation: Your Ultimate Guide to Easy Deployment and Fast Inferencing - Neural Magic https://neuralmagic.com/blog/image-segmentationyour-ultimate-guide-to-easy-deployment-and-fast-inferencing/ 2 comments
Linked pages
- Denoising Diffusion Probabilistic Models (DDPM) https://nn.labml.ai/diffusion/ddpm/index.html 54 comments
- labml.ai Annotated PyTorch Paper Implementations https://nn.labml.ai/index.html 17 comments
- Regret Minimization in Games with Incomplete Information (CFR) https://nn.labml.ai/cfr/index.html 11 comments
- Generalized Advantage Estimation (GAE) https://nn.labml.ai/rl/ppo/gae.html 6 comments
- Normalization Layers https://nn.labml.ai/normalization/index.html 3 comments
- PonderNet: Learning to Ponder https://nn.labml.ai/adaptive_computation/ponder_net/index.html 1 comment
- Zero-DP Memory Optimization https://nn.labml.ai/scaling/zero3/index.html 0 comments
- Graph Attention Networks v2 (GATv2) https://nn.labml.ai/graphs/gatv2/index.html 0 comments
- Evidential Deep Learning to Quantify Classification Uncertainty https://nn.labml.ai/uncertainty/evidence/index.html 0 comments
- Gated Linear Units and Variants https://nn.labml.ai/transformers/glu_variants/simple.html 0 comments
- Compressive Transformer https://nn.labml.ai/transformers/compressive/index.html 0 comments
- Linear Transformers Are Secretly Fast Weight Memory Systems https://nn.labml.ai/transformers/fast_weights/index.html 0 comments
- Pay Attention to MLPs (gMLP) https://nn.labml.ai/transformers/gmlp/index.html 0 comments
- Patches Are All You Need? (ConvMixer) https://nn.labml.ai/conv_mixer/index.html 0 comments
- Primer: Searching for Efficient Transformers for Language Modeling https://nn.labml.ai/transformers/primer_ez/index.html 0 comments
- Distilling the Knowledge in a Neural Network https://nn.labml.ai/distillation/index.html 0 comments
- Graph Attention Networks (GAT) https://nn.labml.ai/graphs/gat/index.html 0 comments
- DeepNorm https://nn.labml.ai/normalization/deep_norm/index.html 0 comments
- FNet: Mixing Tokens with Fourier Transforms https://nn.labml.ai/transformers/fnet/index.html 0 comments
- An Attention Free Transformer https://nn.labml.ai/transformers/aft/index.html 0 comments