Hacker News
- Transformers Are Graph Neural Networks https://thegradient.pub/transformers-are-graph-neural-networks/ 25 comments
Linking pages
- Why I'm lukewarm on graph neural networks - SingleLunch https://www.singlelunch.com/2020/12/28/why-im-lukewarm-on-graph-neural-networks/ 47 comments
- Beyond Message Passing: a Physics-Inspired Paradigm for Graph Neural Networks https://thegradient.pub/graph-neural-networks-beyond-message-passing-and-weisfeiler-lehman/ 16 comments
- Graph Neural Networks through the lens of Differential Geometry and Algebraic Topology | by Michael Bronstein | Towards Data Science https://towardsdatascience.com/graph-neural-networks-through-the-lens-of-differential-geometry-and-algebraic-topology-3a7c3c22d5f 6 comments
- Towards Deep Learning for Relational Databases | by Gustav Šír | Nov, 2022 | Towards Data Science https://towardsdatascience.com/towards-deep-learning-for-relational-databases-de9adce5bb00 4 comments
- Recent Advances in Deep Learning for Routing Problems | Chaitanya K. Joshi https://www.chaitjo.com/post/deep-learning-for-routing-problems/ 3 comments
- How I Learned to Stop Worrying and Love Graphs https://www.appliedexploration.com/p/graph-neural-networks-future-of-ai 1 comment
- Aman's AI Journal • Primers • Transformers https://aman.ai/primers/ai/transformers/ 0 comments
- How I Learned to Stop Worrying and Love Graphs https://www.appliedexploration.com/p/graph-neural-networks-future-of-ai?s=w 0 comments
- What is Geometric Deep Learning?. Deep Learning 🤖 on graphs and in 3D | by Flawnson Tong | Medium https://towardsdatascience.com/what-is-geometric-deep-learning-b2adb662d91d 0 comments
Linked pages
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- Understanding searches better than ever before https://www.blog.google/products/search/search-language-understanding-bert/ 109 comments
- Better Language Models and Their Implications https://openai.com/blog/better-language-models/ 99 comments
- Understanding LSTM Networks -- colah's blog https://colah.github.io/posts/2015-08-Understanding-LSTMs/ 64 comments
- The Illustrated Transformer – Jay Alammar – Visualizing machine learning one concept at a time. https://jalammar.github.io/illustrated-transformer/ 36 comments
- Training a single AI model can emit as much carbon as five cars in their lifetimes | MIT Technology Review https://www.technologyreview.com/s/613630/training-a-single-ai-model-can-emit-as-much-carbon-as-five-cars-in-their-lifetimes/ 30 comments
- Attention Is All You Need - YouTube https://www.youtube.com/watch?v=iDulhoQ2pro 29 comments
- Transformers are Graph Neural Networks | NTU Graph Deep Learning Lab https://graphdeeplearning.github.io/post/transformers-are-gnns/ 19 comments
- A new model and dataset for long-range memory https://deepmind.com/blog/article/A_new_model_and_dataset_for_long-range_memory 13 comments
- Attention and Augmented Recurrent Neural Networks https://distill.pub/2016/augmented-rnns/ 9 comments
- Generative Modeling with Sparse Transformers https://openai.com/blog/sparse-transformer/ 9 comments
- Deep Learning, NLP, and Representations - colah's blog http://colah.github.io/posts/2014-07-NLP-RNNs-Representations/ 9 comments
- Transformer-XL: Unleashing the Potential of Attention Models – Google AI Blog https://ai.googleblog.com/2019/01/transformer-xl-unleashing-potential-of.html 7 comments
- Language, trees, and geometry in neural networks https://pair-code.github.io/interpretability/bert-tree/ 5 comments
- The Annotated Transformer https://nlp.seas.harvard.edu/2018/04/03/attention.html 3 comments
- [1806.01261] Relational inductive biases, deep learning, and graph networks https://arxiv.org/abs/1806.01261 2 comments
- Introducing a Conditional Transformer Language Model for Controllable Generation https://blog.einstein.ai/introducing-a-conditional-transformer-language-model-for-controllable-generation/ 1 comment
- Making Transformer networks simpler and more efficient https://ai.facebook.com/blog/making-transformer-networks-simpler-and-more-efficient/ 1 comment
- MegatronLM: Training Billion+ Parameter Language Models Using GPU Model Parallelism - NVIDIA ADLR https://nv-adlr.github.io/MegatronLM 1 comment
- [1910.10683] Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer https://arxiv.org/abs/1910.10683 1 comment
Related searches:
Search whole site: site:thegradient.pub
Search title: Transformers are Graph Neural Networks
See how to search.