Hacker News
- Transformers Are Graph Neural Networks https://graphdeeplearning.github.io/post/transformers-are-gnns/ 19 comments
Linking pages
Linked pages
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- Understanding searches better than ever before https://www.blog.google/products/search/search-language-understanding-bert/ 109 comments
- Better Language Models and Their Implications https://openai.com/blog/better-language-models/ 99 comments
- Understanding LSTM Networks -- colah's blog https://colah.github.io/posts/2015-08-Understanding-LSTMs/ 64 comments
- The Illustrated Transformer – Jay Alammar – Visualizing machine learning one concept at a time. https://jalammar.github.io/illustrated-transformer/ 36 comments
- Training a single AI model can emit as much carbon as five cars in their lifetimes | MIT Technology Review https://www.technologyreview.com/s/613630/training-a-single-ai-model-can-emit-as-much-carbon-as-five-cars-in-their-lifetimes/ 30 comments
- Attention Is All You Need - YouTube https://www.youtube.com/watch?v=iDulhoQ2pro 29 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- A new model and dataset for long-range memory https://deepmind.com/blog/article/A_new_model_and_dataset_for_long-range_memory 13 comments
- Attention and Augmented Recurrent Neural Networks https://distill.pub/2016/augmented-rnns/ 9 comments
- Generative Modeling with Sparse Transformers https://openai.com/blog/sparse-transformer/ 9 comments
- Deep Learning, NLP, and Representations - colah's blog http://colah.github.io/posts/2014-07-NLP-RNNs-Representations/ 9 comments
- Transformer-XL: Unleashing the Potential of Attention Models – Google AI Blog https://ai.googleblog.com/2019/01/transformer-xl-unleashing-potential-of.html 7 comments
- Language, trees, and geometry in neural networks https://pair-code.github.io/interpretability/bert-tree/ 5 comments
- The Annotated Transformer https://nlp.seas.harvard.edu/2018/04/03/attention.html 3 comments
- [1806.01261] Relational inductive biases, deep learning, and graph networks https://arxiv.org/abs/1806.01261 2 comments
- Introducing a Conditional Transformer Language Model for Controllable Generation https://blog.einstein.ai/introducing-a-conditional-transformer-language-model-for-controllable-generation/ 1 comment
- Making Transformer networks simpler and more efficient https://ai.facebook.com/blog/making-transformer-networks-simpler-and-more-efficient/ 1 comment
- MegatronLM: Training Billion+ Parameter Language Models Using GPU Model Parallelism - NVIDIA ADLR https://nv-adlr.github.io/MegatronLM 1 comment
- [1910.10683] Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer https://arxiv.org/abs/1910.10683 1 comment
Related searches:
Search whole site: site:graphdeeplearning.github.io
Search title: Transformers are Graph Neural Networks | NTU Graph Deep Learning Lab
See how to search.