Linking pages
Linked pages
- PyTorch http://pytorch.org/ 100 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- [1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding https://arxiv.org/abs/1810.04805 25 comments
- TensorFlow Lite | ML for Mobile and Edge Devices https://www.tensorflow.org/lite/ 22 comments
- Medium https://medium.com/m/signin?isDraft=1&operation=login&redirect=https%3A%2F%2Fmedium.com%2F%40jamie_34747%2F79d382edf22b%3Fsource%3D 19 comments
- Generative Modeling with Sparse Transformers https://openai.com/blog/sparse-transformer/ 9 comments
- [1503.02531] Distilling the Knowledge in a Neural Network https://arxiv.org/abs/1503.02531 5 comments
- [1904.10509] Generating Long Sequences with Sparse Transformers https://arxiv.org/abs/1904.10509 1 comment
- TensorRT SDK | NVIDIA Developer https://developer.nvidia.com/tensorrt 0 comments
- Looking Back at Google’s Research Efforts in 2018 – Google AI Blog https://ai.googleblog.com/2019/01/looking-back-at-googles-research.html 0 comments
- Learn how to make BERT smaller and faster | The Rasa Blog | Rasa https://blog.rasa.com/compressing-bert-for-faster-prediction-2/ 0 comments
- NVIDIA Data Center Deep Learning Product Performance | NVIDIA Developer https://developer.nvidia.com/deep-learning-performance-training-inference 0 comments
- http://www.cs.cornell.edu/~caruana/compression.kdd06.pdf 0 comments
- [1909.11687] Extremely Small BERT Models from Mixed-Vocabulary Training https://arxiv.org/abs/1909.11687 0 comments
- https://openreview.net/pdf?id=H1eA7AEtvS 0 comments
- 🏎 Smaller, faster, cheaper, lighter: Introducing DistilBERT, a distilled version of BERT | by Victor Sanh | HuggingFace | Medium https://medium.com/huggingface/distilbert-8cf3380435b5 0 comments
Related searches:
Search whole site: site:blog.inten.to
Search title: Speeding up BERT. How to make BERT models faster | by Grigory Sapunov | Intento
See how to search.