Linking pages
Linked pages
- [1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding https://arxiv.org/abs/1810.04805 25 comments
- Medium https://medium.com/m/signin?isDraft=1&operation=login&redirect=https%3A%2F%2Fmedium.com%2F%40jamie_34747%2F79d382edf22b%3Fsource%3D 19 comments
- [1503.02531] Distilling the Knowledge in a Neural Network https://arxiv.org/abs/1503.02531 5 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/pytorch-pretrained-BERT 0 comments
Related searches:
Search whole site: site:towardsdatascience.com
Search title: Distilling BERT — How to achieve BERT performance using Logistic Regression | by Dima Shulga | Towards Data Science
See how to search.