Linking pages
- GitHub - IndicoDataSolutions/finetune: Scikit-learn style model finetuning for NLP https://github.com/IndicoDataSolutions/finetune/tree/master 8 comments
- A Visual Guide to Using BERT for the First Time – Jay Alammar – Visualizing machine learning one concept at a time. http://jalammar.github.io/a-visual-guide-to-using-bert-for-the-first-time/ 1 comment
- 2019 — Year of BERT and Transformer | by Manu Suryavansh | Towards Data Science https://towardsdatascience.com/2019-year-of-bert-and-transformer-f200b53d05b9?sk=77913662dd96ce5de77998341504902f&source=friends_link 0 comments
- The 5 Components Towards Building Production-Ready Machine Learning Systems — James Le https://jameskle.com/writes/testing-and-deployment 0 comments
- Machine Learning on Mobile and at the Edge: 2019 industry year-in-review | by Austin Kodra | Heartbeat https://heartbeat.fritz.ai/machine-learning-on-mobile-and-at-the-edge-2019-industry-year-in-review-3d783a404da4 0 comments
- Researchers find machine learning models still struggle to detect hate speech | VentureBeat https://venturebeat.com/2021/01/06/researchers-find-machine-learning-models-still-struggle-to-detect-hate-speech/ 0 comments
- Speeding up BERT. How to make BERT models faster | by Grigory Sapunov | Intento https://blog.inten.to/speeding-up-bert-5528e18bb4ea 0 comments
- Unlabelled Data’s Stock is Rising | by Connor Shorten | Towards Data Science https://medium.com/@connorshorten300/unlabelled-datas-stock-is-rising-71ed1cf909b7 0 comments
Linked pages
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- [1803.03635] The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks https://arxiv.org/abs/1803.03635 32 comments
- GitHub - google-research/bert: TensorFlow code and pre-trained models for BERT https://github.com/google-research/bert 21 comments
- [1503.02531] Distilling the Knowledge in a Neural Network https://arxiv.org/abs/1503.02531 5 comments
- GitHub - nyu-mll/jiant: jiant is an nlp toolkit https://github.com/nyu-mll/jiant 4 comments
- Nvidia trains world’s largest Transformer-based language model | VentureBeat https://venturebeat.com/2019/08/13/nvidia-trains-worlds-largest-transformer-based-language-model/ 3 comments
- GitHub - facebookresearch/XLM: PyTorch original implementation of Cross-lingual Language Model Pretraining. https://github.com/facebookresearch/XLM 0 comments
- Learn how to make BERT smaller and faster | The Rasa Blog | Rasa https://blog.rasa.com/compressing-bert-for-faster-prediction-2/ 0 comments
- GitHub - huggingface/swift-coreml-transformers: Swift Core ML 3 implementations of GPT-2, DistilGPT-2, BERT, and DistilBERT for Question answering. Other Transformers coming soon! https://github.com/huggingface/swift-coreml-transformers 0 comments
- http://www.cs.cornell.edu/~caruana/compression.kdd06.pdf 0 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/pytorch-transformers 0 comments
- [1907.11692] RoBERTa: A Robustly Optimized BERT Pretraining Approach https://arxiv.org/abs/1907.11692 0 comments