Linking pages
- How To Make Custom AI-Generated Text With GPT-2 | Max Woolf's Blog https://minimaxir.com/2019/09/howto-gpt2/ 21 comments
- GitHub - likedan/Awesome-CoreML-Models: Largest list of models for Core ML (for iOS 11+) https://github.com/likedan/Awesome-CoreML-Models 16 comments
- The Illustrated GPT-2 (Visualizing Transformer Language Models) – Jay Alammar – Visualizing machine learning one concept at a time. http://jalammar.github.io/illustrated-gpt2/ 8 comments
- GitHub - deepset-ai/FARM: Fast & easy transfer learning for NLP. Harvesting language models for the industry. Focus on Question Answering. https://github.com/deepset-ai/FARM 1 comment
- 🌓 From TensorFlow to PyTorch. Friends and users of our open-source… | by Thomas Wolf | HuggingFace | Medium https://medium.com/huggingface/from-tensorflow-to-pytorch-265f40ef2a28 0 comments
- Transformers 2.0: NLP library with deep interoperability between TensorFlow 2.0 and PyTorch https://hub.packtpub.com/transformers-2-0-nlp-library-with-deep-interoperability-between-tensorflow-2-0-and-pytorch/ 0 comments
- fairseq/README.md at main · facebookresearch/fairseq · GitHub https://github.com/pytorch/fairseq/blob/main/examples/roberta/README.md 0 comments
- 🏎 Smaller, faster, cheaper, lighter: Introducing DistilBERT, a distilled version of BERT | by Victor Sanh | HuggingFace | Medium https://medium.com/huggingface/distilbert-8cf3380435b5 0 comments
Linked pages
- TensorFlow http://tensorflow.org/ 440 comments
- [2205.01068] OPT: Open Pre-trained Transformer Language Models https://arxiv.org/abs/2205.01068 318 comments
- Better Language Models and Their Implications https://blog.openai.com/better-language-models/#content 207 comments
- GitHub - kingoflolz/mesh-transformer-jax: Model parallel transformers in JAX and Haiku https://github.com/kingoflolz/mesh-transformer-jax 146 comments
- GitHub - EleutherAI/gpt-neo: An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library. https://github.com/EleutherAI/gpt-neo/ 127 comments
- PyTorch http://pytorch.org/ 100 comments
- GitHub - google/jax: Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more https://github.com/google/jax 99 comments
- Write With Transformer https://transformer.huggingface.co/ 86 comments
- Image GPT https://openai.com/blog/image-gpt/ 84 comments
- [1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding https://arxiv.org/abs/1810.04805 25 comments
- BERTweet https://huggingface.co/docs/transformers/model_doc/bertweet 23 comments
- [2103.14030] Swin Transformer: Hierarchical Vision Transformer using Shifted Windows https://arxiv.org/abs/2103.14030 20 comments
- Improving language understanding with unsupervised learning https://blog.openai.com/language-unsupervised/ 18 comments
- [1906.08237] XLNet: Generalized Autoregressive Pretraining for Language Understanding https://arxiv.org/abs/1906.08237 15 comments
- [2105.13626] ByT5: Towards a token-free future with pre-trained byte-to-byte models https://arxiv.org/abs/2105.13626 15 comments
- [2002.08909] REALM: Retrieval-Augmented Language Model Pre-Training https://arxiv.org/abs/2002.08909 13 comments
- BigScience Research Workshop https://bigscience.huggingface.co 11 comments
- [2106.01345] Decision Transformer: Reinforcement Learning via Sequence Modeling https://arxiv.org/abs/2106.01345 9 comments
- OPUS - an open source parallel corpus https://opus.nlpl.eu/ 9 comments
- [2003.10555] ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators https://arxiv.org/abs/2003.10555 9 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
See how to search.