Hacker News
- Transformers 2.0: Deep interoperability between TensorFlow 2.0 and PyTorch https://github.com/huggingface/transformers 3 comments
- [N] Handling very large models on small setups with Hugging Face's Accelerate library https://github.com/huggingface/transformers 2 comments machinelearning
- [P] OSLO: Open Source framework for Large-scale transformer Optimization https://github.com/huggingface/transformers 4 comments machinelearning
- [D] How do pretrained tokenizers work? https://github.com/huggingface/transformers 7 comments machinelearning
- Rust tokenization for transformers https://github.com/huggingface/transformers 4 comments rust
- Huggingface Releases New Tokenizers Library Written in Rust https://github.com/huggingface/transformers 6 comments rust
Linking pages
- GitHub - normal-computing/outlines: Generative Model Programming https://github.com/normal-computing/outlines 304 comments
- GitHub - microsoft/visual-chatgpt: VisualChatGPT https://github.com/microsoft/visual-chatgpt 229 comments
- PyTorch 2.0 | PyTorch https://pytorch.org/get-started/pytorch-2.0/ 153 comments
- uv: Unified Python packaging https://astral.sh/blog/uv-unified-python-packaging 151 comments
- GitHub - neuml/txtai: 💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows https://github.com/neuml/txtai 138 comments
- GitHub - raghavan/PdfGptIndexer: An efficient tool for indexing and searching PDF text data using OpenAI's GPT-2 model and FAISS (Facebook AI Similarity Search) index, designed for rapid information retrieval and superior search accuracy. https://github.com/raghavan/PdfGptIndexer 137 comments
- Building LLM applications for production https://huyenchip.com/2023/04/11/llm-engineering.html 136 comments
- GitHub - karpathy/minGPT: A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training https://github.com/karpathy/minGPT 133 comments
- GitHub - microsoft/unilm: Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities https://github.com/microsoft/unilm 104 comments
- GitHub - suno-ai/bark: 🔊 Text-Prompted Generative Audio Model https://github.com/suno-ai/bark 93 comments
- GitHub - clovaai/donut: Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022 https://github.com/clovaai/donut 90 comments
- GitHub - InternLM/InternLM: Official release of InternLM2.5 base and chat models. 1M context support https://github.com/InternLM/InternLM 89 comments
- Ruff v0.4.0: a hand-written recursive descent parser for Python https://astral.sh/blog/ruff-v0.4.0 72 comments
- GitHub - EleutherAI/gpt-neox: An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library. https://github.com/EleutherAI/gpt-neox 67 comments
- GitHub - deepdoctection/deepdoctection: A Repo For Document AI https://github.com/deepdoctection/deepdoctection 63 comments
- txtai https://neuml.github.io/txtai/ 58 comments
- GitHub - code-kern-ai/refinery: The data scientist's open-source choice to scale, assess and maintain natural language data. Treat training data like a software artifact. https://github.com/code-kern-ai/refinery 57 comments
- GitHub - ankane/transformers-ruby: State-of-the-art transformers for Ruby https://github.com/ankane/transformers-ruby 56 comments
- GitHub - xenova/transformers.js: State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server! https://github.com/xenova/transformers.js 55 comments
- Open-Core Companies Are Not Incentivized To Make Their Projects Good https://www.plural.sh/blog/open-core-companies/ 50 comments
Linked pages
- TensorFlow http://tensorflow.org/ 440 comments
- [2205.01068] OPT: Open Pre-trained Transformer Language Models https://arxiv.org/abs/2205.01068 318 comments
- Better Language Models and Their Implications https://blog.openai.com/better-language-models/#content 207 comments
- GitHub - BlinkDL/RWKV-LM: RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding. https://github.com/BlinkDL/RWKV-LM 179 comments
- GitHub - kingoflolz/mesh-transformer-jax: Model parallel transformers in JAX and Haiku https://github.com/kingoflolz/mesh-transformer-jax 146 comments
- GitHub - EleutherAI/gpt-neo: An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library. https://github.com/EleutherAI/gpt-neo/ 127 comments
- PyTorch http://pytorch.org/ 100 comments
- GitHub - google/jax: Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more https://github.com/google/jax 99 comments
- Write With Transformer https://transformer.huggingface.co/ 86 comments
- Image GPT https://openai.com/blog/image-gpt/ 84 comments
- [1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding https://arxiv.org/abs/1810.04805 25 comments
- BERTweet https://huggingface.co/docs/transformers/model_doc/bertweet 23 comments
- [2103.14030] Swin Transformer: Hierarchical Vision Transformer using Shifted Windows https://arxiv.org/abs/2103.14030 20 comments
- Improving language understanding with unsupervised learning https://blog.openai.com/language-unsupervised/ 18 comments
- [1906.08237] XLNet: Generalized Autoregressive Pretraining for Language Understanding https://arxiv.org/abs/1906.08237 15 comments
- [2105.13626] ByT5: Towards a token-free future with pre-trained byte-to-byte models https://arxiv.org/abs/2105.13626 15 comments
- [2209.10655] Mega: Moving Average Equipped Gated Attention https://arxiv.org/abs/2209.10655 14 comments
- [2002.08909] REALM: Retrieval-Augmented Language Model Pre-Training https://arxiv.org/abs/2002.08909 13 comments
- GitHub - s-JoL/Open-Llama: The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF. https://github.com/s-JoL/Open-Llama 13 comments
- BigScience Research Workshop https://bigscience.huggingface.co 11 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
See how to search.