Linking pages
- GitHub - bigscience-workshop/petals: 🌸 Run 100B+ language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading https://github.com/bigscience-workshop/petals 155 comments
- GitHub - llSourcell/DoctorGPT: DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private. https://github.com/llSourcell/DoctorGPT 75 comments
- GitHub - Lightning-AI/lit-llama: Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed. https://github.com/Lightning-AI/lit-llama 69 comments
- How to make LLMs go fast https://vgel.me/posts/faster-inference/ 54 comments
- A quest for very long context: Part 1 - by The AI Dude https://naxalpha.substack.com/p/a-quest-for-very-long-context-part 42 comments
- GitHub - chris-alexiuk/alpaca-lora: Instruct-tune LLaMA on consumer hardware https://github.com/chris-alexiuk/alpaca-lora 30 comments
- Mini-post: first look at LLaMA. Background | by Enryu | Mar, 2023 | Medium https://medium.com/@enryu9000/mini-post-first-look-at-llama-4403517d41a1 27 comments
- Transformer Math 101 | EleutherAI Blog https://blog.eleuther.ai/transformer-math/ 13 comments
- GitHub - tloen/alpaca-lora: Instruct-tune LLaMA on consumer hardware https://github.com/tloen/alpaca-lora 11 comments
- GitHub - artidoro/qlora: QLoRA: Efficient Finetuning of Quantized LLMs https://github.com/artidoro/qlora 5 comments
- GitHub - bigcode-project/starcoder: Home of StarCoder: fine-tuning & inference! https://github.com/bigcode-project/starcoder 3 comments
- Low-Rank Pruning of Llama2 https://mobiusml.github.io/low-rank-llama2/ 3 comments
- GitHub - taprosoft/llm_finetuning: Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes) https://github.com/taprosoft/llm_finetuning 2 comments
- HQQ quantization https://mobiusml.github.io/hqq_blog/ 2 comments
- GitHub - learning-at-home/hivemind: Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world. https://github.com/learning-at-home/hivemind 1 comment
- BigDL/python/llm at main · intel-analytics/BigDL · GitHub https://github.com/intel-analytics/BigDL/tree/main/python/llm 1 comment
- The best open source software of 2023 | InfoWorld https://www.infoworld.com/article/3709196/the-best-open-source-software-of-2023.html 1 comment
- GitHub - bigcode-project/starcoder2 https://github.com/bigcode-project/starcoder2 1 comment
- How to deploy your deep learning side project on a budget – Lucky's Notes https://luckytoilet.wordpress.com/2023/06/20/how-to-deploy-your-deep-learning-side-project-on-a-budget/ 0 comments
- GitHub - RUCAIBox/LLMSurvey: The official GitHub page for the survey paper "A Survey of Large Language Models". https://github.com/RUCAIBox/LLMSurvey 0 comments
Linked pages
- [2208.07339] LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale https://arxiv.org/abs/2208.07339 33 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- LLM.int8() and Emergent Features — Tim Dettmers https://timdettmers.com/2022/08/17/llm-int8-and-emergent-features/ 15 comments
- Start Locally | PyTorch https://pytorch.org/get-started/locally/ 3 comments
- A Gentle Introduction to 8-bit Matrix Multiplication for transformers at scale using transformers, accelerate and bitsandbytes https://huggingface.co/blog/hf-bitsandbytes-integration 0 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - TimDettmers/bitsandbytes: 8-bit CUDA functions for PyTorch
See how to search.