- A beginner-friendly repo for fine-tuning LLMs with different quantization techniques in one package. There is also sample guide to deploy your own API server or chatUI. https://github.com/taprosoft/llm_finetuning 2 comments learnmachinelearning
Linked pages
- GitHub - ggerganov/llama.cpp: Port of Facebook's LLaMA model in C/C++ https://github.com/ggerganov/llama.cpp 286 comments
- GitHub - microsoft/guidance: A guidance language for controlling large language models. https://github.com/microsoft/guidance 190 comments
- GitHub - oobabooga/text-generation-webui: A Gradio web UI for Large Language Models. https://github.com/oobabooga/text-generation-webui 41 comments
- GitHub - tloen/alpaca-lora: Instruct-tune LLaMA on consumer hardware https://github.com/tloen/alpaca-lora 11 comments
- GitHub - psf/black: The uncompromising Python code formatter https://github.com/psf/black 8 comments
- GitHub - pre-commit/pre-commit: A framework for managing and maintaining multi-language pre-commit hooks. https://github.com/pre-commit/pre-commit 0 comments
- GitHub - qwopqwop200/GPTQ-for-LLaMa: 4 bits quantization of LLaMa using GPTQ https://github.com/qwopqwop200/GPTQ-for-LLaMa 0 comments
- GitHub - johnsmith0031/alpaca_lora_4bit https://github.com/johnsmith0031/alpaca_lora_4bit 0 comments
- GitHub - TimDettmers/bitsandbytes: 8-bit CUDA functions for PyTorch https://github.com/TimDettmers/bitsandbytes 0 comments
- WizardLM/WizardLM-30B-V1.0 · Hugging Face https://huggingface.co/WizardLM/WizardLM-30B-V1.0 0 comments
- GitHub - vllm-project/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs https://github.com/vllm-project/vllm 0 comments
- GitHub - turboderp/exllama: A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights. https://github.com/turboderp/exllama 0 comments