Linking pages
Linked pages
- How To Finetune GPT Like Large Language Models on a Custom Dataset - Lightning AI https://lightning.ai/pages/blog/how-to-finetune-gpt-like-large-language-models-on-a-custom-dataset/ 122 comments
- GitHub - Lightning-AI/lit-llama: Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed. https://github.com/Lightning-AI/lit-llama 69 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- Finetuning LLMs on a Single GPU Using Gradient Accumulation https://lightning.ai/pages/blog/gradient-accumulation/ 3 comments
- [2007.14062] Big Bird: Transformers for Longer Sequences https://arxiv.org/abs/2007.14062 0 comments
- [2010.11929] An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale https://arxiv.org/abs/2010.11929 0 comments
- Accelerating Large Language Models with Mixed-Precision Techniques - Lightning AI https://lightning.ai/pages/community/tutorial/accelerating-large-language-models-with-mixed-precision-techniques/ 0 comments
Related searches:
Search whole site: site:lightning.ai
Search title: Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch - Lightning AI
See how to search.