Hacker News
- Liger-Kernel: Efficient Triton kernels for LLM training https://github.com/linkedin/Liger-Kernel 2 comments
- [P] Liger Kernel: One line to make LLM Training +20% faster and -60% memory https://github.com/linkedin/Liger-Kernel 17 comments machinelearning
Linking pages
- Qwen2.5: A Party of Foundation Models! | Qwen https://qwenlm.github.io/blog/qwen2.5/ 38 comments
- GitTrends - September 1 2024 - GitTrends https://gitstars.substack.com/p/gittrends-september-1-2024 0 comments
- GitHub - axolotl-ai-cloud/axolotl: Go ahead and axolotl questions https://github.com/axolotl-ai-cloud/axolotl 0 comments
- Liger Kernels Leap the CUDA Moat: A Case Study with Liger, LinkedIn's SOTA Training Kernels on AMD GPU https://embeddedllm.com/blog/cuda-to-rocm-portability-case-study-liger-kernel 0 comments
- How to fine-tune open LLMs in 2025 with Hugging Face https://www.philschmid.de/fine-tune-llms-in-2025 0 comments
Linked pages
- GitHub - unslothai/unsloth: 2x faster 50% less memory LLM finetuning https://github.com/unslothai/unsloth 122 comments
- https://arxiv.org/pdf/2002.05202.pdf 10 comments
- [2104.09864] RoFormer: Enhanced Transformer with Rotary Position Embedding https://arxiv.org/abs/2104.09864 8 comments
- CrossEntropyLoss — PyTorch 2.4 documentation https://pytorch.org/docs/stable/generated/torch.nn.CrossEntropyLoss.html 4 comments
- GitHub - microsoft/DeepSpeed: DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective. https://github.com/microsoft/DeepSpeed 1 comment
Would you like to stay up to date with Computer science? Checkout Computer science
Weekly.
Related searches:
Search whole site: site:github.com
Search title: GitHub - linkedin/Liger-Kernel: Efficient Triton Kernels for LLM Training
See how to search.